mlx_localllm 0.1.0 copy "mlx_localllm: ^0.1.0" to clipboard
mlx_localllm: ^0.1.0 copied to clipboard

retracted

A high-performance local LLM plugin for macOS using Apple's MLX framework.

mlx_localllm #

English | 简体中文


English #

A Flutter plugin for high-performance localized LLM inference on macOS using Apple's MLX framework. It provides zero-latency interaction by loading models directly into the application process.

Features #

  • 🚀 Native Acceleration: Built on MLX for optimal performance on Apple Silicon.
  • 📦 In-Process: No external dependencies like Ollama required.
  • 🌐 Robust Downloader: Built-in support for HuggingFace and mirrors (hf-mirror.com) with resilient progress reporting.
  • 🛠️ Easy Integration: Comprehensive API for model management and inference.

Requirements #

  • macOS: 14.0 or higher.
  • Hardware: Apple Silicon (M1, M2, M3, etc.).
  • Flutter: 3.0.0 or higher.

Installation #

  1. Add the dependency to your pubspec.yaml:

    flutter pub add mlx_localllm
    
  2. CRITICAL STEP: Since this plugin relies on Swift Package Manager (SPM) for MLX, you must run the configuration script in your macos directory:

    # Navigate to your project's macos directory
    cd macos
    # Run the setup script provided by the plugin
    ruby ../path/to/mlx_localllm/scripts/setup_example_macos.rb
    

    (Note: For production apps, you may need to manually add the SPM dependencies MLX, MLXLLM, Hub, and Tokenizers to your Runner target in Xcode if you don't use the script).

Usage #

Check Support

bool supported = await MlxLocalllm.isSupported();

Download Model

MlxLocalllm.modelManager.downloadModel('mlx-community/Qwen2.5-0.5B-Instruct-4bit')
  .listen((progress) {
    print('Download progress: ${progress.progress * 100}%');
  });

Run Inference

await MlxLocalllm.inferenceEngine.loadModel('mlx-community/Qwen2.5-0.5B-Instruct-4bit');
String response = await MlxLocalllm.inferenceEngine.generate('Hello, who are you?');
print(response);

简体中文 #

基于 Apple MLX 框架的 macOS 高性能本地大模型推理 Flutter 插件。通过将模型直接加载到应用进程中,提供零延迟的交互体验。

特性 #

  • 🚀 原生加速: 基于 MLX 针对 Apple Silicon 深度优化。
  • 📦 进程内推理: 无需安装 Ollama 等外部服务,集成度高。
  • 🌐 鲁棒下载器: 内置对 HuggingFace 及其镜像站(如 hf-mirror.com)的支持,优化了进度上报稳定性。
  • 🛠️ 全功能 API: 提供完整的模型下载、删除、状态检查库及推理引擎。

系统要求 #

  • 系统: macOS 14.0 及以上。
  • 硬件: Apple Silicon 系列芯片 (M1, M2, M3 等)。
  • Flutter: 3.0.0 及以上。

安装指南 #

  1. pubspec.yaml 中添加依赖:

    flutter pub add mlx_localllm
    
  2. 关键步骤: 该插件通过 Swift Package Manager (SPM) 引用 MLX,由于 CocoaPods 与 SPM 的集成限制,您需要在项目的 macos 目录运行配置脚本:

    cd macos
    # 运行插件自带的配置脚本(指向实际插件路径)
    ruby path/to/mlx_localllm/scripts/setup_example_macos.rb
    

快速开始 #

硬件支持检测

bool supported = await MlxLocalllm.isSupported();

模型下载

MlxLocalllm.modelManager.downloadModel('mlx-community/Qwen2.5-0.5B-Instruct-4bit')
  .listen((progress) {
    print('下载进度: ${progress.progress * 100}%');
  });

执行推理

await MlxLocalllm.inferenceEngine.loadModel('mlx-community/Qwen2.5-0.5B-Instruct-4bit');
String response = await MlxLocalllm.inferenceEngine.generate('你好,请做下自我介绍');
print(response);

许可证 #

MIT License.

1
likes
0
points
212
downloads

Publisher

unverified uploader

Weekly Downloads

A high-performance local LLM plugin for macOS using Apple's MLX framework.

Repository (GitHub)
View/report issues

License

unknown (license)

Dependencies

flutter, http, path, path_provider, plugin_platform_interface

More

Packages that depend on mlx_localllm

Packages that implement mlx_localllm