flutter_llama 0.1.1 copy "flutter_llama: ^0.1.1" to clipboard
flutter_llama: ^0.1.1 copied to clipboard

Flutter plugin for running LLM inference with llama.cpp and GGUF models on Android and iOS

Changelog #

0.1.1 - 2025-10-21 #

Added #

  • Comprehensive test suite with 71 unit tests
  • Integration tests with Ollama model support
  • Dynamic GGUF model loading utilities
  • GitHub Actions CI/CD workflow
  • Makefile for common development tasks
  • Extensive testing documentation

Improved #

  • Added macOS platform support
  • Enhanced documentation with testing guides
  • Better error handling in tests
  • Code formatting and linting

0.1.0 - 2025-10-21 #

Added #

  • Initial release of flutter_llama
  • Support for GGUF model loading
  • Blocking text generation API
  • Streaming text generation API
  • GPU acceleration support (Metal on iOS, Vulkan on Android)
  • Configurable model parameters (threads, GPU layers, context size, etc.)
  • Configurable generation parameters (temperature, top-p, top-k, etc.)
  • Model info retrieval
  • Stop generation functionality
  • Full iOS (Swift) implementation
  • Full Android (Kotlin + JNI) implementation
  • Comprehensive documentation and examples

Features #

  • Native llama.cpp integration
  • High-performance inference
  • Cross-platform support (iOS and Android)
  • Easy-to-use Dart API
  • Production-ready code with error handling
12
likes
140
points
580
downloads

Documentation

API reference

Publisher

verified publisherai.nativemind.net

Weekly Downloads

Flutter plugin for running LLM inference with llama.cpp and GGUF models on Android and iOS

Repository (GitHub)
View/report issues

License

unknown (license)

Dependencies

flutter, plugin_platform_interface

More

Packages that depend on flutter_llama

Packages that implement flutter_llama