edge_veda 1.1.1
edge_veda: ^1.1.1 copied to clipboard
On-device LLM inference SDK for Flutter. Run Llama, Phi, and other language models locally with Metal GPU acceleration on iOS devices.
Changelog #
All notable changes to the Edge Veda Flutter SDK will be documented in this file.
The format is based on Keep a Changelog, and this project adheres to Semantic Versioning.
1.1.1 - 2026-02-09 #
Fixed #
- License corrected to Apache 2.0 (was incorrectly MIT in pub.dev package)
- README rewritten with accurate capabilities and real soak test metrics
- CHANGELOG cleaned up to reflect only shipped features
1.1.0 - 2026-02-08 #
Added #
- Vision (VLM): SmolVLM2-500M support for real-time camera-to-text inference
- Chat Session API: Multi-turn conversation management with context overflow summarization
- Chat Templates: Llama 3 Instruct, ChatML, and generic template formats
- System Prompt Presets: Built-in assistant, coder, and creative personas
- VisionWorker: Persistent isolate for vision inference (model loads once, reused across frames)
- FrameQueue: Drop-newest backpressure for camera frame processing
- RuntimePolicy: Adaptive QoS with thermal/battery/memory-aware hysteresis
- TelemetryService: iOS thermal state, battery level, memory polling via MethodChannel
- PerfTrace: JSONL performance trace logger for soak test analysis
- Soak Test Screen: 15-minute automated vision benchmark in demo app
initVision()anddescribeImage()APIsCameraUtilsfor BGRA/YUV420 to RGB conversion- Context indicator (turn count + usage bar) in demo Chat tab
- New Chat button and persona picker in demo app
Changed #
- Upgraded llama.cpp from b4658 to b7952
- XCFramework rebuilt with all symbols including
ev_vision_get_last_timings - Demo app redesigned with dark theme, 3-tab navigation (Chat, Vision, Settings)
- Chat tab rewritten to use ChatSession API (no direct generate() calls)
- All FFI bindings now eager (removed lazy workaround for missing symbols)
- Constrained ffi to <2.1.0 (avoids objective_c simulator crash)
Fixed #
- Xcode 26 debug blank executor: export
_mainin podspec symbol whitelist - RuntimePolicy evaluate() de-escalation when pressure improves but persists
1.0.0 - 2026-02-04 #
Added #
- Core SDK: On-device LLM inference via llama.cpp with Metal GPU on iOS
- Dart FFI: 37 native function bindings via
DynamicLibrary.process() - Streaming: Token-by-token generation with
CancelTokencancellation - Model Management: Download, cache, SHA-256 verify, delete
- Memory Monitoring: RSS tracking, pressure callbacks, configurable limits
- Isolate Safety: All FFI calls in
Isolate.run(), persistentStreamingWorker - XCFramework: Device arm64 + simulator arm64 static library packaging
- Demo App: Chat screen with streaming, model selection, benchmark mode
- Exception Hierarchy: 10 typed exceptions mapped from native error codes