Local AI.
Unleashed with MLX.
The open source, native macOS app for running LLMs locally. Maximum performance, complete privacy.
Blazing Fast Tokens/Sec
Built directly on Apple's MLX framework to deliver industry-leading inference speeds on M-series chips.
Native Feel
Built with SwiftUI for a truly Mac-like experience. Smooth animations, frosted glass, and seamless integration.
Open Source
Transparency is key. Inspect the code, contribute features, and run models with zero compiled secrets.
Benchmarks. Off the charts.
Leveraging the Neural Engine to deliver token generation speeds that leave others in the dust.
Llama-3-8B-Instruct on MacBook Pro M3 Max
Run the best models.
Llama 3
Mistral
Gemma
Falcon
Phi-3
StarCoder