No cloud, no internet, no tracking.
Machine learning is a field of computer science where systems are taught to learn patterns and make decisions from data, without being explicitly programmed for every specific task.
The goal is to enable computers to predict outcomes or classify new, unseen data. For example, if you show it thousands of pictures labeled "cat" and "dog," the machine learns the features that define a cat versus a dog.
The model weights live on your phone. Prompts, drafts and answers never touch a server — there isn't one to touch.
Airplane mode? Subway? Flight? No difference. Inference happens on your SoC, never on someone else's server.
Save reusable system prompts once and pick the right one for any model. Keep tone, role and output format consistent across every session.
Temperature, Top-P, Top-K, Min-P, repetition penalty, seed, context size — each model remembers its own settings.
Models are multi-gigabyte. Download them to any folder — internal storage, SD card, or an external drive. Move them between locations any time.
Browse community-made models optimized for mobile — compact enough to fit on your phone, powerful enough to be useful.
Browse the curated list or paste your own model URL. Sizes range from 267 MB to 5.4 GB.
Reliable background downloads with notifications, speed, and ETA — with automatic resume when the connection drops.
Your conversation history stays on your device. Reasoning is shown inline. No accounts, no API keys.
The whole app is translated into 28 languages. On first launch, we'll pick a model that speaks yours.
Built on llama.cpp with GGUF quantized models. Native C++ inference with ARM-optimized kernels means less heat, more tokens per second, and no unexpected background drain.
Still stuck? Open an issue on GitHub or reach out directly.
Seconds to install. Minutes to download a model. Then you're done with the cloud.