Llamatik is a growing collection of local-first AI products designed for privacy, control, and real-world usefulness.
Chat with AI, generate content, and run models locally — all on your device.
Use the Llamatik App for private, on-device AI, or explore the platform behind it with the Llamatik Library.
Most AI today depends entirely on remote servers. That means your data leaves your device, latency increases, and costs scale with usage.
Llamatik takes a different approach.
By running AI locally when possible, Llamatik gives you more control over your data, faster interactions, and a more sustainable way to use AI — without constant reliance on the cloud.
Private AI, on your device
A simple, powerful AI assistant that runs locally.
No accounts, no tracking, and no unnecessary cloud dependency.
The platform behind the app
A Kotlin Multiplatform library for integrating AI across Android, iOS, desktop, and server.

Your data stays on your device. No hidden tracking, and no external processing unless you choose it.
Run AI where it makes sense — on-device when possible, remote when needed.
Reducing unnecessary cloud usage means less energy consumption and a more efficient AI stack.
Llamatik Library focuses on the hard parts of integrating AI — so you can focus on building great products.
Instead of stitching together native code, bindings, and platform-specific logic, Llamatik provides a unified Kotlin-first abstraction that works across platforms and environments.
Run LLMs fully offline using native llama.cpp bindings compiled for each platform. No network required, no data leakage.
A single, shared Kotlin API for Android, iOS, desktop, and server — with expect/actual handled for you.
Use HTTP-based inference when models are too large or when centralized execution is required — without changing your app logic.
Built-in support for common LLM use cases like text generation, chat-style prompts, and vector embeddings.
Works with modern GGUF-based models such as LLaMA, Mistral, and Phi.
No heavy frameworks, no cloud SDKs. Just Kotlin, native binaries, and full control over your stack.
implementation("com.llamatik:library")
Cloud-based AI is powerful — but it comes with trade-offs: latency, recurring costs, privacy concerns, and vendor lock-in.
Llamatik lets you run models directly on user devices using native inference powered by llama.cpp, reducing infrastructure costs while keeping sensitive data local.
When you need scale or centralized execution, you can switch to remote inference using the same core platform.
Local-first by default. Remote when needed.

Llamatik is open source and built in the open.
You can inspect the code, understand how it works, and use it in your own projects.
Whether you use the app or build with the library, everything is designed with transparency and long-term trust in mind.
Powered by battle-tested tools used in production apps.
Compose Multiplatform
Material Design 3
Download the app or try Llamatik directly in your browser.