Open Source · Android · 100% Local
Mias runs AI models directly on your Android device. Ask questions, write code, get help — even when you're offline. Built for students who live on their phone.
Why Mias
No subscription. No login. No data leaving your phone.
In the library, on the bus, in a dead zone — Mias keeps working. Your assistant isn't tied to Wi-Fi or mobile data.
Every conversation is encrypted locally. No OpenAI, no Anthropic, no Firebase. Zero cloud calls — ever.
Simple question? Runs a 1 GB model to save battery. Complex code? Offloads to your PC's GPU over Wi-Fi.
Mias builds a local memory from your conversations — topics you care about, how you like to be answered.
Phone getting hot? Mias automatically drops to a lighter model. Stays fast without cooking your battery.
Read files, search the web, set reminders, open apps. Not just a chatbox — it can act on your device.
How it works
Mias reads your intent and routes it to the best available model — on-device or your desktop if connected.
ONNX Runtime or llama.cpp processes your query on-device (NPU/CPU) or on your PC's GPU via Tailscale.
Response streams back in real-time. If a tool was needed — file read, web fetch — you see what ran and why.
Key points are compressed into long-term memory every few hours. Your next session is already smarter.
Models
All models are downloaded directly from HuggingFace. You pick what to install.
| Model | Size | Runs on | Hardware | Best for |
|---|---|---|---|---|
| MobileLLM-R1.5 | 1.1 GB | Phone | CPU | Quick questions, always-on, low battery use |
| Gemma-4 INT4 | 3.2 GB | Phone | NPU | General chat, faster responses, daily use |
| Qwen3-Coder-Next Q4_K_M | 18.5 GB | Desktop | GPU | Code, research, complex reasoning tasks |
Setup
# Clone the repo git clone https://github.com/nikhlgoel/mias.git cd mias # Build (first time ~5 min) ./gradlew assembleDebug # Install on phone (USB debugging on) adb install -r app/build/outputs/apk/debug/app-debug.apk # Done. Launch the app, register biometric, # go to Brain Market and download a model.
# Build the server container cd desktop docker build -t mias-desktop:latest . # Download Qwen3 model (~18 GB, one-time) pip install huggingface-hub huggingface-cli download Qwen/Qwen3-Coder-Next-GGUF \ Qwen3-Coder-Next-32B-Q4_K_M.gguf # Run the server docker run --gpus all -p 8400:8400 mias-desktop:latest # Verify curl localhost:8400/health → {"status":"ready","device":"cuda"}
# Install Tailscale on both devices # Android: Play Store → Tailscale # Desktop: tailscale up --accept-routes tailscale status # Copy your desktop IP from output # In the app: # Settings → Networking → enter IP → Test # You'll see ✓ Connected
# For the Android app: Java 21+ Android Studio + SDK 35 USB debugging enabled 4 GB RAM phone minimum # For desktop offload (optional): Docker Python 3.11+ CUDA 12.0+ GPU (or CPU, slower) ~20 GB free disk space
Privacy
What Mias will never do:
What Mias enforces by design:
Open Source
MIT-adjacent stack. Full source on GitHub. No black boxes.