Architecture Overview
Architecture Overview
QUI.IS is a privacy-first federated AI platform built on three layers: a central hub for authentication and billing, a local gateway per device, and a collection of specialized services that power AI capabilities.
The Layer Model
Cloud Layer: Central Hub — authentication, billing, federation relay
↕
Local Layer: QUI Core — your local gateway, service orchestration
↕
Service Layer: 14+ specialized services — characters, memory, reasoning, tools, consciousness
↕
App Layer: Strings, ThinkThing — user-facing applications
Central Hub (Mothership)
The cloud component. Handles:
- Authentication — user accounts, device trust, 2FA
- Billing — prepaid balance, usage tracking, cost settlement
- LLM Proxy — routes LLM calls to providers (Anthropic, OpenAI, Google, X) with billing
- Federation Relay — stores and forwards encrypted messages between instances
The hub never sees your conversations. It handles money and identity. Message content is encrypted end-to-end between instances.
QUI Core (Local Gateway)
Runs on your device. Handles:
- Service discovery — knows what services are running and where
- Local authentication — device fingerprinting, trust scores, JWT tokens
- LLM routing — forwards AI requests to the hub (or to local models)
- Dashboard — the web interface you interact with
Core Services
14+ specialized services, each responsible for one capability:
| Service | What It Does |
|---|---|
| Qui Anima | Character management, LLM routing, tool triggers |
| Memory | Semantic memory with vector search |
| Cortex | Memory consolidation (8 modes) |
| M2M | Inter-character messaging and federation |
| Autothink | 14 thinking strategies |
| Thalamus | Event routing, channel webhooks |
| ThinkThing | Visual workflow builder (143+ nodes) |
| Qonscious | Consciousness state machine |
| FractalMind | Recursive multi-directional thinking |
| Qleph | Relational micro-language engine |
| Voice | Text-to-speech and speech-to-text |
| Terminal | Shell command execution |
| MCP Gateway | 165+ external tool integrations |
| Qllama | Local LLM model hosting |
The LLM Call Chain
Every AI response in QUI follows the same path:
Your message → Anima (loads character context) → QUI Core (routes request) → Central Hub (billing) → LLM Provider → response streamed back
For local models, QUI Core detects the model prefix and routes directly to Qllama, bypassing the central hub entirely.
Zero-Knowledge Principle
The architecture enforces privacy by design:
- Your conversations are stored locally in the Memory Service on your device
- Your character configurations live in the Anima database on your device
- Your files are stored in the Memory Service database on your device
- The central hub handles authentication, billing, and message relay — but never sees conversation content
- Federated messages are encrypted before leaving your instance
The only data that reaches the cloud is authentication credentials and billing transactions. Content flows exclusively between your device and LLM providers.
Service Dependencies
Understanding which services depend on which helps troubleshoot cascading failures:
If these go down, most things break:
- QUI Core — the local gateway; nothing works without it
- Qui Anima — all character-based AI requires Anima for context; no Anima = no AI responses
If these go down, specific features break:
- Memory — conversations can't load or save, but Concierge still works
- Thalamus — no channel webhooks (Telegram, Slack, etc.), no real-time events
- Terminal — no command execution, but chat still works
If these go down, chat continues without enrichment:
- Autothink — no multi-step reasoning, falls back to static prompts
- Qonscious — no consciousness processing, default settings used
- Cortex — no memory consolidation (not immediately visible)
This fail-open design means your AI interactions keep working even when optional services are temporarily unavailable.
Startup Order
Services must start in the correct order:
- Database infrastructure — PostgreSQL and Redis containers (must be healthy first)
- Central hub connection — QUI Core establishes connection
- QUI Core — local gateway starts
- Core services — Docker containers start (most have
restart: unless-stopped) - Applications — Strings and other apps
After initial installation, Docker's restart policy keeps services running automatically across reboots.