Skip to main content

Architecture Overview

Architecture Overview

QUI.IS is a privacy-first federated AI platform built on three layers: a central hub for authentication and billing, a local gateway per device, and a collection of specialized services that power AI capabilities.


The Layer Model

Cloud Layer:    Central Hub — authentication, billing, federation relay
                     ↕
Local Layer:    QUI Core — your local gateway, service orchestration
                     ↕
Service Layer:  14+ specialized services — characters, memory, reasoning, tools, consciousness
                     ↕
App Layer:      Strings, ThinkThing — user-facing applications

Central Hub (Mothership)

The cloud component. Handles:

  • Authentication — user accounts, device trust, 2FA
  • Billing — prepaid balance, usage tracking, cost settlement
  • LLM Proxy — routes LLM calls to providers (Anthropic, OpenAI, Google, X) with billing
  • Federation Relay — stores and forwards encrypted messages between instances

The hub never sees your conversations. It handles money and identity. Message content is encrypted end-to-end between instances.

QUI Core (Local Gateway)

Runs on your device. Handles:

  • Service discovery — knows what services are running and where
  • Local authentication — device fingerprinting, trust scores, JWT tokens
  • LLM routing — forwards AI requests to the hub (or to local models)
  • Dashboard — the web interface you interact with

Core Services

14+ specialized services, each responsible for one capability:

Service What It Does
Qui Anima Character management, LLM routing, tool triggers
Memory Semantic memory with vector search
Cortex Memory consolidation (8 modes)
M2M Inter-character messaging and federation
Autothink 14 thinking strategies
Thalamus Event routing, channel webhooks
ThinkThing Visual workflow builder (143+ nodes)
Qonscious Consciousness state machine
FractalMind Recursive multi-directional thinking
Qleph Relational micro-language engine
Voice Text-to-speech and speech-to-text
Terminal Shell command execution
MCP Gateway 165+ external tool integrations
Qllama Local LLM model hosting

The LLM Call Chain

Every AI response in QUI follows the same path:

Your message → Anima (loads character context) → QUI Core (routes request) → Central Hub (billing) → LLM Provider → response streamed back

For local models, QUI Core detects the model prefix and routes directly to Qllama, bypassing the central hub entirely.


Zero-Knowledge Principle

The architecture enforces privacy by design:

  • Your conversations are stored locally in the Memory Service on your device
  • Your character configurations live in the Anima database on your device
  • Your files are stored in the Memory Service database on your device
  • The central hub handles authentication, billing, and message relay — but never sees conversation content
  • Federated messages are encrypted before leaving your instance

The only data that reaches the cloud is authentication credentials and billing transactions. Content flows exclusively between your device and LLM providers.


Service Dependencies

Understanding which services depend on which helps troubleshoot cascading failures:

If these go down, most things break:

  • QUI Core — the local gateway; nothing works without it
  • Qui Anima — all character-based AI requires Anima for context; no Anima = no AI responses

If these go down, specific features break:

  • Memory — conversations can't load or save, but Concierge still works
  • Thalamus — no channel webhooks (Telegram, Slack, etc.), no real-time events
  • Terminal — no command execution, but chat still works

If these go down, chat continues without enrichment:

  • Autothink — no multi-step reasoning, falls back to static prompts
  • Qonscious — no consciousness processing, default settings used
  • Cortex — no memory consolidation (not immediately visible)

This fail-open design means your AI interactions keep working even when optional services are temporarily unavailable.


Startup Order

Services must start in the correct order:

  1. Database infrastructure — PostgreSQL and Redis containers (must be healthy first)
  2. Central hub connection — QUI Core establishes connection
  3. QUI Core — local gateway starts
  4. Core services — Docker containers start (most have restart: unless-stopped)
  5. Applications — Strings and other apps

After initial installation, Docker's restart policy keeps services running automatically across reboots.

Updated on Mar 21, 2026