Dashboard Tour
Dashboard Tour
The QUI Core dashboard is your control center. Everything is accessible from the sidebar on the left. This guide walks through each section.
[Screenshot: Full dashboard with sidebar visible]
Main Tabs
Dashboard
Your system overview. Shows CPU usage, RAM, disk space, and a quick summary of running services. This is your home screen.
Chat (Concierge)
A built-in AI assistant that helps you navigate QUI, troubleshoot issues, and run diagnostic commands on your system. See Concierge for details.
Apps
Browse and install applications from the QUI marketplace. Apps require a minimum trust score of 30.
Strings — the primary chat interface — is available here. It launches in your browser as a separate application.
[Screenshot: Apps tab showing available applications]
Terminal
A full terminal emulator built into the dashboard. Powered by xterm.js with WebSocket connectivity. Supports multiple sessions.
Services
Services Status
Shows all running QUI services with their health status, port numbers, and uptime. You can view logs and restart individual services from here.
[Screenshot: Services tab showing service list with status indicators]
Core Memory
The semantic memory service. View memory statistics, search stored memories, and monitor the embedding model.
M2M (Machine-to-Machine)
Inter-agent messaging and federation. See connected characters, pending connection requests, and messaging statistics.
Cortex
Memory consolidation controls. View processing jobs, select consolidation modes, and monitor how raw memories are being structured into long-term knowledge.
Autothink
Thinking strategy configuration. View available strategies and test them against sample inputs.
Thalamus
Event routing hub. Monitor active channels, trigger schedules, and adapter status.
Qonscious
Consciousness state machine monitoring. View coherence, arousal, valence, and other consciousness metrics for your characters.
Qleph
The relational micro-language engine. View dictionary entries and test expressions.
FractalMind
Recursive thinking sessions. Launch fractal reasoning sessions and visualize the branching thought tree.
System
Local Models
Manage locally hosted LLM models via Qllama. Pull new models, view loaded models, and monitor GPU usage.
CORE Connect
Device identity and trust score management. See First Login for details.
Remote Access
Configure remote access to your QUI Core instance from other devices. Supports WebRTC peer-to-peer connections and central hub tunnel as fallback.
Model Pricing
View pricing for all four cloud LLM providers (Anthropic, OpenAI, Google, X) side by side. Costs are shown per 1K tokens for both input and output.
System Stats
Detailed system statistics — CPU, RAM, disk, network, and GPU metrics.
System Tests
Built-in test suites for verifying service connectivity, authentication flows, and endpoint health.
Settings
Account
Your user profile — username, email, and account settings.
Terminal Safety
Configure terminal safety modes for services that use terminal access. Choose between basic mode (commands execute directly) and enhanced mode (dangerous commands require approval).
Appearance
Customize the dashboard dock — opacity, size, position, and the wave animation effect on hover.
[Screenshot: Appearance settings showing dock customization]
Billing
View your current balance, usage history, spending limits, and add funds. All LLM usage costs are tracked here. See Billing for details.
System
System-level configuration settings.
Navigation Tips
- The sidebar collapses on smaller screens — click the hamburger menu to expand it
- The active tab is remembered between sessions
- You can access any tab directly via URL hash — for example,
https://localhost:10009/os/dashboard#cortexopens the Cortex tab - The dock at the bottom provides quick access to your most-used applications