🎵 Now playing in the lab: Loading...

🌙 nocturne.mediagato.com

Local AI assistant for the Mediagato network

🧠 Ollama Models ⚡ FastAPI Streaming 🗂️ SQLite History 🔒 On-Prem Deployment

Nocturne delivers a stylish, fast local assistant with modern UX expectations. Built for Mediagato staff and trusted collaborators, it combines streaming chat, persistent history, and optional web search—all powered by local Ollama models with zero external API dependencies.

🧠 Backend Core

  • FastAPI with streaming responses
  • Session-based authentication
  • SQLite chat history storage
  • Optional DuckDuckGo web search

🖥️ Frontend

  • SSE-driven streaming UI
  • Persistent conversation history
  • Model selector dropdown
  • One-click message copy

⚙️ Runtime & Infra

  • systemd-managed services
  • Ubuntu 24.04 LTS on Intel NUC
  • 1TB storage for models & history
  • CPU-only inference (no GPU required)

🎨 User Experience

  • Mediagato dark theme with purple accents
  • Glassmorphism UI effects
  • Mobile-optimized responsive layout
  • Compact history dropdown navigation

Model Inventory

Model Size Primary Role
qwen2.5:1.5b 986 MB Default balanced assistant
tinyllama:1.1b 637 MB Ultra-fast fallback
llama3.2:1b 1.3 GB Meta Llama 3.2 compact
gemma2:2b 1.6 GB Google Gemma v2
qwen2.5-coder:1.5b 986 MB Code-focused variant
phi3:mini 2.2 GB Microsoft Phi-3
deepseek-coder:1.3b-instruct 776 MB Lightweight coder
deepseek-coder:6.7b-instruct 3.8 GB Highest-quality coder

With Ollama storage relocated to a 1TB volume and systemd supervision for both FastAPI and Ollama, Nocturne maintains rapid local inference without sacrificing retention. All chat history is stored in SQLite, ensuring conversations persist across sessions while keeping data entirely on-premises.