Model Hub Dashboard
Discover, download, and manage Large Language Models
Overview
The Model Hub provides a comprehensive interface for discovering, downloading, and managing Large Language Models (LLMs). It integrates with multiple sources including local Ollama (auto-managed), HuggingFace Hub, Google Vertex AI, and cloud APIs.
Supported Sources
Ollama, HuggingFace, Cloud APIs
Hardware Check
Pre-flight VRAM validation
Download Progress
Real-time tracking
Access URL
/models
Interface Layout
Local Models Tab
Browse and manage locally-hosted models via Ollama integration.
Features:
- HuggingFace search for GGUF models
- Installed Ollama inventory
- Active download progress
- Quantization selection
Example Interface:
┌─────────────────────────────────────┐
│ [Search HuggingFace...] [🔍] │
├─────────────────────────────────────┤
│ [llama3:7b] [Pull] [Available] │
│ [deepseek-r1:32b] [✓] [Active] │
│ [qwen3:72b] [▼ 45%] [Downloading]│
└─────────────────────────────────────┘
API Catalog Tab
Discover and use models from cloud API providers.
Supported Providers:
- Google (Gemini 2.5, Gemini 2.0)
- OpenAI (GPT-4o, GPT-4o-mini)
- Anthropic (Claude 3.5+)
- OpenRouter (400+ models)
Provider Status:
✓ Google (Active)
✓ OpenAI (Active)
⚠ Anthropic (No Key)
Model Details Panel
When you select a model, a detailed sidebar appears showing model specifications and hardware compatibility.
Model Info
Pre-flight Check
Available Quants
Suggested for your GPU: Q4_K_M
Quick Start Guide
Discover Models
- Navigate to
/models - Click the API Catalog tab to see cloud models
- Switch to Local Models tab for local/HuggingFace search
- Search for models like "deepseek-r1" or "llama-4"
Check Hardware Compatibility
- Select any model from the list
- Click Check VRAM button
- Review the pre-flight check results
- If insufficient VRAM, try lower quantization
Download & Use
- Click Pull Model on desired model
- Watch progress bar track download
- Once complete, go to Chat page
- Click Quick Model Swap FAB
- Select newly downloaded model
Search Tips
Recommended Models
deepseek-coder- Coding modelsqwen3.5- Qwen modelsllama-4- Latest Llamagemma-3- Lightweight
Search Filters
- Auto-filters GGUF format only
- Excludes sharded models
- Excludes vision models (mmproj)
- Shows available quantizations
Ready to Configure?
Learn how to set up local LLMs and cloud API providers with detailed configuration guides.
Documentation Menu
Model Hub Dashboard
Discover, download, and manage Large Language Models
Overview
The Model Hub provides a comprehensive interface for discovering, downloading, and managing Large Language Models (LLMs). It integrates with multiple sources including local Ollama (auto-managed), HuggingFace Hub, Google Vertex AI, and cloud APIs.
Supported Sources
Ollama, HuggingFace, Cloud APIs
Hardware Check
Pre-flight VRAM validation
Download Progress
Real-time tracking
Access URL
/models
Interface Layout
Local Models Tab
Browse and manage locally-hosted models via Ollama integration.
Features:
- HuggingFace search for GGUF models
- Installed Ollama inventory
- Active download progress
- Quantization selection
Example Interface:
┌─────────────────────────────────────┐
│ [Search HuggingFace...] [🔍] │
├─────────────────────────────────────┤
│ [llama3:7b] [Pull] [Available] │
│ [deepseek-r1:32b] [✓] [Active] │
│ [qwen3:72b] [▼ 45%] [Downloading]│
└─────────────────────────────────────┘
API Catalog Tab
Discover and use models from cloud API providers.
Supported Providers:
- Google (Gemini 2.5, Gemini 2.0)
- OpenAI (GPT-4o, GPT-4o-mini)
- Anthropic (Claude 3.5+)
- OpenRouter (400+ models)
Provider Status:
✓ Google (Active)
✓ OpenAI (Active)
⚠ Anthropic (No Key)
Model Details Panel
When you select a model, a detailed sidebar appears showing model specifications and hardware compatibility.
Model Info
Pre-flight Check
Available Quants
Suggested for your GPU: Q4_K_M
Quick Start Guide
Discover Models
- Navigate to
/models - Click the API Catalog tab to see cloud models
- Switch to Local Models tab for local/HuggingFace search
- Search for models like "deepseek-r1" or "llama-4"
Check Hardware Compatibility
- Select any model from the list
- Click Check VRAM button
- Review the pre-flight check results
- If insufficient VRAM, try lower quantization
Download & Use
- Click Pull Model on desired model
- Watch progress bar track download
- Once complete, go to Chat page
- Click Quick Model Swap FAB
- Select newly downloaded model
Search Tips
Recommended Models
deepseek-coder- Coding modelsqwen3.5- Qwen modelsllama-4- Latest Llamagemma-3- Lightweight
Search Filters
- Auto-filters GGUF format only
- Excludes sharded models
- Excludes vision models (mmproj)
- Shows available quantizations
Ready to Configure?
Learn how to set up local LLMs and cloud API providers with detailed configuration guides.