harbor
MCP ServerFreeOne command brings a complete pre-wired LLM stack with hundreds of services to explore.
Capabilities14 decomposed
docker compose-based service orchestration with dynamic configuration resolution
Medium confidenceHarbor abstracts Docker Compose through a CLI system that dynamically resolves and merges compose files based on requested services, hardware capabilities (GPU detection via has_nvidia()), and user profiles. The orchestration engine uses a 'Lego-like' modular approach where each service is a pluggable module, with the core harbor.sh script handling service lifecycle management through functions like run_up() for starting services with flags like --tail or --open. Configuration is merged via compose_with_options() which combines base compose files with service-specific overrides.
Uses dynamic compose file merging with hardware-aware profile selection (compose_with_options + has_nvidia detection) rather than static configuration, enabling single-command deployment across heterogeneous hardware without manual intervention
Simpler than Kubernetes for local AI stacks but more flexible than Docker Compose alone because it automates the 'wiring' between services (e.g., connecting UI to inference backend) based on what's actually deployed
environment variable management with profile-based configuration
Medium confidenceHarbor provides a dedicated env_manager() function in harbor.sh (lines 1257-1350) that handles get, set, and list operations for the .env file, enabling users to configure services through environment variables without editing files directly. The system supports profile-based configuration through profiles/default.env, allowing users to switch between different hardware profiles, model selections, and service configurations. Configuration changes are persisted to the .env file and automatically loaded on subsequent service starts.
Implements a dedicated env_manager() CLI function with get/set/list operations instead of requiring users to edit .env files directly, combined with profile-based configuration switching (profiles/default.env) for hardware-aware deployments
More user-friendly than raw Docker Compose environment variable management because it provides CLI commands for configuration instead of requiring file editing, and supports profile switching for different hardware setups
service dependency resolution and automatic wiring with compose file merging
Medium confidenceHarbor implements automatic service dependency resolution through its compose file merging system (compose_with_options function in harbor.sh lines 402-520). When a user requests a service, Harbor analyzes service metadata to identify required dependencies, then merges the appropriate compose files in dependency order. This ensures that if a user enables a RAG service, the required vector database and embedding model services are automatically started. The system prevents circular dependencies and validates that all required services are available before starting the stack.
Implements automatic dependency resolution through compose file merging (compose_with_options) that analyzes service metadata to identify and start required dependencies in correct order, preventing broken configurations and circular dependencies
More intelligent than manual Docker Compose because it automatically resolves and starts dependencies, and more reliable than ad-hoc service startup because it validates dependency chains before starting services
version synchronization and model management across services
Medium confidenceHarbor includes version synchronization logic (routines/models/hf.ts, routines/models/llamacpp.ts) that manages model versions across different inference backends. The system tracks which models are available in each backend (Ollama, llama.cpp, HuggingFace), handles model downloads and caching, and ensures version consistency when switching backends. Users can specify model versions through environment variables, and Harbor automatically downloads the correct version for the selected backend. The system supports model quantization variants (e.g., 4-bit, 8-bit) and automatically selects the appropriate variant based on available hardware.
Implements version synchronization and model management (routines/models/hf.ts, llamacpp.ts) that tracks model availability across backends, handles downloads and caching, and automatically selects quantization variants based on hardware
More integrated than manual model management because it automates downloads and version tracking, and more flexible than single-backend model management because it supports multiple backends with different quantization variants
observability and evaluation services for llm monitoring and testing
Medium confidenceHarbor includes observability and evaluation services that enable monitoring of LLM inference (latency, throughput, token usage) and evaluation of model outputs (quality metrics, safety checks). These services integrate with Harbor Boost to collect metrics from every LLM request, and provide dashboards and APIs for analyzing performance. The system supports custom evaluation modules that can be plugged into the request/response pipeline to assess output quality, detect hallucinations, or check for safety violations.
Provides observability and evaluation services that integrate with Harbor Boost to collect metrics from every LLM request and support custom evaluation modules for quality assessment and safety checking
More integrated than external monitoring tools because it's built into Harbor's request pipeline, and more flexible than fixed evaluation metrics because it supports custom evaluation modules
custom service creation and harbor boost module development framework
Medium confidenceHarbor provides a framework for creating custom services and Harbor Boost modules that extend the platform's capabilities. Custom services are defined as Docker Compose services with metadata declarations, while Boost modules are Python classes that hook into the LLM request/response pipeline. The framework includes templates, documentation, and integration testing utilities to help developers build and test custom extensions. Custom services are automatically discovered and integrated into the service catalog, and Boost modules can be enabled through configuration without modifying Harbor core.
Provides a framework for creating custom services (Docker Compose + metadata) and Boost modules (Python classes) that extend Harbor without forking, with automatic discovery and integration into the service catalog
More extensible than closed platforms because it provides clear extension points and templates, and more integrated than plugin systems because custom services are first-class citizens in Harbor's service model
service catalog with metadata-driven discovery and tagging
Medium confidenceHarbor maintains a curated service catalog (app/src/serviceMetadata.ts lines 8-103) with over 50 AI-related services organized by Harbor Service Tags (HST). Each service has associated metadata including category (LLM backends, frontends, satellite services, RAG tools), dependencies, port mappings, and integration patterns. The catalog enables users to discover available services, understand their purpose, and understand how they integrate with other services in the stack. Service metadata drives the dynamic composition of Docker Compose files and the Harbor Desktop App's UI.
Implements a declarative service catalog (serviceMetadata.ts) with Harbor Service Tags (HST) for categorization, enabling metadata-driven service discovery and composition rather than requiring users to manually understand service relationships
More discoverable than raw Docker Compose because services are tagged and categorized with explicit metadata, making it easier for users to find and understand available services without reading documentation
harbor boost — llm proxy with python-based module system for optimization
Medium confidenceHarbor Boost is an optimizing LLM proxy layer (services/boost/pyproject.toml) built with a Python-based module system that intercepts LLM requests and applies transformations such as prompt optimization, response caching, cost tracking, and multi-provider routing. The module system allows users to create custom Boost modules that hook into the request/response pipeline. Boost acts as a middleware between client applications and inference backends (Ollama, llama.cpp, OpenAI), enabling advanced features like artifact generation and visualization without modifying the underlying models.
Implements a Python-based module system for LLM request/response transformation that allows users to create custom optimization logic (caching, routing, artifact generation) without modifying Harbor core or client applications
More flexible than static LLM proxies because the module system enables custom transformations, and more lightweight than full LLM orchestration frameworks because it focuses specifically on request/response optimization
hardware-aware service composition with gpu detection and fallback profiles
Medium confidenceHarbor includes hardware detection logic (has_nvidia() function in harbor.sh lines 262-264) that automatically detects NVIDIA GPUs and applies GPU-accelerated compose files when available. The system maintains separate compose file variants for CPU-only and GPU-accelerated deployments, allowing the same service configuration to run optimally on different hardware without user intervention. If GPU detection fails or GPUs are unavailable, Harbor automatically falls back to CPU-only compose files, ensuring services remain functional across heterogeneous hardware.
Implements automatic GPU detection (has_nvidia()) with compose file variants that enable the same service configuration to run optimally on CPU or GPU hardware without user intervention or manual profile switching
More user-friendly than Docker Compose alone because it automatically selects the right hardware profile, and more flexible than cloud-only solutions because it supports heterogeneous on-premises hardware
mcp (model context protocol) service integration and tool-use orchestration
Medium confidenceHarbor integrates with the Model Context Protocol (MCP) to enable LLM agents to call external tools and services as function calls. Services in Harbor expose MCP-compatible tool schemas that agents can discover and invoke. The integration allows agents to use tools like web search (SearXNG), code execution, file operations, and custom service APIs without hardcoding tool logic into the agent. Tool schemas are defined in service metadata and automatically registered with the MCP server when services start.
Integrates services with Model Context Protocol (MCP) to enable LLM agents to discover and call Harbor services as tools, with schemas defined in service metadata rather than hardcoded in agent logic
More flexible than hardcoded tool integration because MCP schemas are declarative and discoverable, and more standardized than custom tool APIs because it uses the MCP protocol supported by multiple LLM frameworks
harbor desktop app — tauri-based gui for service management and monitoring
Medium confidenceHarbor includes a Tauri-based desktop application (app/src-tauri/tauri.conf.json) that provides a graphical interface for managing services, viewing service status, configuring settings, and monitoring logs. The app is built with Tauri (Rust backend, TypeScript/React frontend) for cross-platform compatibility and minimal resource overhead. The app communicates with the Harbor CLI and Docker daemon to provide real-time service status, enable one-click service start/stop, and visualize service dependencies and health.
Provides a Tauri-based desktop application (TypeScript/React frontend, Rust backend) that mirrors CLI functionality with visual service management, real-time status monitoring, and configuration UI without requiring CLI knowledge
More lightweight than web-based dashboards because Tauri uses native OS windows, and more integrated than separate monitoring tools because it's built specifically for Harbor's service model
multi-backend llm inference with ollama, llama.cpp, and cloud provider support
Medium confidenceHarbor supports multiple LLM inference backends including Ollama (local model serving), llama.cpp (CPU-optimized inference), and cloud providers (OpenAI, Anthropic) through a unified interface. Each backend is a pluggable service with its own compose file and configuration. Harbor's LiteLLM Gateway (satellite service) provides a unified API endpoint that routes requests to the selected backend, allowing applications to switch backends without code changes. Model selection and backend routing are configured through environment variables and Harbor Boost modules.
Provides pluggable LLM backend services (Ollama, llama.cpp, cloud providers) with unified API routing through LiteLLM Gateway, enabling backend switching through environment variables and Harbor Boost modules without application code changes
More flexible than single-backend solutions because it supports local and cloud inference with unified routing, and more integrated than separate inference services because backends are pre-configured and automatically wired together
rag (retrieval-augmented generation) service integration with knowledge base management
Medium confidenceHarbor includes satellite services for RAG workflows, including vector databases, document indexers, and retrieval engines. Services like SearXNG provide web search integration for RAG, while knowledge base services enable semantic search over local documents. Harbor's service composition automatically wires RAG services together (e.g., connecting a document indexer to a vector database), and Harbor Boost modules can intercept LLM requests to augment prompts with retrieved context. The system supports multiple RAG patterns including semantic search, hybrid search, and multi-hop retrieval.
Integrates RAG services (vector databases, document indexers, web search via SearXNG) with automatic service wiring and Harbor Boost module hooks for prompt augmentation, enabling end-to-end RAG without custom integration code
More integrated than standalone RAG libraries because services are pre-configured and automatically connected, and more flexible than cloud RAG APIs because it supports local-only deployments and custom retrieval logic
web ui frontends (open webui, comfyui) with unified service routing
Medium confidenceHarbor includes pre-configured web UI frontends including Open WebUI (chat interface for LLMs) and ComfyUI (node-based workflow builder for image generation). These frontends are automatically wired to the appropriate inference backends through Harbor's service composition. Open WebUI connects to LLM backends (Ollama, llama.cpp, cloud providers) through the LiteLLM Gateway, while ComfyUI connects to image generation models. The frontends are accessible through a unified landing page service that provides navigation and service discovery.
Provides pre-configured Open WebUI and ComfyUI frontends that automatically route to Harbor's inference backends through LiteLLM Gateway, eliminating manual UI-to-backend wiring and providing unified access through a landing page service
More integrated than standalone UI projects because frontends are pre-wired to Harbor backends, and more user-friendly than CLI-only access because it provides visual interfaces for non-technical users
Capabilities are decomposed by AI analysis. Each maps to specific user intents and improves with match feedback.
Related Artifactssharing capabilities
Artifacts that share capabilities with harbor, ranked by overlap. Discovered automatically through the match graph.
openapi-servers
OpenAPI Tool Servers
stable-diffusion-webui-docker
Easy Docker setup for Stable Diffusion with user-friendly UI
Minima
** - Local RAG (on-premises) with MCP server.
ai-goofish-monitor
基于 Playwright 和AI实现的闲鱼多任务实时/定时监控与智能分析系统,配备了功能完善的后台管理UI。帮助用户从闲鱼海量商品中,找到心仪产品。
TaskingAI
The open source platform for AI-native application development.
OpenAgents
[COLM 2024] OpenAgents: An Open Platform for Language Agents in the Wild
Best For
- ✓solo developers building local LLM infrastructure
- ✓teams deploying self-hosted AI stacks without cloud dependencies
- ✓researchers prototyping multi-service AI systems on heterogeneous hardware
- ✓developers managing multiple Harbor deployments with different configurations
- ✓teams sharing Harbor setups with environment-specific overrides
- ✓users without direct file system access who need to configure services
- ✓developers building complex multi-service AI stacks
- ✓teams managing Harbor deployments with many interdependent services
Known Limitations
- ⚠Abstraction adds complexity — debugging requires understanding both Harbor's resolution logic and underlying Docker Compose
- ⚠Hardware detection is limited to NVIDIA GPUs; AMD/Intel GPU support requires manual compose file configuration
- ⚠Service interdependencies are implicit in compose files — circular dependencies or missing services can cause silent failures
- ⚠Profile system is file-based (profiles/default.env) — no built-in validation of configuration values before service startup
- ⚠Environment variable changes require service restart to take effect — no hot-reload capability
- ⚠No conflict detection when multiple profiles define overlapping variables
Requirements
Input / Output
UnfragileRank
UnfragileRank is computed from adoption signals, documentation quality, ecosystem connectivity, match graph feedback, and freshness. No artifact can pay for a higher rank.
Repository Details
Last commit: Apr 16, 2026
About
One command brings a complete pre-wired LLM stack with hundreds of services to explore.
Categories
Alternatives to harbor
Are you the builder of harbor?
Claim this artifact to get a verified badge, access match analytics, see which intents users search for, and manage your listing.
Get the weekly brief
New tools, rising stars, and what's actually worth your time. No spam.
Data Sources
Looking for something else?
Search →