local-deep-research
BenchmarkFreeLocal Deep Research achieves ~95% on SimpleQA benchmark (tested with GPT-4.1-mini). Supports local and cloud LLMs (Ollama, Google, Anthropic, ...). Searches 10+ sources - arXiv, PubMed, web, and your private documents. Everything Local & Encrypted.
Capabilities16 decomposed
multi-source iterative research with llm-driven query refinement
Medium confidenceExecutes deep, multi-turn research workflows that iteratively refine queries based on LLM analysis of intermediate results. The system searches 10+ sources (arXiv, PubMed, web via Brave/SearXNG, private documents) in a coordinated loop, with each iteration using LLM reasoning to identify gaps and reformulate queries. Research execution is managed through a service-oriented architecture with thread-safe settings context, enabling parallel research tasks while maintaining isolation per user and per research session.
Implements LLM-driven query refinement loop where each research iteration analyzes gaps in current results and reformulates queries, rather than executing a static search plan. This is coordinated through a Research Service that manages execution lifecycle with thread-safe context management, enabling concurrent research tasks with per-user isolation via SQLCipher encrypted databases.
Outperforms single-pass research tools (Perplexity, traditional RAG) by iteratively deepening search based on LLM reasoning about gaps, achieving ~95% accuracy on SimpleQA benchmark while maintaining full local deployment and encryption for sensitive research.
per-user encrypted database with pbkdf2-derived key derivation
Medium confidenceProvides per-user data isolation through SQLCipher databases encrypted with AES-256-CBC, where each user's password is derived via PBKDF2-HMAC-SHA512 with 256,000 iterations and a per-user random salt. The database architecture separates user data (research history, collections, settings) from system configuration, with automatic encryption key management and password-based access control. Database encryption check utilities verify SQLCipher compatibility at startup.
Uses PBKDF2-HMAC-SHA512 with 256,000 iterations and per-user random salt to derive encryption keys directly from user passwords, eliminating the need for external key management systems. This approach is implemented through database/encryption_check.py and database/sqlcipher_compat.py modules that verify SQLCipher availability and handle key derivation transparently.
Provides stronger per-user isolation than application-level encryption (which shares keys) and simpler deployment than external key management (no KMS infrastructure needed), while maintaining NIST-compliant key derivation parameters.
flask web application with real-time research ui and result streaming
Medium confidenceProvides a web-based user interface built with Flask backend and modern frontend (likely React or Vue.js based on build system references). The web UI enables real-time research execution with streaming result updates, research history management, and collection/library organization. Frontend communicates with Flask backend via REST API, with WebSocket support for real-time status updates during long-running research.
Implements Flask web application with real-time research UI that streams results as they are discovered, rather than waiting for complete research execution. Frontend build system enables modern JavaScript framework integration with hot reloading for development.
More interactive than CLI tools by providing real-time progress visualization and result streaming, while maintaining same encryption and per-user isolation as backend.
thread-safe settings and context management for concurrent research execution
Medium confidenceImplements thread-safe settings management through context variables that enable concurrent research tasks to maintain isolated configuration and state. Each research execution gets its own context (LLM provider, search sources, user credentials) that is thread-local, preventing cross-contamination between concurrent requests. Settings are loaded from environment variables and configuration files with runtime override capability.
Implements thread-safe settings through Python contextvars, enabling each research execution to maintain isolated configuration without global state. This allows concurrent research tasks with different LLM providers or search sources to execute simultaneously.
More robust than global configuration variables by preventing cross-contamination between concurrent requests, while simpler than request-scoped dependency injection frameworks.
benchmarking system with simpleqa evaluation and accuracy metrics
Medium confidenceIncludes built-in benchmarking infrastructure that evaluates research quality against the SimpleQA benchmark, measuring accuracy, citation correctness, and source attribution. The benchmarking system executes research on benchmark queries, compares results against ground truth, and generates accuracy reports. This enables quantitative evaluation of research quality across different LLM providers and configurations.
Includes built-in benchmarking against SimpleQA with ~95% accuracy achieved with GPT-4.1-mini, enabling quantitative evaluation of research quality. Benchmarking system generates detailed accuracy reports comparing citation correctness and source attribution.
More comprehensive than manual testing by providing automated benchmarking against standardized dataset, while enabling comparison across LLM providers and configurations.
document download and management with automatic metadata extraction
Medium confidenceAutomatically downloads and manages research documents (PDFs, web pages) discovered during research, with automatic metadata extraction (title, authors, publication date). Downloaded documents are stored in encrypted database with full-text indexing for later search. Metadata extraction uses heuristics and optional OCR for PDFs, enabling documents to be cited and referenced in future research.
Automatically downloads and indexes research documents discovered during research, with automatic metadata extraction and storage in encrypted database. Downloaded documents are indexed for full-text search in future research.
More integrated than manual document management by automatically downloading and indexing documents discovered during research, while maintaining encryption and per-user isolation.
news and subscription management for continuous research updates
Medium confidenceEnables subscription to research topics with automatic periodic research execution and result delivery. The system maintains topic subscriptions in encrypted database, executes research on subscribed topics at configured intervals (daily, weekly, monthly), and delivers results via email or web UI notifications. Subscription management includes filtering, deduplication, and archival of subscription results.
Implements subscription system that automatically executes research on topics at configured intervals and delivers results via email or web UI. Subscription results are stored in encrypted database with deduplication and filtering.
More integrated than external alert services (Google Alerts, Feedly) by using same research engine and maintaining results in encrypted database for historical analysis.
report generation and export in multiple formats
Medium confidenceGenerates research reports from research results with support for multiple export formats (markdown, HTML, PDF, JSON). Report generation includes automatic formatting, citation insertion, table of contents generation, and optional styling. Exported reports can be shared externally while maintaining citation metadata for verification.
Generates research reports in multiple formats (markdown, HTML, PDF, JSON) with automatic citation insertion and formatting. Report generation is integrated into research workflow, enabling one-click export.
More integrated than external report generators by supporting multiple formats natively and maintaining citation metadata throughout export process.
multi-provider llm abstraction with unified interface
Medium confidenceAbstracts multiple LLM providers (OpenAI, Anthropic, Google, Mistral, Ollama) behind a unified Python interface, enabling runtime provider switching without code changes. Configuration is managed through environment variables and thread-safe settings context, with provider-specific parameters (temperature, max_tokens, system prompts) normalized across APIs. The abstraction handles provider-specific response formats, streaming behavior, and error handling transparently.
Implements provider abstraction through thread-safe settings context that enables runtime provider switching without code changes. Configuration is centralized in LLM Provider Configuration system with environment variable overrides, allowing different research tasks to use different providers simultaneously while maintaining consistent API surface.
More flexible than LangChain's provider abstraction by supporting local Ollama as first-class citizen and enabling per-task provider selection, while simpler than building custom provider wrappers for each LLM API.
search engine integration layer with 10+ source coordination
Medium confidenceCoordinates searches across 10+ sources (arXiv, PubMed, Brave web search, SearXNG, Google Scholar, private document collections) through a unified search interface. The Search Engine Integration Layer abstracts provider-specific APIs, query syntax, and result formats into a common result schema with source attribution. Search execution is parallelized where possible, with configurable timeouts and fallback behavior when sources are unavailable.
Implements unified search interface that abstracts 10+ heterogeneous sources (academic APIs, web search, private RAG) with source-specific query translation and result normalization. Search execution is parallelized through async/await patterns with configurable per-source timeouts, enabling fast fallback when sources are slow or unavailable.
Broader source coverage than single-provider search (Brave, Google) by combining academic (arXiv, PubMed), web (Brave, SearXNG), and private document sources in unified interface, while maintaining local deployment option via self-hosted SearXNG.
rag-based private document indexing and retrieval
Medium confidenceIndexes private documents (PDFs, markdown, text) into vector embeddings using local or cloud embedding models, enabling semantic search across document collections. Documents are stored in per-user encrypted databases with metadata (source, date, collection), and retrieval uses cosine similarity search on embeddings to find relevant passages. The RAG system integrates with the research workflow to supplement search results from public sources with organization-specific knowledge.
Implements RAG system with per-user encrypted storage of documents and embeddings, enabling private document search without external vector databases. Document indexing is integrated into research workflow, allowing seamless combination of public source results with private document retrieval in single research execution.
Simpler deployment than external vector databases (Pinecone, Weaviate) by storing embeddings in encrypted SQLCipher, while maintaining semantic search capability through local or cloud embedding models.
citation tracking and source attribution with evidence chains
Medium confidenceAutomatically tracks citations throughout the research process, maintaining evidence chains that link claims in the final report back to original sources. The citation handler (citation_handler.py) extracts source metadata (URL, publication date, authors) from search results and embeds citations in generated content. Reports can be exported with full citation metadata in multiple formats (markdown with footnotes, HTML with hyperlinks, JSON with source provenance).
Implements citation tracking through evidence chains that link claims in generated reports back to original sources, with support for multiple export formats. Citation handler maintains source metadata throughout research execution and generates formatted citations in markdown, HTML, and JSON formats.
More comprehensive than simple URL citations by tracking full evidence chains and supporting multiple citation formats, while maintaining source metadata in encrypted database for audit trails.
rest api with async request handling and long-running research tasks
Medium confidenceExposes research capabilities through a REST API built on Flask with async task execution for long-running research workflows. API endpoints support both synchronous queries (for quick searches) and asynchronous research execution (for deep research that may take 30+ seconds). Task status can be polled via job IDs, with results cached in encrypted database. API authentication uses per-user tokens derived from encrypted credentials.
Implements REST API with async task execution for long-running research, allowing clients to submit research requests and poll results without blocking. API authentication uses per-user encrypted credentials, and task results are cached in encrypted database for audit trails.
Simpler than building custom API wrappers by providing built-in async task handling and per-user authentication, while maintaining full encryption of cached results.
python programmatic api (ldrclient) for direct integration
Medium confidenceProvides a Python client library (LDRClient) that enables direct programmatic access to research capabilities without REST API overhead. The client handles user authentication, database access, and research execution through a simple Python interface. Supports both synchronous and asynchronous research execution, with context managers for automatic resource cleanup.
Provides Python client library with both sync and async interfaces, enabling direct library access without REST API overhead. Client handles authentication and database access transparently, with context managers for resource cleanup.
Lower latency than REST API for Python applications by eliminating HTTP overhead, while maintaining same encryption and per-user isolation guarantees.
cli tools (ldr, ldr-web) for command-line research execution
Medium confidenceProvides command-line interfaces for research execution and web server management. The `ldr` command enables direct research queries from shell scripts, while `ldr-web` manages the Flask web application lifecycle. CLI tools support configuration via environment variables and command-line flags, with output formatting options (JSON, markdown, plain text).
Provides both research execution (`ldr`) and deployment management (`ldr-web`) CLI tools, enabling shell script integration and CI/CD automation. CLI tools support multiple output formats and configuration via environment variables.
Simpler than building custom shell wrappers around Python API by providing native CLI with built-in formatting and error handling.
mcp server (ldr-mcp) for claude desktop and ai assistant integration
Medium confidenceImplements Model Context Protocol (MCP) server that enables integration with Claude Desktop and other AI assistants. The MCP server exposes research capabilities as tools that Claude can invoke directly, with automatic result formatting and context injection. This enables AI assistants to perform research as part of their reasoning process without external API calls.
Implements MCP server that exposes research as native tools for Claude Desktop, enabling AI assistants to invoke research as part of their reasoning without external API integration. Results are automatically formatted for context injection.
Tighter integration than REST API by using MCP protocol native to Claude, enabling research invocation as part of assistant reasoning rather than external tool calls.
Capabilities are decomposed by AI analysis. Each maps to specific user intents and improves with match feedback.
Related Artifactssharing capabilities
Artifacts that share capabilities with local-deep-research, ranked by overlap. Discovered automatically through the match graph.
gpt-researcher
An autonomous agent that conducts deep research on any data using any LLM providers
GPT Researcher
Autonomous agent for comprehensive research reports.
gpt-researcher
An autonomous agent that conducts deep research on any data using any LLM providers
onyx
Open Source AI Platform - AI Chat with advanced features that works with every LLM
py-gpt
Desktop AI Assistant powered by GPT-5, GPT-4, o1, o3, Gemini, Claude, Ollama, DeepSeek, Perplexity, Grok, Bielik, chat, vision, voice, RAG, image and video generation, agents, tools, MCP, plugins, speech synthesis and recognition, web search, memory, presets, assistants,and more. Linux, Windows, Mac
atlas-mcp-server
A Model Context Protocol (MCP) server for ATLAS, a Neo4j-powered task management system for LLM Agents - implementing a three-tier architecture (Projects, Tasks, Knowledge) to manage complex workflows. Now with Deep Research.
Best For
- ✓researchers and academics building automated literature review pipelines
- ✓teams deploying privacy-critical research infrastructure on-premise
- ✓developers integrating deep research capabilities into larger AI agent systems
- ✓teams deploying on shared infrastructure or untrusted cloud environments
- ✓organizations with data residency or encryption-at-rest compliance requirements
- ✓developers building multi-tenant research platforms with strong privacy guarantees
- ✓non-technical users who need research capabilities without CLI
- ✓teams collaborating on research with shared web interface
Known Limitations
- ⚠Research execution latency scales with number of sources and LLM response time; typical multi-turn research takes 30-120 seconds depending on query complexity
- ⚠Query refinement relies on LLM reasoning quality; weaker models may produce suboptimal follow-up queries
- ⚠Private document search requires pre-indexing via RAG pipeline; real-time document addition has ~5-10 second indexing latency per document
- ⚠No built-in deduplication across sources; duplicate results may appear in final research output
- ⚠SQLCipher compilation required on bare metal; Docker images include pre-compiled binaries but custom builds need SQLCipher development headers
- ⚠Key derivation with 256,000 PBKDF2 iterations adds ~100-200ms latency to database initialization per user session
Requirements
Input / Output
UnfragileRank
UnfragileRank is computed from adoption signals, documentation quality, ecosystem connectivity, match graph feedback, and freshness. No artifact can pay for a higher rank.
Repository Details
Last commit: Apr 22, 2026
About
Local Deep Research achieves ~95% on SimpleQA benchmark (tested with GPT-4.1-mini). Supports local and cloud LLMs (Ollama, Google, Anthropic, ...). Searches 10+ sources - arXiv, PubMed, web, and your private documents. Everything Local & Encrypted.
Categories
Alternatives to local-deep-research
Are you the builder of local-deep-research?
Claim this artifact to get a verified badge, access match analytics, see which intents users search for, and manage your listing.
Get the weekly brief
New tools, rising stars, and what's actually worth your time. No spam.
Data Sources
Looking for something else?
Search →