awesome-llm-apps
AgentFree100+ AI Agent & RAG apps you can actually run — clone, customize, ship.
Capabilities13 decomposed
multi-framework agent scaffolding with framework-agnostic patterns
Medium confidenceProvides 100+ production-ready agent implementations across three primary frameworks (Agno, LangChain/LangGraph, and native Python) organized by complexity tier (starter, advanced single-agent, multi-agent). Each implementation includes complete dependency specifications, environment configuration templates, and runnable entry points, allowing developers to clone and immediately execute agents without framework-specific boilerplate. The repository uses a tiered complexity model where starter agents demonstrate basic tool-calling patterns, advanced agents implement planner-executor architectures with state management, and multi-agent systems showcase coordination via message passing or shared context.
Organizes 100+ implementations across three distinct frameworks (Agno, LangChain/LangGraph, native) with explicit complexity tiers (starter/advanced/expert) and domain-specific examples (finance, travel, research), enabling side-by-side framework comparison and progressive learning paths. Most agent repositories focus on a single framework; this one treats framework diversity as a feature.
Broader framework coverage and clearer complexity progression than single-framework tutorials; more production-focused than academic agent papers but less opinionated than framework-specific docs
retrieval-augmented generation (rag) pattern library with multiple retrieval strategies
Medium confidenceImplements 8+ distinct RAG architectures (basic retrieval, corrective RAG, hybrid retrieval, database routing, agentic RAG, autonomous RAG, RAG with reasoning) with working code for each pattern. Each implementation demonstrates a specific retrieval strategy: basic RAG uses vector similarity search, corrective RAG adds a grading step to filter irrelevant chunks, hybrid RAG combines vector and keyword search, database routing uses an LLM to select which database to query, and agentic RAG treats retrieval as a tool the agent can invoke iteratively. Implementations support multiple vector databases (Pinecone, Weaviate, Chroma, FAISS) and document sources (PDFs, web pages, databases, code repositories).
Provides 8+ distinct RAG patterns (basic, corrective, hybrid, database routing, agentic, autonomous, reasoning-enhanced) with working implementations for each, allowing developers to compare trade-offs between retrieval quality and latency. Most RAG tutorials show only basic vector search; this library treats RAG as a design space with multiple valid solutions.
More comprehensive RAG pattern coverage than LangChain's built-in RAG examples; more practical than academic RAG papers with runnable code for each pattern
investment and finance agent with real-time market data integration
Medium confidenceImplements specialized agents for financial analysis and investment decisions that integrate real-time market data, financial APIs, and domain-specific reasoning. The investment agent can fetch stock prices, analyze financial statements, calculate metrics (P/E ratio, dividend yield), and provide investment recommendations. Integration with financial data providers (Alpha Vantage, Finnhub, or similar) enables real-time market data access. The agent uses domain-specific prompts and reasoning patterns for financial analysis, handles numerical precision and currency conversions, and provides citations to data sources. Examples include portfolio analysis agents, stock recommendation agents, and market trend analysis agents.
Provides investment agent implementations with real-time market data integration, financial metric calculations, and domain-specific reasoning patterns. Demonstrates how to handle numerical precision, currency conversions, and financial data sources. Most agent tutorials are generic; this library includes domain-specific agents for finance.
More specialized than generic agents but less comprehensive than dedicated financial analysis platforms; useful for prototyping financial agents
web scraping agent with browser automation and dynamic content handling
Medium confidenceImplements agents that can browse the web, scrape content, and extract information from dynamic websites using browser automation (Selenium, Playwright, or Puppeteer). The web scraping agent can navigate websites, interact with forms and buttons, wait for dynamic content to load, and extract structured data. Integration with agent frameworks allows the agent to decide what to scrape, how to navigate, and how to extract information based on user requests. Examples include competitive intelligence agents that scrape competitor websites, price monitoring agents that track product prices, and content aggregation agents that gather information from multiple sources. The agent handles JavaScript-heavy sites and can wait for content to load before extraction.
Provides web scraping agent implementations with browser automation, dynamic content handling, and integration with agent frameworks. Demonstrates how agents can decide what to scrape and how to navigate websites. Most agent tutorials don't include web scraping; this library treats it as a legitimate agent capability with appropriate caveats.
More practical than generic scraping tutorials; enables agent-driven scraping but with significant latency and resource trade-offs vs direct HTTP scraping
corrective and hybrid rag with relevance grading and multi-strategy retrieval
Medium confidenceImplements advanced RAG patterns that improve retrieval quality beyond basic vector similarity search. Corrective RAG adds a grading step where an LLM evaluates whether retrieved documents are relevant to the query; if not, the system reformulates the query and retrieves again. Hybrid RAG combines multiple retrieval strategies (vector similarity, keyword search, semantic search) and ranks results by combining scores from different methods. Implementations demonstrate how to define relevance criteria, implement grading logic, and combine retrieval scores. The corrective approach trades latency for quality (additional LLM calls), while hybrid approaches balance different retrieval strengths.
Provides implementations of corrective RAG (with relevance grading and query reformulation) and hybrid RAG (combining vector and keyword search) with explicit trade-offs between quality and latency. Demonstrates how to define and implement relevance criteria. Most RAG tutorials show only basic vector search; this library treats quality improvement as a design pattern.
More sophisticated than basic RAG but with documented latency costs; more practical than academic RAG papers with working code
model context protocol (mcp) agent integration with multi-provider tool binding
Medium confidenceDemonstrates MCP protocol integration for agents that need to interact with external systems (GitHub, Notion, browsers, file systems) through standardized tool schemas. Implementations show how to define MCP tool specifications (input schemas, descriptions), bind them to agent frameworks (Agno, LangChain), and handle tool execution with error recovery. The repository includes examples of travel planning agents using MCP for flight/hotel APIs, GitHub agents using MCP for repository operations, and browser automation agents using MCP for web scraping, all following the MCP specification for tool discovery and invocation.
Provides working MCP implementations for diverse use cases (travel planning, GitHub operations, browser automation, Notion integration) with explicit tool schema definitions and error handling patterns. Demonstrates how MCP standardizes tool discovery and invocation across different external systems, reducing boilerplate compared to custom API wrappers.
More comprehensive MCP examples than official MCP documentation; more standardized than custom tool-calling implementations but less mature than framework-specific tool ecosystems
multi-agent coordination with message passing and shared context
Medium confidenceImplements multi-agent systems where specialized agents (e.g., SEO auditor, content writer, technical reviewer) coordinate via message passing or shared state to solve complex tasks. Examples include an SEO audit team where one agent crawls websites, another analyzes content, and a third generates recommendations; a home renovation agent where one agent gathers requirements, another estimates costs, and a third creates project plans. Coordination patterns include sequential task handoff (agent A completes, passes results to agent B), parallel execution with result aggregation, and hierarchical delegation (manager agent assigns tasks to worker agents). Implementations use either explicit message queues or shared context objects to pass information between agents.
Provides concrete multi-agent examples (SEO audit team, home renovation agent) with explicit coordination patterns (message passing, shared context, hierarchical delegation) and implementation code. Most agent tutorials focus on single agents; this library treats multi-agent coordination as a first-class pattern with multiple architectural approaches.
More practical multi-agent examples than academic papers; more detailed than framework docs but less opinionated than specialized multi-agent frameworks like AutoGen
research agent with iterative planning and web search integration
Medium confidenceImplements research agents that decompose complex research queries into sub-questions, search the web for relevant information, synthesize findings, and iteratively refine results. The research agent uses a planner-executor pattern: a planner LLM breaks down 'research X' into specific search queries, an executor searches the web and retrieves documents, and a synthesizer combines results into a coherent report. Integration with Google Gemini Interactions API enables real-time web search within agent reasoning loops. The agent can iterate — if initial results are insufficient, it generates follow-up queries and searches again. Outputs include structured research reports with source citations and confidence scores.
Combines planner-executor-synthesizer architecture with iterative refinement and real-time web search via Gemini Interactions API, enabling agents to conduct research beyond their training data. Most research agents use static RAG; this implementation treats web search as a first-class agent capability with iterative improvement.
More sophisticated than basic web search agents; tightly integrated with Gemini's native search capabilities but less portable than framework-agnostic approaches
voice agent with speech-to-text and text-to-speech synthesis
Medium confidenceImplements voice-based agents that accept audio input, transcribe it to text, process through an LLM agent, and synthesize responses back to speech. The voice agent pipeline uses a speech-to-text service (e.g., Google Speech-to-Text, Deepgram) to convert audio to text, passes the text to an agent for processing, and uses a text-to-speech service (e.g., Google TTS, ElevenLabs) to convert the agent's response back to audio. Implementations handle audio streaming, real-time transcription, and low-latency synthesis. Examples include voice-based travel planners, customer service agents, and accessibility-focused applications.
Provides end-to-end voice agent implementations with explicit handling of audio streaming, transcription, agent processing, and synthesis. Demonstrates integration with multiple speech services (Google, Deepgram, ElevenLabs) and latency optimization patterns. Most agent tutorials are text-only; this library treats voice as a first-class interaction modality.
More complete voice agent examples than framework docs; more practical than academic speech processing papers but less specialized than dedicated voice AI platforms
persistent conversation memory with context management
Medium confidenceImplements agents with persistent conversation history and context management, allowing multi-turn interactions where the agent remembers previous exchanges and maintains coherent context. Patterns include simple conversation history (storing all messages), summarization-based memory (periodically summarizing old messages to save tokens), entity-based memory (tracking important entities and their attributes), and hybrid approaches combining multiple memory strategies. Implementations use local storage (SQLite, JSON files) or external services (Redis, Supabase) for persistence. The agent can retrieve relevant context from history, update memory as new information emerges, and manage context window size to stay within LLM token limits.
Provides multiple memory strategies (simple history, summarization, entity-based, hybrid) with working implementations and storage backends (SQLite, Redis, Supabase). Demonstrates explicit token management and context window optimization. Most agent tutorials assume stateless interactions; this library treats persistent memory as essential for real-world agents.
More comprehensive memory patterns than framework defaults; more practical than academic memory papers but less specialized than dedicated memory systems like Mem0
domain-specific agent templates for specialized data sources
Medium confidenceProvides pre-built agent templates for interacting with specific data sources: GitHub agents for repository analysis and code search, PDF chat agents for document Q&A, YouTube transcript agents for video content analysis, and similar domain-specific implementations. Each template includes data source connectors (GitHub API client, PDF parser, YouTube API), specialized prompts for the domain, and example use cases. The GitHub agent can search repositories, analyze code, and answer questions about codebases; the PDF agent can extract text, handle multi-page documents, and cite specific pages; the YouTube agent can fetch transcripts and summarize video content. Templates are designed to be cloned and customized for specific domains.
Provides ready-to-use agent templates for specific data sources (GitHub, PDF, YouTube) with data connectors, domain-specific prompts, and example use cases. Treats domain-specific agents as a pattern worth standardizing rather than requiring custom implementation for each source.
More practical than generic agent tutorials; more specialized than framework docs but less comprehensive than dedicated tools for each domain
local llm agent execution with ollama and deepseek integration
Medium confidenceDemonstrates running agents entirely locally using open-source LLMs (Deepseek, Mistral, Llama) via Ollama, eliminating dependency on cloud LLM APIs. Implementations show how to configure Agno or LangChain agents to use local Ollama endpoints, handle model-specific prompt formatting, and manage local inference latency. Examples include local RAG agents (combining local LLM with local vector database like FAISS), local research agents (using local search or document retrieval), and local multi-agent systems. The local approach trades cloud API costs for local compute resources and enables offline operation.
Provides complete local agent implementations (RAG, research, multi-agent) using Ollama and open-source models, with explicit latency and quality trade-offs documented. Demonstrates how to configure agents for local inference and handle model-specific prompt formatting. Most agent tutorials assume cloud APIs; this library treats local execution as a viable alternative with specific use cases.
More practical local agent examples than Ollama docs; enables privacy and cost optimization but with quality/latency trade-offs vs cloud APIs
streamlit ui generation for agent visualization and interaction
Medium confidenceProvides Streamlit-based UI templates for visualizing agent execution, displaying reasoning steps, and enabling user interaction with agents. Implementations show how to build agent dashboards that display agent state, tool calls, and reasoning traces in real-time. Streamlit integration allows rapid UI prototyping without frontend development — agents can be wrapped with a Streamlit app that handles user input, displays agent responses, and visualizes execution flow. Examples include research agent dashboards showing search queries and results, multi-agent system dashboards showing agent coordination, and RAG dashboards showing retrieved documents and relevance scores.
Provides Streamlit templates for agent visualization and interaction, enabling rapid UI prototyping without frontend development. Demonstrates how to display agent reasoning, tool calls, and execution traces in real-time. Most agent tutorials focus on backend logic; this library treats UI as an important part of the agent experience.
Faster to prototype than custom web frameworks; more limited than production web frameworks but sufficient for demos and internal tools
Capabilities are decomposed by AI analysis. Each maps to specific user intents and improves with match feedback.
Related Artifactssharing capabilities
Artifacts that share capabilities with awesome-llm-apps, ranked by overlap. Discovered automatically through the match graph.
star the repo
to get notified when new templates ship.**
FinRobot
FinRobot: An Open-Source AI Agent Platform for Financial Analysis using LLMs 🚀 🚀 🚀
AgenticRAG-Survey
Agentic-RAG explores advanced Retrieval-Augmented Generation systems enhanced with AI LLM agents.
GenerativeAIExamples
Generative AI reference workflows optimized for accelerated infrastructure and microservice architecture.
500-AI-Agents-Projects
The 500 AI Agents Projects is a curated collection of AI agent use cases across various industries. It showcases practical applications and provides links to open-source projects for implementation, illustrating how AI agents are transforming sectors such as healthcare, finance, education, retail, a
awesome-generative-ai-guide
A one stop repository for generative AI research updates, interview resources, notebooks and much more!
Best For
- ✓developers new to LLM agents seeking reference implementations
- ✓teams evaluating multiple agent frameworks before committing to one
- ✓builders prototyping domain-specific agents (finance, travel, research) with minimal setup time
- ✓teams building knowledge-base Q&A systems or document search applications
- ✓developers optimizing retrieval quality beyond basic vector similarity
- ✓builders integrating RAG into existing LLM applications
- ✓fintech teams building investment analysis tools
- ✓developers creating financial advisory agents
Known Limitations
- ⚠No unified abstraction layer — switching between frameworks requires rewriting agent logic, not just swapping imports
- ⚠Examples assume familiarity with Python async/await patterns and basic LLM concepts; minimal pedagogical scaffolding for absolute beginners
- ⚠Framework versions in examples may drift from latest releases; requires manual dependency updates for production use
- ⚠No built-in testing harness or evaluation framework — quality assurance is left to the implementer
- ⚠Vector database setup (Pinecone, Weaviate) requires external service provisioning; FAISS examples are local-only and don't scale to millions of documents
- ⚠Chunk size, overlap, and embedding model choices are hardcoded in examples; no adaptive chunking or dynamic embedding selection
Requirements
Input / Output
UnfragileRank
UnfragileRank is computed from adoption signals, documentation quality, ecosystem connectivity, match graph feedback, and freshness. No artifact can pay for a higher rank.
Repository Details
Last commit: Apr 19, 2026
About
100+ AI Agent & RAG apps you can actually run — clone, customize, ship.
Categories
Alternatives to awesome-llm-apps
Are you the builder of awesome-llm-apps?
Claim this artifact to get a verified badge, access match analytics, see which intents users search for, and manage your listing.
Get the weekly brief
New tools, rising stars, and what's actually worth your time. No spam.
Data Sources
Looking for something else?
Search →