awesome-LLM-resources
ModelFree🧑🚀 全世界最好的LLM资料总结(多模态生成、Agent、辅助编程、AI审稿、数据处理、模型训练、模型推理、o1 模型、MCP、小语言模型、视觉语言模型) | Summary of the world's best LLM resources.
Capabilities15 decomposed
bilingual hierarchical resource catalog indexing and navigation
Medium confidenceOrganizes 300+ LLM ecosystem resources across 25+ categories using a bilingual (Chinese/English) hierarchical markdown structure deployed via Jekyll GitHub Pages. The catalog uses a consistent section pattern with category headers, resource links, and descriptions that enable both human browsing and programmatic discovery through GitHub's raw markdown API. Each resource is tagged with domain (foundation, deployment, multimodal, etc.) enabling cross-domain navigation and filtering.
Uses a bilingual hierarchical organization (Chinese-first naming convention) across 25+ domain categories (Foundation & Training, RAG Systems, Agentic RL, Multimodal Systems, etc.) with 1,278-line single-file architecture enabling GitHub Pages deployment without backend infrastructure. Integrates DeepWiki architectural analysis to provide technical context for each category section.
More comprehensive and domain-specific than Papers with Code or Hugging Face Model Hub for LLM ecosystem discovery; bilingual support and architectural depth analysis differentiates from English-only awesome lists.
foundation and training resource aggregation with data-to-model pipeline mapping
Medium confidenceCatalogs 40+ resources spanning data processing, model training, fine-tuning frameworks, and reinforcement learning approaches. The catalog maps the complete pipeline from raw data curation through foundation model training, including tools for data annotation (Label Studio, Argilla), preprocessing (Hugging Face Datasets), fine-tuning (Unsloth, LLaMA-Factory), and agentic RL (veRL, AReaL). Resources are organized by training methodology (supervised fine-tuning, RLHF, DPO, GRPO) enabling builders to select appropriate frameworks for their training objectives.
Uniquely maps agentic reinforcement learning frameworks (veRL, AReaL, slime, Agent Lightning) alongside traditional fine-tuning, reflecting the shift toward reasoning model training. Includes specialized sections for GRPO (Group Relative Policy Optimization) and reasoning model training pipelines used in DeepSeek-R1 replication.
More comprehensive than Papers with Code for training infrastructure; includes both data processing and RL training frameworks in one taxonomy, whereas most resources separate these concerns.
advanced reasoning and o1/o3 model resource aggregation
Medium confidenceCatalogs 15+ resources for advanced reasoning models (OpenAI o1, o3, DeepSeek-R1) and open-source reasoning model implementations. The catalog maps how reasoning models differ from standard LLMs (chain-of-thought training, test-time compute, verification), including training approaches (GRPO, RL-based reasoning) and inference patterns. Resources span both commercial reasoning APIs and open-source implementations, enabling builders to understand and implement advanced reasoning capabilities.
Focuses specifically on advanced reasoning models (o1, o3, DeepSeek-R1) and their training approaches (GRPO, RL-based reasoning), reflecting the emerging frontier of reasoning-focused LLMs. Includes both commercial APIs and open-source implementations, enabling builders to understand and replicate reasoning capabilities.
Uniquely focused on reasoning model training and implementation; most LLM resources treat reasoning as a capability of standard models rather than a distinct model category.
small and efficient model resource aggregation with optimization technique mapping
Medium confidenceCatalogs 25+ small and efficient LLM models (Phi, TinyLlama, Mistral 7B, Qwen, Gemma) organized by optimization approach: quantization (GPTQ, AWQ, GGUF), distillation, pruning, and architectural efficiency. The catalog maps how efficient models trade off capability for size/speed, including benchmarks on standard tasks. Resources span both pre-optimized models and optimization frameworks, enabling builders to select or create efficient models for resource-constrained deployments.
Organizes efficient models by optimization approach (quantization, distillation, pruning, architectural efficiency) rather than just model name. Includes both pre-optimized models (Phi, TinyLlama) and optimization frameworks, reflecting the spectrum from ready-to-use to custom optimization.
More optimization-technique-focused than individual model documentation; enables builders to understand efficiency tradeoffs and select or create efficient models matching their constraints.
model context protocol (mcp) resource aggregation with integration pattern guidance
Medium confidenceCatalogs resources for Model Context Protocol (MCP), a standardized protocol for LLM context management and tool integration. The catalog maps MCP implementations, client libraries, and server implementations, including integration patterns with LLM applications. Resources span both MCP specification documentation and practical implementations, enabling builders to understand and implement MCP-based context management and tool orchestration.
Focuses specifically on Model Context Protocol (MCP) as a standardized approach to context management and tool integration, distinct from custom tool calling implementations. Maps MCP specification, client libraries, and server implementations, reflecting the emerging standardization of LLM context protocols.
Uniquely focused on MCP standardization; most LLM resources treat tool integration as framework-specific rather than protocol-based.
learning resources aggregation spanning books, courses, and technical papers
Medium confidenceCatalogs 50+ learning resources organized by format: books (LLM fundamentals, prompt engineering, RAG), courses (university courses, online platforms), and technical papers (foundational research, recent advances). The catalog maps resources by topic (transformer architecture, fine-tuning, agents, multimodal) and audience level (beginner, intermediate, advanced), enabling learners to find appropriate educational materials for their background and goals.
Organizes learning resources by format (books, courses, papers) and topic (transformers, fine-tuning, agents, multimodal) rather than just listing materials. Includes both foundational resources and cutting-edge research papers, reflecting the breadth of LLM knowledge.
More topic-and-format-focused than general learning platforms; enables learners to find specific educational materials for their background and goals.
interactive demo and model arena discovery for comparative evaluation
Medium confidenceCatalogs 10+ interactive platforms (Hugging Face Spaces, OpenRouter, Chatbot Arena, Together Playground) enabling side-by-side model comparison and evaluation. The catalog maps how platforms enable comparative evaluation (same prompt across models, user voting, leaderboards) and integration with multiple model providers. Resources span both community-driven arenas (Chatbot Arena) and commercial platforms (OpenRouter), enabling builders to evaluate models before integration.
Focuses on interactive platforms enabling side-by-side model comparison and community-driven evaluation, distinct from automated benchmarking. Includes both community arenas (Chatbot Arena) and commercial platforms (OpenRouter), reflecting the spectrum from open to managed evaluation.
More interactive-and-comparative-focused than static benchmarks; enables real-time model evaluation and community-driven quality assessment.
inference and serving framework discovery with deployment pattern guidance
Medium confidenceAggregates 30+ inference serving frameworks (vLLM, TensorRT-LLM, SGLang, Ollama, LM Studio) organized by deployment pattern (local, cloud, edge, batch). The catalog maps frameworks to specific optimization techniques (quantization, batching, KV-cache optimization) and hardware targets (CPU, GPU, mobile). Resources include both open-source inference engines and commercial serving platforms, enabling builders to select frameworks matching their latency, throughput, and cost requirements.
Organizes inference frameworks by deployment pattern (local, cloud, edge, batch) rather than just framework name, with explicit mapping to optimization techniques (quantization, batching, KV-cache) and hardware targets. Includes both open-source engines (vLLM, SGLang, Ollama) and commercial platforms (Together AI, Replicate).
More deployment-pattern-focused than framework-specific documentation; enables builders to find solutions by use case (low-latency API, batch processing, edge deployment) rather than learning individual framework APIs.
rag system component discovery with pipeline architecture mapping
Medium confidenceCatalogs 50+ RAG components organized by pipeline stage: document ingestion (LlamaIndex, LangChain), vector databases (Pinecone, Weaviate, Milvus, Qdrant), retrieval optimization (BM25, semantic search, hybrid retrieval), and generation orchestration. The catalog maps how components integrate into end-to-end RAG pipelines, including chunking strategies, embedding models, reranking, and prompt engineering techniques. Resources span both framework-level solutions (LlamaIndex, LangChain) and specialized components (vector databases, rerankers).
Maps RAG systems by pipeline stage (ingestion → chunking → embedding → retrieval → reranking → generation) with explicit component categories, enabling builders to understand integration points. Includes both high-level frameworks (LlamaIndex, LangChain) and specialized components (Qdrant, Milvus, Rerankers), reflecting the modular RAG ecosystem.
More pipeline-architecture-focused than individual framework documentation; enables builders to understand how components fit together rather than learning one framework's abstractions.
ai agents and orchestration framework catalog with tool-use pattern mapping
Medium confidenceAggregates 40+ agent frameworks (AutoGen, LangGraph, CrewAI, Swarm, Rigging) organized by orchestration pattern: multi-agent coordination, tool calling, memory management, and planning strategies. The catalog maps how frameworks implement agent capabilities (function calling, state management, tool registry) and integration points with LLM APIs (OpenAI, Anthropic, Ollama). Resources include both high-level agent frameworks and lower-level orchestration primitives, enabling builders to select frameworks matching their agent complexity and coordination requirements.
Organizes agent frameworks by orchestration pattern (multi-agent coordination, tool calling, memory management, planning) rather than just framework name. Includes both high-level frameworks (AutoGen, CrewAI) and lower-level primitives (LangGraph, Swarm), reflecting the spectrum from abstraction to control.
More pattern-focused than individual framework documentation; enables builders to understand orchestration approaches (hierarchical vs peer-to-peer) and select frameworks matching their coordination requirements.
coding assistant and development tool resource aggregation
Medium confidenceCatalogs 25+ coding-focused LLM tools (GitHub Copilot, Cursor, Codeium, Aider, Continue) organized by capability: code completion, refactoring, debugging, code review, and test generation. The catalog maps tools by integration point (IDE plugins, CLI, web-based) and supported languages/frameworks. Resources include both commercial coding assistants and open-source alternatives, enabling developers to select tools matching their development workflow and language preferences.
Organizes coding tools by capability (completion, refactoring, debugging, review) and integration point (IDE, CLI, web) rather than just tool name. Includes both commercial (GitHub Copilot, Cursor) and open-source (Aider, Continue) options, enabling developers to evaluate alternatives.
More capability-focused than individual tool documentation; enables developers to find tools for specific coding tasks (refactoring, debugging) rather than learning one tool's full feature set.
search and research tool discovery with information retrieval pattern mapping
Medium confidenceAggregates 20+ search and research tools (Perplexity, Tavily, Exa, Metaphor, web search APIs) organized by retrieval pattern: web search, academic paper search, semantic search, and real-time information retrieval. The catalog maps how tools integrate with LLM applications through APIs, including search result formatting and citation handling. Resources span both consumer-facing search tools and developer-oriented search APIs, enabling builders to select tools matching their information retrieval requirements.
Organizes search tools by retrieval pattern (web search, academic papers, semantic search, real-time) rather than just tool name. Includes both consumer tools (Perplexity) and developer APIs (Tavily, Exa), reflecting the spectrum from user-facing to programmatic search.
More pattern-focused than individual search tool documentation; enables builders to understand retrieval approaches and select tools matching their information needs.
multimodal system resource aggregation spanning vision, audio, and video
Medium confidenceCatalogs 60+ multimodal resources organized by modality: vision-language models (GPT-4V, Claude Vision, LLaVA), video generation (Sora, Runway, Pika), image generation (DALL-E, Midjourney, Stable Diffusion), speech systems (Whisper, TTS, voice cloning), and unified multimodal models (Gemini, GPT-4o). The catalog maps how multimodal models integrate with LLM applications, including input/output format handling and API integration patterns. Resources span both commercial APIs and open-source models, enabling builders to select tools matching their multimodal requirements.
Organizes multimodal resources by modality (vision, audio, video, unified) rather than just model name. Includes both commercial APIs (OpenAI, Anthropic, Runway) and open-source models (LLaVA, Stable Diffusion, Whisper), reflecting the spectrum from managed services to self-hosted solutions.
More modality-focused than individual model documentation; enables builders to understand multimodal capabilities and select tools matching their input/output requirements.
evaluation and benchmarking framework discovery with metric-based organization
Medium confidenceAggregates 30+ evaluation frameworks and benchmarks organized by evaluation type: LLM capability benchmarks (MMLU, HumanEval, MATH), RAG evaluation (RAGAS, TruLens), agent evaluation (AgentBench), and safety/alignment evaluation. The catalog maps how evaluation frameworks measure specific capabilities (reasoning, coding, knowledge, safety) and integrate with model development pipelines. Resources span both standardized benchmarks (MMLU, HumanEval) and specialized evaluation tools (RAGAS for RAG, TruLens for observability).
Organizes evaluation frameworks by evaluation type (capability benchmarks, RAG evaluation, agent evaluation, safety) rather than just framework name. Includes both standardized benchmarks (MMLU, HumanEval) and specialized tools (RAGAS, TruLens, AgentBench), reflecting the diversity of evaluation needs.
More evaluation-type-focused than individual benchmark documentation; enables teams to find appropriate evaluation tools for their specific use case (RAG, agents, safety).
llm api service comparison and integration guidance
Medium confidenceCatalogs 15+ LLM API providers (OpenAI, Anthropic, Google, Meta, Mistral, Qwen, Together AI, Replicate) organized by provider type: frontier models (GPT-4, Claude), open-source model APIs (Mistral, Qwen, Llama), and specialized providers (Together AI for fine-tuning, Replicate for inference). The catalog maps API capabilities (model versions, context length, pricing, rate limits) and integration patterns, enabling builders to select providers matching their cost, latency, and capability requirements.
Organizes LLM providers by provider type (frontier models, open-source APIs, specialized services) rather than just provider name. Includes both commercial APIs (OpenAI, Anthropic, Google) and open-source model APIs (Mistral, Qwen, Together AI), reflecting the spectrum from proprietary to open models.
More provider-type-focused than individual API documentation; enables builders to understand provider categories and select services matching their cost, capability, and control requirements.
Capabilities are decomposed by AI analysis. Each maps to specific user intents and improves with match feedback.
Related Artifactssharing capabilities
Artifacts that share capabilities with awesome-LLM-resources, ranked by overlap. Discovered automatically through the match graph.
awesome-generative-ai
A curated list of Generative AI tools, works, models, and references
LlamaIndex
Transform enterprise data into powerful LLM applications...
OpenAI: o3 Deep Research
o3-deep-research is OpenAI's advanced model for deep research, designed to tackle complex, multi-step research tasks. Note: This model always uses the 'web_search' tool which adds additional cost.
modelcontextprotocol.io
for comprehensive guides, best practices, and technical details on implementing MCP servers.
@splicr/mcp-server
Splicr MCP server — route what you read to what you're building
Orygo AI
Harness AI to centralize, search, and manage knowledge effortlessly across...
Best For
- ✓LLM practitioners building production systems who need ecosystem overview
- ✓Teams evaluating multiple framework options across foundation, deployment, and application layers
- ✓Non-English speakers seeking Chinese-language LLM resources and documentation
- ✓Researchers mapping the LLM landscape for comparative analysis
- ✓ML teams training custom LLMs on proprietary data
- ✓Researchers implementing RLHF, DPO, or agentic RL training pipelines
- ✓Organizations migrating from closed-source models to open-source fine-tuning
- ✓Data engineers building data processing pipelines for model training
Known Limitations
- ⚠No programmatic API — requires parsing markdown or GitHub API to extract structured data
- ⚠Manual curation means resource freshness depends on community contributions; no automated staleness detection
- ⚠No versioning or release tracking for linked projects — links may point to outdated versions
- ⚠Search functionality limited to GitHub's text search; no semantic or category-based filtering interface
- ⚠Catalog links to external frameworks; no integrated training environment or unified API
- ⚠No guidance on framework selection criteria (e.g., when to use Unsloth vs LLaMA-Factory)
Requirements
Input / Output
UnfragileRank
UnfragileRank is computed from adoption signals, documentation quality, ecosystem connectivity, match graph feedback, and freshness. No artifact can pay for a higher rank.
Repository Details
Last commit: Apr 18, 2026
About
🧑🚀 全世界最好的LLM资料总结(多模态生成、Agent、辅助编程、AI审稿、数据处理、模型训练、模型推理、o1 模型、MCP、小语言模型、视觉语言模型) | Summary of the world's best LLM resources.
Categories
Alternatives to awesome-LLM-resources
Are you the builder of awesome-LLM-resources?
Claim this artifact to get a verified badge, access match analytics, see which intents users search for, and manage your listing.
Get the weekly brief
New tools, rising stars, and what's actually worth your time. No spam.
Data Sources
Looking for something else?
Search →