llm-course
ModelFreeCourse to get into Large Language Models (LLMs) with roadmaps and Colab notebooks.
Capabilities17 decomposed
structured-learning-roadmap-navigation
Medium confidenceOrganizes LLM education into three progressive learning tracks (Fundamentals, Scientist, Engineer) with explicit entry points and dependency mapping, implemented as a single markdown hub that links to ~150+ external resources. Users navigate via a hierarchical section structure that maps learning paths to specific topics, with each topic following a consistent pattern of curated articles, videos, and tools. The architecture uses a documentation-first approach where the README.md acts as a central knowledge graph rather than containing executable code.
Uses a three-track learning path architecture (Fundamentals/Scientist/Engineer) with explicit optional vs. core topic designation, enabling learners to skip prerequisites based on background. Most LLM courses use linear progression; this enables parallel tracks with clear entry points.
More structured and goal-oriented than generic LLM resource lists (e.g., Awesome-LLM), with explicit learning paths vs. flat collections of links
theoretical-topic-curation-with-external-linking
Medium confidenceAggregates 24 theoretical topics across three learning paths and embeds curated external references (articles, papers, videos, tools) directly within each topic section. Implementation uses a consistent topic section pattern where each topic links to 3-8 external resources selected for pedagogical value. The curation layer filters and organizes content from diverse sources (research papers, blog posts, YouTube, GitHub projects) into a single navigable structure without duplicating content.
Implements a consistent topic section pattern (theory + curated resources + tools) across 24 topics, enabling predictable navigation. Each topic embeds ~3-8 hand-selected external resources rather than generating them, ensuring quality over quantity.
More curated and pedagogically structured than raw resource aggregators; provides context and organization vs. flat link collections like Awesome-LLM
rag-and-vector-storage-architecture-guidance
Medium confidenceProvides educational content on Retrieval Augmented Generation (RAG) and vector storage systems, covering vector databases (Pinecone, Weaviate, Milvus), embedding models, retrieval strategies, and advanced RAG techniques (re-ranking, query expansion, hybrid search). Content is organized as two dedicated sections within the LLM Engineer track and links to vector database documentation, embedding model resources, and RAG frameworks (LangChain, LlamaIndex). This capability enables practitioners to build knowledge-grounded LLM applications without fine-tuning.
Separates basic RAG and advanced RAG into distinct sections, with coverage of vector databases, embedding models, and retrieval strategies. Links to both foundational RAG papers and practical frameworks (LangChain, LlamaIndex), enabling end-to-end RAG system building.
More comprehensive than single-framework tutorials; more practical than research papers because it includes tool recommendations and architecture patterns
llm-agents-and-tool-orchestration-guidance
Medium confidenceProvides educational content on building LLM agents that can plan, reason, and use tools to accomplish complex tasks. Content covers agent architectures (ReAct, Chain-of-Thought), tool calling and function schemas, planning strategies, and agent frameworks (LangChain, AutoGPT, CrewAI). This capability is organized as a dedicated section within the LLM Engineer track and links to agent research papers, framework documentation, and implementation examples. Enables practitioners to build autonomous systems that go beyond simple prompt-response interactions.
Provides dedicated agent section with coverage of agent architectures (ReAct, Chain-of-Thought), tool calling patterns, and multi-agent orchestration. Links to both foundational agent research and practical frameworks, enabling practitioners to build agents from scratch or using existing frameworks.
More comprehensive than single-framework tutorials; more practical than research papers because it includes framework recommendations and implementation patterns
inference-optimization-and-serving-strategies
Medium confidenceProvides educational content on optimizing LLM inference for latency and throughput, covering techniques like batching, caching, quantization, and serving frameworks (vLLM, TensorRT-LLM, Ollama). Content is organized as a dedicated section within the LLM Engineer track and links to optimization papers, serving framework documentation, and performance benchmarks. This capability enables practitioners to deploy models efficiently and meet production latency/throughput requirements.
Provides dedicated inference optimization section with coverage of multiple optimization techniques (batching, caching, quantization) and serving frameworks. Links to both optimization research and practical framework documentation, enabling practitioners to choose and implement optimization strategies.
More comprehensive than single-framework documentation; more practical than research papers because it includes framework comparisons and implementation guidance
llm-deployment-and-infrastructure-patterns
Medium confidenceProvides educational content on deploying LLMs to production, covering containerization (Docker), orchestration (Kubernetes), cloud platforms (AWS, GCP, Azure), monitoring, and operational considerations. Content is organized as a dedicated section within the LLM Engineer track and links to deployment frameworks, cloud documentation, and best practices. This capability enables practitioners to move models from development to production with proper infrastructure, monitoring, and reliability patterns.
Provides dedicated deployment section with coverage of containerization, orchestration, cloud platforms, and operational considerations. Links to both deployment frameworks and cloud documentation, enabling practitioners to deploy models across different infrastructure options.
More LLM-specific than generic DevOps guides; more practical than research papers because it includes tool recommendations and architecture patterns
llm-security-and-safety-considerations
Medium confidenceProvides educational content on securing LLM applications and addressing safety concerns, covering prompt injection attacks, data privacy, model poisoning, adversarial robustness, and compliance considerations. Content is organized as a dedicated section within the LLM Engineer track and links to security research, safety frameworks, and best practices. This capability enables practitioners to build LLM applications with appropriate security and safety guardrails.
Provides dedicated security section with coverage of prompt injection, data privacy, model poisoning, and compliance. Links to both security research and practical frameworks, enabling practitioners to implement security and safety measures appropriate to their threat model.
More LLM-specific than generic security guides; more practical than research papers because it includes implementation guidance and best practices
evaluation-and-benchmarking-frameworks
Medium confidenceProvides educational content on evaluating LLM quality and performance, covering automatic metrics (BLEU, ROUGE, BERTScore), human evaluation, benchmarks (MMLU, HellaSwag, TruthfulQA), and evaluation frameworks. Content is organized as a dedicated section within the LLM Scientist track and links to evaluation papers, benchmark datasets, and evaluation tools. This capability enables practitioners to measure model quality and compare different models or training approaches.
Provides dedicated evaluation section with coverage of automatic metrics, human evaluation, and standard benchmarks. Links to both evaluation research and practical frameworks, enabling practitioners to measure model quality comprehensively.
More comprehensive than single-metric tutorials; more practical than research papers because it includes benchmark datasets and evaluation tools
new-trends-and-emerging-techniques-curation
Medium confidenceProvides curated content on emerging LLM techniques and research trends, covering recent advances in model architecture, training methods, and applications. Content is organized as a dedicated section within the LLM Scientist track and links to recent research papers, blog posts, and tools implementing new techniques. This capability enables practitioners to stay current with rapidly evolving LLM field and understand cutting-edge approaches.
Provides dedicated section for emerging techniques and trends, enabling practitioners to discover and evaluate cutting-edge approaches. Most LLM courses focus on established techniques; this section bridges the gap to research frontiers.
More curated than raw research feeds; more accessible than academic conferences because content is organized and contextualized
hands-on-colab-notebook-integration
Medium confidenceProvides 23 executable Jupyter notebooks hosted on Google Colab that implement theoretical concepts from the course, organized into four categories: Automated Tools (8), Fine-tuning (6), Quantization (4), and Advanced Techniques (5). Notebooks are embedded as links within relevant course sections, creating a tight coupling between theory and practice. Each notebook implements specific techniques (e.g., LoRA fine-tuning, GGUF quantization, model merging) with runnable code that requires only a Google account and GPU access.
Organizes 23 notebooks into four functional categories (Automated Tools, Fine-tuning, Quantization, Advanced) with direct embedding in course sections, creating a theory-to-practice pipeline. Notebooks are hosted on Colab (zero setup) rather than requiring local installation, lowering barrier to entry.
More accessible than local notebook repositories because Colab requires no setup; more integrated than standalone notebooks because they're linked to specific course topics
llm-fundamentals-prerequisite-track
Medium confidenceProvides an optional foundational learning path covering Mathematics for Machine Learning, Python for Machine Learning, Neural Networks, and Natural Language Processing. This track is marked as optional (not required for advanced learners) and spans lines 74-157 of the README, serving as a prerequisite for both Scientist and Engineer tracks. Implementation uses a modular topic structure where each fundamental topic links to external resources (textbooks, courses, tutorials) rather than providing original content.
Explicitly marks fundamentals as optional and modular, allowing learners with existing ML knowledge to skip directly to Scientist/Engineer tracks. Most LLM courses require linear progression through basics; this enables flexible entry points.
More flexible than linear ML courses because prerequisites are optional; more focused than general ML curricula because resources are curated for LLM practitioners
llm-scientist-research-and-training-track
Medium confidenceProvides a core learning path (8 topics, lines 159-304) focused on building and training LLMs from scratch, covering LLM Architecture, Pre-Training Models, Post-Training Datasets, Supervised Fine-Tuning, Preference Alignment, Evaluation, Quantization, and New Trends. This track is designed for researchers and practitioners wanting to understand model internals and training pipelines. Implementation uses the same topic-curation pattern as Fundamentals but with deeper technical content (papers, research blogs, training frameworks).
Organizes 8 core research topics in a logical progression (Architecture → Pre-Training → Post-Training → Evaluation → Optimization), with each topic linking to both foundational papers and recent research. Includes dedicated quantization and evaluation sections that bridge theory and practice.
More research-focused than engineering-oriented courses; provides deeper technical content than introductory LLM guides but less practical than deployment-focused resources
llm-engineer-production-and-deployment-track
Medium confidenceProvides a core learning path (8 topics, lines 305-440) focused on deploying and operating LLMs in production, covering Running LLMs, Vector Storage, Retrieval Augmented Generation (RAG), Advanced RAG, Agents, Inference Optimization, Deployment, and Security. This track is designed for engineers building LLM applications and systems. Implementation uses the same topic-curation pattern but emphasizes tools, frameworks, and operational concerns over research papers.
Organizes 8 production-focused topics in a logical pipeline (Running → Storage → Retrieval → Agents → Optimization → Deployment → Security), with emphasis on tools and frameworks rather than research. Includes dedicated sections for RAG and Agents, which are critical for production LLM applications.
More operations-focused than research-oriented courses; provides practical deployment guidance vs. theoretical LLM courses that lack production context
transformer-architecture-educational-content
Medium confidenceProvides comprehensive educational material on transformer architecture fundamentals, covering decoder-only architectures (GPT, Llama, Mistral), tokenization methods, attention mechanism variants, and text generation strategies. Content is organized as a dedicated section within the LLM Scientist track and links to foundational papers (Attention is All You Need), implementation guides, and visual explanations. This capability serves as the architectural foundation for understanding all downstream topics (pre-training, fine-tuning, quantization).
Organizes transformer architecture as a dedicated foundational section with explicit coverage of decoder-only vs. encoder-decoder variants, tokenization, and attention mechanisms. Most LLM courses assume transformer knowledge; this provides structured learning for those needing to build it from scratch.
More comprehensive than blog post explanations; more accessible than original research papers because it curates multiple explanations and implementations
pre-training-and-dataset-curation-guidance
Medium confidenceProvides educational content on pre-training LLMs from scratch and curating post-training datasets, covering model initialization, training objectives (next-token prediction, masked language modeling), dataset composition, and scaling laws. Content is organized as two dedicated sections within the LLM Scientist track and links to research papers (Chinchilla, Scaling Laws), dataset resources (Common Crawl, Wikipedia), and training frameworks (Hugging Face Transformers, Megatron). This capability bridges architecture understanding with practical training considerations.
Separates pre-training and post-training dataset considerations into distinct sections, with explicit coverage of scaling laws and dataset composition. Links to both foundational research (Chinchilla scaling laws) and practical resources (dataset repositories, training frameworks).
More comprehensive than blog posts on pre-training; more practical than pure research papers because it includes tool recommendations and dataset resources
fine-tuning-and-preference-alignment-implementation
Medium confidenceProvides educational content and 6 executable notebooks on supervised fine-tuning (SFT) and preference alignment techniques (RLHF, DPO, IPO). Content covers fine-tuning methodologies, dataset preparation, and alignment algorithms, with notebooks implementing LoRA fine-tuning, full fine-tuning, and preference alignment on Colab. This capability enables practitioners to adapt pre-trained models to specific tasks and align outputs with human preferences without requiring massive compute.
Provides both theoretical content (alignment algorithms, fine-tuning trade-offs) and 6 executable notebooks implementing SFT and preference alignment. Notebooks cover both efficient (LoRA) and full fine-tuning, enabling practitioners to choose based on their constraints.
More comprehensive than single-technique tutorials; more accessible than research papers because notebooks provide working code and step-by-step guidance
quantization-techniques-and-optimization
Medium confidenceProvides educational content and 4 executable notebooks on quantization techniques for reducing model size and inference latency, covering post-training quantization (PTQ), quantization-aware training (QAT), and specific formats (GGUF, GPTQ, AWQ). Content links to research papers on quantization methods and includes notebooks implementing quantization pipelines on Colab. This capability enables deployment of large models on resource-constrained hardware without significant quality loss.
Provides 4 dedicated quantization notebooks covering multiple formats (GGUF, GPTQ, AWQ) with explicit trade-off analysis. Most courses treat quantization as a single technique; this provides format-specific guidance and working implementations.
More practical than research papers on quantization because it includes working code; more comprehensive than single-format tutorials because it covers multiple quantization methods
Capabilities are decomposed by AI analysis. Each maps to specific user intents and improves with match feedback.
Related Artifactssharing capabilities
Artifacts that share capabilities with llm-course, ranked by overlap. Discovered automatically through the match graph.
Roadmap
A roadmap connecting many of the most important concepts in machine learning, how to learn them, and what tools to use to perform them.
awesome-generative-ai-guide
A one stop repository for generative AI research updates, interview resources, notebooks and much more!
AgentGuide
https://adongwanai.github.io/AgentGuide | AI Agent开发指南 | LangGraph实战 | 高级RAG | 转行大模型 | 大模型面试 | 算法工程师 | 面试题库 | 强化学习|数据合成
generative-ai-for-beginners
21 Lessons, Get Started Building with Generative AI
ai-notes
notes for software engineers getting up to speed on new AI developments. Serves as datastore for https://latent.space writing, and product brainstorming, but has cleaned up canonical references under the /Resources folder.
AI and Machine Learning Roadmaps
Roadmaps featuring essential concepts, learning methods, and the tools to put them into...
Best For
- ✓self-taught developers transitioning into LLM engineering
- ✓ML practitioners wanting a structured path from theory to production
- ✓teams building internal LLM knowledge bases
- ✓researchers wanting a curated bibliography for LLM topics
- ✓educators building course materials from vetted sources
- ✓practitioners needing quick access to both theory and tools
- ✓teams building knowledge-grounded chatbots and Q&A systems
- ✓practitioners wanting to add domain knowledge to LLMs without fine-tuning
Known Limitations
- ⚠No interactive quizzes or progress tracking — purely reference-based navigation
- ⚠Requires external tool access (Colab, GitHub) to execute notebooks
- ⚠Content updates depend on manual curation; no automated resource discovery
- ⚠No quality scoring or difficulty ratings for external resources — all links treated equally
- ⚠Curation is manual and static; no dynamic ranking based on community feedback
- ⚠External links may break or become outdated without automated monitoring
Requirements
Input / Output
UnfragileRank
UnfragileRank is computed from adoption signals, documentation quality, ecosystem connectivity, match graph feedback, and freshness. No artifact can pay for a higher rank.
Repository Details
Last commit: Feb 5, 2026
About
Course to get into Large Language Models (LLMs) with roadmaps and Colab notebooks.
Categories
Alternatives to llm-course
Are you the builder of llm-course?
Claim this artifact to get a verified badge, access match analytics, see which intents users search for, and manage your listing.
Get the weekly brief
New tools, rising stars, and what's actually worth your time. No spam.
Data Sources
Looking for something else?
Search →