Prompt Engineering Guide
RepositoryFreeGuide and resources for prompt engineering.
Capabilities15 decomposed
multi-language prompt engineering documentation platform
Medium confidenceServes comprehensive prompt engineering educational content across 11 languages using Next.js 13 with Nextra 2.13 static site generation. The platform implements a middleware-based internationalization system that routes users to language-specific content (e.g., pages/introduction/basics.en.mdx, pages/introduction/basics.ar.mdx) with automatic language detection and manual override capabilities. Content is organized hierarchically through _meta.json files that define navigation structure per language, enabling consistent UX across locales while maintaining independent content management.
Uses Nextra 2.13's built-in i18n system with file-based language routing (_meta.{lang}.json) rather than URL parameters, enabling clean SEO-friendly URLs and automatic language-specific navigation hierarchies without additional routing logic
Simpler than Docusaurus i18n setup because language variants are defined declaratively in metadata files rather than requiring separate site instances or complex routing configuration
structured prompt engineering technique documentation with visual diagrams
Medium confidenceProvides comprehensive documentation of 15+ prompting techniques (Zero-Shot, Few-Shot, Chain-of-Thought, Tree of Thoughts, ReAct, RAG, PAL, Self-Consistency, Prompt Chaining, APE) organized as MDX pages with embedded PNG diagrams illustrating technique workflows. Each technique page includes conceptual explanation, implementation patterns, code examples, and visual architecture diagrams (e.g., img/ape-zero-shot-cot.png, img/active-prompt.png) that show how techniques compose with LLM inference. The documentation structure enables cross-referencing between techniques and provides practical guidance on when to apply each approach.
Organizes prompting techniques as a taxonomy with visual workflow diagrams showing how each technique structures LLM reasoning, rather than treating them as isolated tips. Includes technique composition patterns (e.g., CoT + Self-Consistency) showing how techniques can be layered for improved reliability.
More comprehensive than scattered blog posts because it provides unified documentation of 15+ techniques with consistent structure, visual diagrams, and cross-references showing technique relationships and composition patterns
fine-tuning guidance for model customization
Medium confidenceDocuments fine-tuning approaches for customizing LLMs (e.g., GPT-4o fine-tuning) with guidance on when fine-tuning is appropriate vs. prompt engineering, data preparation strategies, and evaluation metrics. The guide covers training data requirements, cost-benefit analysis, and how to combine fine-tuning with prompt engineering for optimal results. It includes examples of fine-tuning for domain-specific tasks and comparison with few-shot prompting effectiveness.
Provides decision framework for fine-tuning vs. prompt engineering rather than assuming fine-tuning is always better, with cost-benefit analysis and guidance on when each approach is appropriate. Includes data preparation patterns specific to fine-tuning.
More strategic than fine-tuning API documentation because it helps teams decide whether fine-tuning is worth the investment; more practical than academic papers because it includes concrete data preparation and cost analysis
synthetic dataset generation with llms
Medium confidenceDocuments techniques for using LLMs to generate synthetic training data, including prompt engineering patterns for data generation, quality control strategies, and diversity mechanisms. The guide covers how to structure generation prompts to produce varied, high-quality synthetic examples, validation approaches to ensure synthetic data quality, and use cases where synthetic data is most effective (e.g., data augmentation, privacy-preserving datasets). Includes examples of generating synthetic datasets for classification, NER, and other NLP tasks.
Focuses on prompt engineering for synthetic data generation, providing patterns for designing generation prompts that produce diverse, high-quality examples. Includes quality validation strategies specific to synthetic data.
More practical than general data augmentation guides because it specifically addresses LLM-based generation; more comprehensive than single-task examples because it covers multiple NLP tasks and quality control strategies
ai agent architecture and context engineering guide
Medium confidenceDocuments agent design patterns and context engineering strategies for building autonomous LLM agents, including agent framework components (planning, reasoning, tool use), context management for agents, and patterns for agent-environment interaction. The guide covers how to structure agent prompts for effective reasoning, manage context across multiple agent steps, and design agent workflows. It includes examples of ReAct agents, planning-based agents, and hierarchical agent architectures.
Provides comprehensive agent design patterns including context engineering strategies for managing agent state across multiple reasoning steps, rather than treating agents as simple tool-calling wrappers. Includes patterns for hierarchical agents and agent composition.
More comprehensive than single-framework documentation because it covers multiple agent architectures and design patterns; more practical than academic papers because it includes implementation guidance and context management strategies
bias detection and mitigation in llm outputs
Medium confidenceDocuments techniques for identifying and mitigating biases in LLM-generated content, including bias categories (gender, racial, cultural), detection strategies through prompting, and mitigation patterns. The guide covers how to structure prompts to reduce bias, validate outputs for bias, and implement fairness checks. It includes examples of biased outputs, detection prompts, and mitigation strategies for different bias types.
Focuses specifically on bias detection and mitigation through prompting rather than treating bias as a general safety concern, providing concrete detection patterns and mitigation strategies. Includes categorization of bias types and domain-specific detection approaches.
More actionable than general fairness frameworks because it provides specific prompting patterns for bias detection and mitigation; more comprehensive than scattered blog posts because it covers multiple bias types and detection strategies
prompt chaining and multi-step workflow orchestration
Medium confidenceDocuments prompt chaining techniques for decomposing complex tasks into sequences of LLM calls, including workflow design patterns, context passing between steps, and error handling strategies. The guide covers how to structure individual prompts in a chain, manage outputs from one step as inputs to the next, and handle failures in multi-step workflows. It includes examples of chaining for complex reasoning tasks, content generation pipelines, and data processing workflows.
Provides systematic patterns for designing prompt chains including context passing strategies and error handling, rather than treating chaining as simple sequential prompting. Includes workflow design patterns for different task types.
More comprehensive than scattered examples because it provides systematic design patterns for multi-step workflows; more practical than academic papers because it includes implementation guidance and error handling strategies
interactive jupyter notebook examples for prompt engineering techniques
Medium confidenceProvides executable Jupyter notebooks (pe-chatgpt-adversarial.ipynb, pe-pal.ipynb) demonstrating prompt engineering techniques with live code examples that can be run in Colab or local environments. Notebooks include step-by-step implementation of techniques like Program-Aided Language Models (PAL) and adversarial prompting, with actual API calls to LLMs, output examples, and explanations of results. This enables hands-on learning where practitioners can modify prompts, observe LLM responses, and experiment with parameter variations in real-time.
Provides fully executable notebooks with real LLM API integration rather than pseudocode or static examples, allowing learners to modify prompts and immediately observe model behavior changes. Includes adversarial prompting examples showing actual jailbreak attempts and model responses.
More practical than documentation-only guides because code can be executed and modified in real-time; more reproducible than blog post examples because notebooks capture exact API calls and responses
llm model capability and parameter reference documentation
Medium confidenceMaintains comprehensive documentation of LLM models (GPT-4, ChatGPT, open-source models) with detailed parameter explanations (temperature, top_p, frequency_penalty, etc.), context window sizes, cost information, and capability matrices. Documentation is organized by model family (OpenAI, open-source) and includes guidance on when to use each parameter setting and how parameter choices affect prompt engineering effectiveness. The reference includes model-specific features like context caching (Gemini), fine-tuning capabilities (GPT-4o), and function calling support across providers.
Provides unified reference documentation for both proprietary (OpenAI, Anthropic) and open-source models with consistent parameter explanations and capability matrices, rather than requiring developers to consult separate provider documentation for each model
More accessible than scattered provider documentation because it consolidates model information in one place with consistent formatting and cross-model comparisons; includes practical guidance on parameter tuning that provider docs don't always explain
prompt engineering research paper collection and synthesis
Medium confidenceCurates and synthesizes findings from peer-reviewed prompt engineering research papers organized by topic (RAG, LLM Agents, prompting techniques). The collection includes paper summaries, key findings, and connections to practical techniques documented in the guide. This bridges academic research and practical application by showing how research insights (e.g., Chain-of-Thought effectiveness from Wei et al.) translate into actionable prompting strategies. The architecture enables cross-referencing between research findings and technique documentation.
Synthesizes academic research findings into practical prompting techniques rather than just listing papers, showing explicit connections between research insights (e.g., CoT improves reasoning) and implementation patterns. Organized by application domain (RAG, Agents) rather than by paper publication date.
More useful than raw paper repositories because it provides curated summaries and connects research to practical techniques; more rigorous than blog posts because it grounds recommendations in peer-reviewed research
adversarial prompting and robustness evaluation guide
Medium confidenceDocuments adversarial prompting techniques (jailbreaks, prompt injection, prompt leaking) with examples of how to attack LLM systems and defensive strategies to mitigate risks. The guide includes categorized attack patterns (e.g., role-playing jailbreaks, encoding-based attacks), code examples showing actual attack implementations, and defensive prompt engineering patterns. This enables security-conscious teams to evaluate their LLM systems' robustness and understand attack surface before deployment.
Provides both attack examples and defensive strategies in one guide, enabling teams to understand threats and implement mitigations. Includes categorized attack patterns (role-playing, encoding, context confusion) showing how different attack vectors work mechanically.
More comprehensive than scattered security advisories because it provides systematic categorization of attack types and defensive patterns; more actionable than academic papers because it includes executable examples and defensive prompt templates
prompt engineering application use-case library
Medium confidenceDocuments real-world applications of prompt engineering across domains (ChatGPT conversational applications, code generation, synthetic data generation, workplace case studies) with concrete examples showing how techniques apply to specific problems. Each use case includes problem statement, prompting approach, code examples, and results. The library demonstrates technique composition (e.g., using CoT for code generation, RAG for domain-specific QA) and shows how to adapt techniques for different domains.
Organizes applications by domain with concrete problem-solution pairs rather than generic technique descriptions, showing how to compose multiple techniques for specific use cases (e.g., CoT + function calling for code generation)
More practical than technique-focused documentation because it shows end-to-end examples of solving real problems; more transferable than single blog posts because it covers multiple domains with consistent structure
prompt reliability and factuality improvement guide
Medium confidenceDocuments techniques for improving LLM output reliability and factuality, including self-consistency prompting, verification strategies, and bias mitigation approaches. The guide covers how to structure prompts to reduce hallucinations, detect factual errors, and improve consistency across multiple generations. It includes practical patterns like generating multiple outputs and voting, fact-checking prompts, and domain-specific validation approaches.
Focuses specifically on reliability and factuality rather than general prompting, providing techniques like self-consistency voting and fact-checking prompts that directly address LLM limitations. Includes patterns for detecting and mitigating hallucinations.
More focused than general prompting guides because it specifically addresses reliability concerns; more practical than theoretical papers because it provides implementable patterns and verification strategies
function calling and tool integration documentation
Medium confidenceDocuments how to use function calling (tool calling) across different LLM providers (OpenAI, Anthropic, Gemini) with schema-based function definitions, parameter handling, and integration patterns. The documentation includes examples of defining function schemas, handling function responses, and chaining multiple function calls. It covers provider-specific differences in function calling APIs and shows how to structure prompts to encourage appropriate function usage.
Provides unified documentation of function calling across multiple providers (OpenAI, Anthropic, Gemini) with explicit comparison of schema differences and provider-specific behaviors, rather than requiring developers to consult separate provider documentation
More comprehensive than single-provider documentation because it shows how to implement function calling portably across providers; more practical than API reference docs because it includes end-to-end examples and schema design patterns
context caching and optimization guide for long-context applications
Medium confidenceDocuments context caching techniques (e.g., Gemini's context caching feature) and optimization strategies for managing large context windows efficiently. The guide covers how to structure prompts for caching, cache invalidation patterns, and cost-benefit analysis of caching vs. re-processing context. It includes examples of caching long documents, system prompts, and few-shot examples to reduce API costs and latency for repeated queries over the same context.
Focuses specifically on context caching as a performance optimization technique, providing strategies for structuring prompts to maximize cache effectiveness and guidance on cache invalidation patterns. Includes cost-benefit analysis for when caching is worthwhile.
More specialized than general optimization guides because it addresses context caching specifically; more practical than provider documentation because it includes architectural patterns for cache-friendly prompt design
Capabilities are decomposed by AI analysis. Each maps to specific user intents and improves with match feedback.
Related Artifactssharing capabilities
Artifacts that share capabilities with Prompt Engineering Guide, ranked by overlap. Discovered automatically through the match graph.
DreamStudio
DreamStudio is an easy-to-use interface for creating images using the Stable Diffusion image generation model.
awesome-generative-ai
A curated list of Generative AI tools, works, models, and references
Ollama
Get up and running with large language models locally.
llm-universe
本项目是一个面向小白开发者的大模型应用开发教程,在线阅读地址:https://datawhalechina.github.io/llm-universe/
Prompt Engineering Guide
Guide and resources for prompt...
Forefront
A Better ChatGPT...
Best For
- ✓global learners seeking prompt engineering education in non-English languages
- ✓teams building multilingual AI education platforms
- ✓content creators translating technical AI documentation
- ✓ML engineers designing prompt-based systems and need to choose techniques
- ✓researchers studying LLM prompting methodologies and their comparative effectiveness
- ✓developers implementing advanced prompting patterns in production applications
- ✓teams with domain-specific tasks where fine-tuning could improve performance
- ✓organizations evaluating whether to invest in fine-tuning vs. prompt engineering
Known Limitations
- ⚠Translation maintenance requires manual updates across 11 language files when core content changes
- ⚠Language detection relies on browser Accept-Language headers; no persistent user language preference storage without external state
- ⚠Static site generation means language-specific content must be pre-built; real-time language switching requires page reload
- ⚠Diagrams are static PNG images; no interactive visualization of technique execution flows
- ⚠Documentation is descriptive rather than prescriptive—lacks automated technique selection based on task characteristics
- ⚠No built-in benchmarking data showing technique performance across different model families and task types
Requirements
Input / Output
UnfragileRank
UnfragileRank is computed from adoption signals, documentation quality, ecosystem connectivity, match graph feedback, and freshness. No artifact can pay for a higher rank.
About
Guide and resources for prompt engineering.
Categories
Alternatives to Prompt Engineering Guide
Are you the builder of Prompt Engineering Guide?
Claim this artifact to get a verified badge, access match analytics, see which intents users search for, and manage your listing.
Get the weekly brief
New tools, rising stars, and what's actually worth your time. No spam.
Data Sources
Looking for something else?
Search →