prompt-engineering

Prompt Engineering Guide

Safety Notice

This listing is imported from skills.sh public index metadata. Review upstream SKILL.md and repository scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "prompt-engineering" with this command: npx skills add eyadsibai/ltk/eyadsibai-ltk-prompt-engineering

Prompt Engineering Guide

Effective prompts, RAG systems, and agent workflows.

When to Use

  • Optimizing LLM prompts

  • Building RAG systems

  • Designing agent workflows

  • Creating few-shot examples

  • Structuring chain-of-thought reasoning

Prompt Structure

Core Components

Component Purpose Include When

Role/Context Set expertise, persona Complex domain tasks

Task Clear instruction Always

Format Output structure Need structured output

Examples Few-shot learning Pattern demonstration needed

Constraints Boundaries, rules Need to limit scope

Prompt Patterns

Pattern Use Case Key Concept

Chain of Thought Complex reasoning "Think step by step"

Few-Shot Pattern learning 2-5 input/output examples

Role Playing Domain expertise "You are an expert X"

Structured Output Parsing needed Specify JSON/format exactly

Self-Consistency Improve accuracy Generate multiple, vote

Chain of Thought Variants

Variant Description When to Use

Standard CoT "Think step by step" Math, logic problems

Zero-Shot CoT Just add "step by step" Quick reasoning boost

Structured CoT Numbered steps Complex multi-step

Self-Ask Ask sub-questions Research-style tasks

Tree of Thought Explore multiple paths Creative/open problems

Key concept: CoT works because it forces the model to show intermediate reasoning, reducing errors in the final answer.

Few-Shot Learning

Example Selection

Criteria Why

Representative Cover common cases

Diverse Show range of inputs

Edge cases Handle boundaries

Consistent format Teach output pattern

Number of Examples

Count Trade-off

0 (zero-shot) Less context, more creative

2-3 Good balance for most tasks

5+ Complex patterns, use tokens

Key concept: Examples teach format more than content. The model learns "how" to respond, not "what" facts to include.

RAG System Design

Architecture Flow

Query → Embed → Search → Retrieve → Augment Prompt → Generate

Chunking Strategies

Strategy Best For Trade-off

Fixed size General documents May split sentences

Sentence-based Precise retrieval Many small chunks

Paragraph-based Context preservation May be too large

Semantic Mixed content More complex

Retrieval Quality Factors

Factor Impact

Chunk size Too small = no context, too large = noise

Overlap Prevents splitting important content

Metadata filtering Narrows search space

Re-ranking Improves relevance of top-k

Hybrid search Combines keyword + semantic

Key concept: RAG quality depends more on retrieval quality than generation quality. Fix retrieval first.

Agent Patterns

ReAct Pattern

Step Description

Thought Reason about what to do

Action Call a tool

Observation Process tool result

Repeat Until task complete

Tool Design Principles

Principle Why

Single purpose Clear when to use

Good descriptions Model selects correctly

Structured inputs Reliable parsing

Informative outputs Model understands result

Error messages Guide retry attempts

Prompt Optimization

Token Efficiency

Technique Savings

Remove redundant instructions 10-30%

Use abbreviations in examples 10-20%

Compress context with summaries 50%+

Remove verbose explanations 20-40%

Quality Improvement

Technique Effect

Add specific examples Reduces errors

Specify output format Enables parsing

Include edge cases Handles boundaries

Add confidence scoring Calibrates uncertainty

Common Task Patterns

Task Key Prompt Elements

Extraction List fields, specify format (JSON), handle missing

Classification List categories, one-shot per category, single answer

Summarization Specify length, focus areas, format (bullets/prose)

Generation Style guide, length, constraints, examples

Q&A Context placement, "based only on context"

Best Practices

Practice Why

Be specific and explicit Reduces ambiguity

Provide clear examples Shows expected format

Specify output format Enables parsing

Test with diverse inputs Find edge cases

Iterate based on failures Targeted improvement

Separate instructions from data Prevent injection

Resources

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

General

document-processing

No summary provided by upstream source.

Repository SourceNeeds Review
General

stripe-payments

No summary provided by upstream source.

Repository SourceNeeds Review
General

file-organization

No summary provided by upstream source.

Repository SourceNeeds Review