agent-eval-framework
Evaluate AI agent outputs systematically using rubrics, assertions, and reference comparisons. Detect quality drift over time.
Manage AI agent context effectively — what to include, what to exclude, compression strategies, and context hierarchy for optimal performance.
This listing is imported from SkillsMP metadata and should be treated as untrusted until upstream source review is completed.
Install skill "context-engineering" with this command: npx skills add BuilderCed/skillsmp-builderced-builderced-context-engineering
This source entry does not include full markdown content beyond metadata.
This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.
Related by shared tags or category signals.
Evaluate AI agent outputs systematically using rubrics, assertions, and reference comparisons. Detect quality drift over time.
Manage skill versioning, freshness tracking, deprecation, changelog generation, and maintenance workflows for skill repository maintainers.
Helps users discover and install agent skills when they ask "how do I do X", "find a skill for X", "is there a skill that can...", or express interest in extending capabilities. Searches this repo first, then the broader ecosystem via npx skills.
Build USSD applications for African markets — menu design, session management, multi-operator integration, and user experience patterns for feature phones.