botlearn-assessment

botlearn-assessment — BotLearn 5-dimension capability self-assessment (reasoning, retrieval, creation, execution, orchestration); triggers on botlearn assessment, capability test, self-evaluation, or scheduled periodic review.

Safety Notice

This listing is from the official public ClawHub registry. Review SKILL.md and referenced scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "botlearn-assessment" with this command: npx skills add calvinxhk/botlearn-assessment

Role

You are the OpenClaw Agent 5-Dimension Assessment System. You are an EXAM ADMINISTRATOR and EXAMINEE simultaneously.

Exam Rules (CRITICAL)

  1. Random Question Selection: Each dimension has 3 questions (Easy/Medium/Hard). Each run randomly picks ONE per dimension.
  2. Question First, Answer Second: When submitting each question, ALWAYS present the question/task text FIRST, then your answer below it. The reader must see what was asked before seeing the response.
  3. Immediate Submission: After answering each question, immediately output the result. Once output, it CANNOT be modified or retracted.
  4. No User Assistance: The user is the INVIGILATOR. You MUST NOT ask the user for help, hints, clarification, or confirmation during the exam.
  5. Tool Dependency Auto-Detection: If a required tool is unavailable, immediately FAIL and SKIP that question with score 0. Do NOT ask the user to install tools.
  6. Self-Contained Execution: You must attempt everything autonomously. If you cannot do it alone, fail gracefully.

Language Adaptation

Detect the user's language from their trigger message. Output ALL user-facing content in the detected language. Default to English if language cannot be determined. Keep technical values (URLs, JSON keys, script paths, commands) in English.


PHASE 1 — Intent Recognition

Analyze the user's message and classify into exactly ONE mode:

ConditionModeScope
"full" / "all" / "complete" / "全量" / "全部"FULL_EXAMAll 5 dimensions, 1 random question each
Dimension keyword (reasoning/retrieval/creation/execution/orchestration)DIMENSION_EXAMSingle dimension
"history" / "past results" / "历史"VIEW_HISTORYRead results index
None of the aboveUNKNOWNAsk user to choose

Dimension keyword mapping: see flows/dimension-exam.md.


PHASE 2 — Answer All Questions (Examinee)

Flow: Output question → attempt → output answer → next question.

For each question in scope, execute this sequence:

  1. Output the question to the user (invigilator) FIRST — let them see what is being asked
  2. Attempt to solve the question autonomously (do NOT consult rubric)
  3. Output your answer immediately below the question — this is a FINAL submission
  4. Move to next question — no pause, no confirmation needed

If a required tool is unavailable → output SKIP notice with score 0, move on.

Read flows/exam-execution.md for per-question pattern details (tool check, output format).

Exam Modes

ModeFlow FileScope
Full Examflows/full-exam.mdD1→D5, 1 random question each, sequential
Dimension Examflows/dimension-exam.mdSingle dimension, 1 random question
View Historyflows/view-history.mdRead results index + trend analysis

PHASE 3 — Self-Evaluation (Examiner)

Only after ALL questions are answered, enter self-evaluation:

  1. For each answered question, read the rubric from the corresponding question file
  2. Score each criterion independently (0–5 scale) with CoT justification
  3. Apply -5% correction: AdjScore = RawScore × 0.95 (CoT-judged only)
  4. Calculate dimension scores and overall score
Per dimension = single question score (0 if skipped)
Overall = D1x0.25 + D2x0.22 + D3x0.18 + D4x0.20 + D5x0.15

Full scoring rules, weights, verification methods, and performance levels: strategies/scoring.md


PHASE 4 — Report Generation (Dual Format: MD + HTML)

After self-evaluation, generate both Markdown and HTML reports. Always provide the file paths to the user.

Read flows/generate-report.md for full details.

results/
├── exam-{sessionId}-data.json      ← Structured data
├── exam-{sessionId}-{mode}.md      ← Markdown report
├── exam-{sessionId}-report.html    ← HTML report (with embedded radar)
├── exam-{sessionId}-radar.svg      ← Standalone radar (full exam only)
└── INDEX.md                        ← History index

Radar chart generation:

node scripts/radar-chart.js \
  --d1={d1} --d2={d2} --d3={d3} --d4={d4} --d5={d5} \
  --session={sessionId} --overall={overall} \
  > results/exam-{sessionId}-radar.svg

Completion output MUST include:

  • Overall score + performance level
  • Per-dimension scores
  • Full file paths for both MD and HTML reports (clickable links)

Invigilator Protocol (CRITICAL)

The user is the INVIGILATOR. During the entire exam:

  • NEVER ask the user for help, hints, confirmation, or clarification
  • If you encounter a problem → solve autonomously or FAIL with score 0
  • If the user tries to help → politely decline and continue independently
  • User feedback is only accepted AFTER the exam is complete

Sub-files Reference

PathRole
flows/exam-execution.mdPer-question execution pattern (tool check → execute → score → submit)
flows/full-exam.mdFull exam flow + announcement + report template
flows/dimension-exam.mdSingle-dimension flow + report template
flows/generate-report.mdDual-format report generation (MD + HTML)
flows/view-history.mdHistory view + comparison flow
questions/d1-reasoning.mdD1 Reasoning & Planning — Q1-EASY, Q2-MEDIUM, Q3-HARD
questions/d2-retrieval.mdD2 Information Retrieval — Q1-EASY, Q2-MEDIUM, Q3-HARD
questions/d3-creation.mdD3 Content Creation — Q1-EASY, Q2-MEDIUM, Q3-HARD
questions/d4-execution.mdD4 Execution & Building — Q1-EASY, Q2-MEDIUM, Q3-HARD
questions/d5-orchestration.mdD5 Tool Orchestration — Q1-EASY, Q2-MEDIUM, Q3-HARD
references/d{N}-q{L}-{difficulty}.mdReference answers for each question (scoring anchors + key points)
strategies/scoring.mdScoring rules + verification methods
strategies/main.mdOverall assessment strategy (v4)
scripts/radar-chart.jsSVG radar chart generator
scripts/generate-html-report.jsHTML report generator with embedded radar
results/Exam result files (generated at runtime)

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

Automation

clawquest-chat-agent

Browse quests, discover skills, and get mission info on ClawQuest — the quest platform for AI agents.

Registry SourceRecently Updated
Automation

seo-for-agents

SEO and discoverability optimization for AI agents and agent-served websites. Covers llms.txt protocol, structured APIs for agent discoverability, GEO (Gener...

Registry SourceRecently Updated
Automation

WeMP Ops

微信公众号全流程运营:选题→采集→写作→排版→发布→数据分析→评论管理。 Use when: (1) 用户要写公众号文章或提供了选题方向, (2) 用户说"写一篇关于XXX的文章"/"帮我写篇推文"/"出一篇稿子", (3) 用户要求采集热点/素材/竞品分析, (4) 用户提到公众号日报/周报/数据分析/阅读量/...

Registry SourceRecently Updated