Business Automation Strategy

# Business Automation Strategy — AfrexAI

Safety Notice

This item is sourced from the public archived skills repository. Treat as untrusted until reviewed.

Copy this and send it to your AI assistant to learn

Install skill "Business Automation Strategy" with this command: npx skills add 1kalin/afrexai-automation-strategy

Business Automation Strategy — AfrexAI

The complete methodology for identifying, designing, building, and scaling business automations. Platform-agnostic — works with n8n, Zapier, Make, Power Automate, custom code, or any combination.

Phase 1: Automation Audit — Find the Gold

Before building anything, map where time and money leak.

Quick ROI Triage

Ask these 5 questions about any process:

  1. How often does it happen? (frequency)
  2. How long does it take? (duration per occurrence)
  3. How many people touch it? (handoffs)
  4. How error-prone is it? (failure rate)
  5. How much does failure cost? (impact)

Process Inventory Template

process_inventory:
  process_name: "[Name]"
  department: "[Sales/Marketing/Ops/Finance/HR/Engineering]"
  owner: "[Person responsible]"
  frequency: "[X per day/week/month]"
  duration_minutes: [time per occurrence]
  monthly_volume: [total occurrences]
  monthly_hours: [volume × duration ÷ 60]
  hourly_cost: [fully loaded employee cost]
  monthly_cost: "$[hours × hourly cost]"
  error_rate: "[X%]"
  error_cost_per_incident: "$[average]"
  handoffs: [number of people involved]
  current_tools: ["tool1", "tool2"]
  automation_potential: "[Full/Partial/Assist/None]"
  complexity: "[Simple/Medium/Complex/Enterprise]"
  dependencies: ["system1", "system2"]
  notes: "[Pain points, workarounds, tribal knowledge]"

Automation Potential Classification

LevelDescriptionHuman RoleExample
FullEnd-to-end automated, no human neededMonitor exceptionsInvoice processing, data sync
PartialAutomated with human approval gatesReview & approveContract generation, hiring workflow
AssistHuman does work, automation helpsExecute with AI assistanceCustomer support, content creation
NoneRequires human judgment/creativityFull ownershipStrategy, relationship building

ROI Calculation

Annual savings = (monthly_hours × 12 × hourly_cost) + (error_rate × volume × 12 × error_cost)
Build cost = development_hours × developer_rate + tool_costs
Payback period = build_cost ÷ (annual_savings ÷ 12) months
ROI = ((annual_savings - annual_tool_cost) ÷ build_cost) × 100%

Decision rules:

  • Payback < 3 months → Build immediately
  • Payback 3-6 months → Build this quarter
  • Payback 6-12 months → Evaluate against alternatives
  • Payback > 12 months → Reconsider (unless strategic)

Phase 2: Prioritization — The Automation Stack Rank

ICE-R Scoring (0-10 each)

DimensionWeightScoring Guide
Impact30%10=saves >$50K/yr, 7=saves >$20K/yr, 5=saves >$5K/yr, 3=saves >$1K/yr
Confidence20%10=proven pattern, 7=similar done before, 5=feasible but new, 3=uncertain
Ease25%10=<1 day, 7=<1 week, 5=<1 month, 3=<3 months, 1=>3 months
Reliability25%10=deterministic, 7=95%+ success, 5=80%+ success, 3=needs frequent fixes
Score = (Impact × 0.30) + (Confidence × 0.20) + (Ease × 0.25) + (Reliability × 0.25)

Quick Win Identification

Automate FIRST (highest ROI, lowest risk):

  1. Data entry / copy-paste between systems
  2. Notification routing (email → Slack → SMS based on rules)
  3. Report generation and distribution
  4. File organization and naming
  5. Status updates across tools
  6. Meeting scheduling and follow-ups
  7. Invoice creation from templates
  8. Lead capture → CRM entry
  9. Onboarding checklists
  10. Backup and archival

Automate LAST (complex, high risk):

  1. Anything involving money transfers without approval
  2. Customer-facing responses without review
  3. Legal/compliance decisions
  4. Hiring/firing workflows
  5. Security-sensitive operations

Phase 3: Platform Selection — Choose Your Weapons

Platform Decision Matrix

FactorNo-Code (Zapier/Make)Low-Code (n8n/Power Automate)Custom CodeAI Agent
Best forSimple integrationsComplex workflowsUnique logicJudgment calls
Build speedHoursDaysWeeksDays-weeks
MaintenanceLowMediumHighMedium
FlexibilityLimitedHighUnlimitedHigh
Cost at scaleExpensiveModerateCheapVaries
Error handlingBasicGoodFull controlVariable
Team skill neededBusiness userTechnical BADeveloperAI engineer
Vendor lock-inHighMediumNoneLow-medium

Selection Decision Tree

Is the process deterministic (same input → same output)?
├── YES: Does it involve >3 systems?
│   ├── YES: Does it need complex branching logic?
│   │   ├── YES → Low-code (n8n/Power Automate)
│   │   └── NO → No-code (Zapier/Make) if budget allows, else n8n
│   └── NO: Is it performance-critical?
│       ├── YES → Custom code
│       └── NO → No-code (simplest wins)
└── NO: Does it need judgment/reasoning?
    ├── YES: Is the judgment pattern learnable?
    │   ├── YES → AI agent with human review
    │   └── NO → Human-assisted automation
    └── NO → Partial automation with human gates

Cost Comparison by Scale

Monthly TasksZapierMaken8n (self-hosted)Custom Code
1,000$30$10$5 (hosting)$50+ (hosting)
10,000$100$30$5$50+
100,000$500+$150$10$50+
1,000,000$2,000+$500+$20$100+

Rule: If you're spending >$200/mo on Zapier/Make, evaluate self-hosted n8n.


Phase 4: Workflow Architecture — Design Before You Build

Workflow Blueprint Template

workflow_blueprint:
  name: "[Descriptive name]"
  id: "WF-[DEPT]-[NUMBER]"
  version: "1.0.0"
  owner: "[Person]"
  priority: "[P0-P3]"
  
  trigger:
    type: "[webhook/schedule/event/manual/condition]"
    source: "[System or schedule]"
    conditions: "[When to fire]"
    dedup_strategy: "[How to prevent double-processing]"
  
  inputs:
    - name: "[field]"
      type: "[string/number/date/object]"
      required: true
      validation: "[rules]"
      source: "[where it comes from]"
  
  steps:
    - id: "step_1"
      action: "[verb: fetch/transform/validate/send/create/update/delete]"
      system: "[target system]"
      description: "[what this step does]"
      input: "[from trigger or previous step]"
      output: "[what it produces]"
      error_handling: "[retry/skip/alert/abort]"
      timeout_seconds: 30
    
    - id: "step_2_branch"
      type: "condition"
      condition: "[expression]"
      true_path: "step_3a"
      false_path: "step_3b"
  
  error_handling:
    retry_policy:
      max_attempts: 3
      backoff: "exponential"
      initial_delay_seconds: 5
    on_failure: "[alert/queue-for-review/fallback]"
    alert_channel: "[Slack/email/SMS]"
    dead_letter_queue: true
  
  monitoring:
    success_metric: "[what defines success]"
    expected_duration_seconds: [max]
    alert_on_duration_exceeded: true
    log_level: "[info/debug/error]"
  
  testing:
    test_data: "[how to generate test inputs]"
    expected_output: "[what success looks like]"
    edge_cases: ["empty input", "duplicate", "malformed data"]

7 Workflow Design Principles

  1. Idempotent by default — Running the same workflow twice with the same input should produce the same result, not duplicates
  2. Fail loudly — Silent failures are worse than crashes. Every error must notify someone
  3. Checkpoint progress — Long workflows should save state so they can resume, not restart
  4. Validate early — Check inputs at the start, not after 10 expensive API calls
  5. Separate concerns — One workflow, one job. Chain workflows, don't build monoliths
  6. Log everything — Timestamps, inputs, outputs, decisions. You WILL need to debug
  7. Human escape hatch — Every automated workflow needs a manual override path

Common Workflow Patterns

PatternWhen to UseExample
SequentialSteps depend on each otherLead → Enrich → Score → Route
Parallel fan-outIndependent stepsSend email + Update CRM + Log analytics
Conditional branchDifferent paths by dataHigh value → Sales, Low value → Nurture
Loop/batchProcess collectionsFor each row in CSV, create record
Approval gateHuman judgment neededContract review before sending
Event-driven chainWorkflow triggers workflowOrder placed → Fulfillment → Shipping → Notification
Retry with fallbackUnreliable external APIsTry API → Retry 3x → Use cached data → Alert
Scheduled sweepPeriodic cleanup/syncNightly: sync CRM → accounting

Phase 5: Integration Architecture — Connect Everything

Integration Quality Checklist

For every system integration:

  • API documentation reviewed
  • Authentication method confirmed (OAuth2/API key/JWT)
  • Rate limits documented (requests/min, requests/day)
  • Webhook support checked (push vs poll)
  • Error response format understood
  • Pagination handling planned
  • Data format confirmed (JSON/XML/CSV)
  • Field mapping documented
  • Test environment available
  • Sandbox/production separation configured

Data Mapping Template

data_mapping:
  source_system: "[System A]"
  target_system: "[System B]"
  sync_direction: "[one-way/bidirectional]"
  sync_frequency: "[real-time/5min/hourly/daily]"
  conflict_resolution: "[source wins/target wins/newest wins/manual]"
  
  field_mappings:
    - source_field: "contact.email"
      target_field: "customer.email_address"
      transform: "lowercase"
      required: true
    - source_field: "contact.company"
      target_field: "customer.organization"
      transform: "trim"
      default: "Unknown"
    - source_field: "contact.created_at"
      target_field: "customer.signup_date"
      transform: "ISO8601 → YYYY-MM-DD"

Rate Limit Strategy

ApproachWhenImplementation
Queue + throttlePredictable volumeProcess queue at 80% of rate limit
Exponential backoffBurst trafficWait 1s, 2s, 4s, 8s on 429 errors
Batch API callsHigh volume CRUDGroup 50-100 records per call
Cache responsesRepeated lookupsCache for TTL matching data freshness needs
Off-peak schedulingNon-urgent syncsRun heavy syncs at 2-4 AM

Phase 6: Error Handling & Reliability — Build It Unbreakable

Error Classification

TypeExampleResponsePriority
TransientAPI timeout, 503Retry with backoffAuto-handle
Rate limit429 Too Many RequestsQueue + throttleAuto-handle
Data validationMissing required fieldLog + skip + alertReview daily
Auth failureToken expiredRefresh + retry, else alertP1 — fix within 1h
Logic errorUnexpected stateHalt + alert + queueP0 — fix immediately
External changeAPI schema changedHalt + alertP0 — fix immediately
CapacityQueue overflowScale + alertP1 — fix within 4h

Dead Letter Queue Pattern

Every workflow should have a DLQ:

  1. Capture — Failed items go to DLQ with full context (input, error, timestamp, step)
  2. Alert — Notify on DLQ growth (>10 items or >1% failure rate)
  3. Review — Daily check of DLQ items
  4. Replay — Ability to reprocess DLQ items after fix
  5. Expire — Auto-archive items older than 30 days with summary

Circuit Breaker Pattern

States: CLOSED (normal) → OPEN (failing) → HALF-OPEN (testing)

CLOSED: Process normally, track failures
  → If failure_count > threshold in window → OPEN

OPEN: Reject all requests, return cached/default
  → After cool_down_period → HALF-OPEN

HALF-OPEN: Allow 1 test request
  → If success → CLOSED
  → If failure → OPEN (reset cool_down)

Thresholds:

  • Simple integrations: 5 failures in 60 seconds
  • Critical paths: 3 failures in 30 seconds
  • Non-critical: 10 failures in 300 seconds

Phase 7: Testing & Validation — Trust But Verify

Automation Test Pyramid

LevelWhatHowWhen
UnitIndividual step logicMock inputs, verify outputEvery change
IntegrationSystem connectionsTest with sandbox APIsWeekly + after changes
End-to-endFull workflow pathRun with test dataBefore deploy + weekly
ChaosFailure scenariosKill steps, corrupt dataMonthly
LoadVolume handling10x normal volumeBefore scaling

Test Scenario Checklist

For every workflow, test:

  • Happy path (normal input, expected output)
  • Empty/null input (missing required fields)
  • Duplicate input (same event twice)
  • Malformed input (wrong types, encoding issues)
  • Boundary values (max length, zero, negative)
  • API down (target system unavailable)
  • Slow response (timeout handling)
  • Partial failure (step 3 of 5 fails)
  • Concurrent execution (two runs at same time)
  • Clock skew / timezone issues
  • Large payload (oversized data)
  • Permission denied (auth issues)

Validation Before Go-Live

go_live_checklist:
  functionality:
    - [ ] All test scenarios pass
    - [ ] Edge cases documented and handled
    - [ ] Error messages are actionable
  
  reliability:
    - [ ] Retry logic tested
    - [ ] Circuit breaker configured
    - [ ] Dead letter queue active
    - [ ] Idempotency verified (run twice, same result)
  
  monitoring:
    - [ ] Success/failure alerts configured
    - [ ] Duration alerts set
    - [ ] Log retention configured
    - [ ] Dashboard created
  
  documentation:
    - [ ] Workflow blueprint updated
    - [ ] Runbook written
    - [ ] Team trained on manual override
  
  rollback:
    - [ ] Previous version preserved
    - [ ] Rollback procedure tested
    - [ ] Data cleanup plan for partial runs

Phase 8: Monitoring & Observability — See Everything

Automation Health Dashboard

automation_dashboard:
  period: "weekly"
  
  summary:
    total_workflows: [count]
    total_executions: [count]
    success_rate: "[X%]"
    avg_duration: "[X seconds]"
    errors_this_period: [count]
    time_saved_hours: [calculated]
    cost_saved: "$[calculated]"
  
  by_workflow:
    - name: "[Workflow name]"
      executions: [count]
      success_rate: "[X%]"
      avg_duration: "[X seconds]"
      p95_duration: "[X seconds]"
      errors: [count]
      error_types: ["type1: count", "type2: count"]
      dlq_items: [count]
      status: "[healthy/degraded/failing]"
  
  alerts_fired: [count]
  manual_interventions: [count]
  
  top_issues:
    - "[Issue 1: description + fix status]"
    - "[Issue 2: description + fix status]"
  
  cost:
    platform_cost: "$[monthly]"
    api_calls_cost: "$[monthly]"
    compute_cost: "$[monthly]"
    total: "$[monthly]"
    cost_per_execution: "$[calculated]"

Alert Rules

MetricWarningCriticalAction
Success rate<95%<90%Investigate + fix
Duration>2x average>5x averageCheck for bottleneck
DLQ size>10 items>50 itemsReview + reprocess
Error spike5 errors/hour20 errors/hourPause + investigate
Queue depth>100 pending>1000 pendingScale + investigate
Cost spike>150% of average>300% of averageAudit + optimize

Weekly Review Questions

  1. Which workflows had the lowest success rate? Why?
  2. Are any workflows consistently slow? What's the bottleneck?
  3. How many manual interventions were needed? Can we eliminate them?
  4. What's in the DLQ? Patterns?
  5. Are we approaching any rate limits?
  6. Total cost vs total time saved — still positive ROI?

Phase 9: Scaling & Optimization — Go From 10 to 10,000

Scaling Checklist

Before scaling any automation:

  • Load tested at 10x current volume
  • Rate limits mapped for all APIs
  • Queue-based architecture (not synchronous chains)
  • Database indexes optimized
  • Caching layer in place
  • Monitoring alerts adjusted for new thresholds
  • Cost projections at scale calculated
  • Fallback/degradation plan documented

Performance Optimization Priority

  1. Eliminate unnecessary API calls — Cache lookups, batch operations
  2. Parallelize independent steps — Don't wait when you don't have to
  3. Optimize data payloads — Only fetch/send fields you need
  4. Use webhooks over polling — Real-time + fewer API calls
  5. Batch processing — Group operations (50-100 per batch)
  6. Async where possible — Don't block on non-critical steps
  7. CDN/cache for static lookups — Country codes, categories, templates
  8. Database query optimization — Indexes, query plans, connection pooling

When to Migrate Platforms

SignalFromTo
Spending >$500/mo on Zapier/MakeNo-codeSelf-hosted n8n
Need custom logic in >50% of workflowsNo-codeLow-code or code
>100K executions/dayAny hostedSelf-hosted or custom
Complex branching breaking visual toolsLow-codeCustom code
Multiple teams building automationsSingle toolPlatform + governance
AI judgment needed in workflowsTraditionalAI agent integration

Phase 10: Governance & Documentation — Keep It Manageable

Automation Registry

Every automation must be registered:

automation_registry_entry:
  id: "WF-[DEPT]-[NUMBER]"
  name: "[Descriptive name]"
  description: "[What it does in one sentence]"
  owner: "[Person]"
  team: "[Department]"
  platform: "[n8n/Zapier/Make/custom]"
  status: "[active/paused/deprecated/testing]"
  created: "[date]"
  last_modified: "[date]"
  last_reviewed: "[date]"
  review_frequency: "[monthly/quarterly]"
  
  business_impact:
    time_saved_monthly_hours: [X]
    cost_saved_monthly: "$[X]"
    error_reduction: "[X%]"
    
  technical:
    trigger: "[type]"
    systems_connected: ["system1", "system2"]
    avg_daily_executions: [X]
    success_rate: "[X%]"
    
  dependencies:
    upstream: ["WF-XXX"]
    downstream: ["WF-YYY"]
    
  documentation:
    blueprint: "[link]"
    runbook: "[link]"
    test_plan: "[link]"

Naming Conventions

Pattern: [DEPT]-[ACTION]-[OBJECT]-[QUALIFIER]
Examples:
  SALES-sync-leads-from-typeform
  FINANCE-generate-invoice-monthly
  HR-onboard-employee-new-hire
  MARKETING-post-content-social-scheduled
  OPS-backup-database-nightly

Change Management for Automations

Change TypeApprovalTestingRollback Plan
Config change (threshold, timing)OwnerQuick smoke testRevert config
Logic change (new branch, new step)Owner + reviewerFull test suitePrevious version
Integration change (new API, new system)Owner + tech leadIntegration + E2EDisconnect + manual
New workflowOwner + stakeholderFull test + pilotDisable workflow
DeprecationOwner + affected teamsVerify replacementsRe-enable

Quarterly Automation Review

  1. Inventory check — Are all automations in the registry? Any rogue workflows?
  2. ROI validation — Is each automation still delivering value?
  3. Health review — Success rates, error trends, DLQ patterns
  4. Cost audit — Platform costs trending up? Optimization opportunities?
  5. Security review — API keys rotated? Permissions still appropriate?
  6. Deprecation candidates — Any automations that should be retired?
  7. Opportunity scan — New processes to automate? Existing ones to improve?

Phase 11: AI-Powered Automations — The Next Level

When to Add AI to Automations

ScenarioAI TypeExample
Classify unstructured textLLMCategorize support tickets
Extract data from documentsLLM + OCRParse invoices, contracts
Generate content from templatesLLMPersonalized emails, reports
Make judgment callsLLM + rulesLead scoring, risk assessment
Summarize informationLLMMeeting notes, research briefs
Route based on intentLLMCustomer request → right team

AI Integration Best Practices

  1. Always validate AI output — LLMs hallucinate. Add validation checks
  2. Set confidence thresholds — Below threshold → human review queue
  3. Log AI decisions — Input, output, confidence, model version
  4. A/B test AI vs rules — Prove AI adds value before committing
  5. Cost-control AI calls — Cache similar inputs, batch where possible
  6. Fallback to rules — If AI is unavailable, have deterministic backup
  7. Review AI decisions weekly — Spot check for quality drift

AI Agent Integration Pattern

ai_agent_step:
  type: "ai_judgment"
  model: "[model name]"
  
  input:
    context: "[relevant data from previous steps]"
    task: "[specific instruction — be precise]"
    output_format: "[JSON schema or structured format]"
    constraints: ["must not", "must always", "if unsure"]
  
  validation:
    confidence_threshold: 0.85
    required_fields: ["field1", "field2"]
    value_ranges:
      score: [0, 100]
      category: ["A", "B", "C"]
    
  on_low_confidence:
    action: "route_to_human"
    queue: "[review queue name]"
    
  on_failure:
    action: "fallback_to_rules"
    rules_engine: "[rule set name]"
    
  monitoring:
    log_all_decisions: true
    sample_rate_for_review: 0.10
    alert_on_confidence_drop: true

Phase 12: Automation Maturity Model

5 Levels of Automation Maturity

LevelNameDescriptionIndicators
1Ad HocManual processes, maybe a few scriptsNo registry, tribal knowledge
2ReactiveAutomate pain points as they ariseSome workflows, no standards
3SystematicPlanned automation programRegistry, testing, monitoring
4OptimizedContinuous improvement, governanceROI tracking, quarterly reviews
5IntelligentAI-augmented, self-healingAdaptive workflows, predictive

Maturity Assessment (Score 1-5 per dimension)

automation_maturity:
  dimensions:
    strategy: [1-5]  # Planned roadmap vs ad hoc
    architecture: [1-5]  # Patterns, standards, reuse
    reliability: [1-5]  # Error handling, monitoring, uptime
    governance: [1-5]  # Registry, change management, reviews
    testing: [1-5]  # Test coverage, validation, chaos
    documentation: [1-5]  # Blueprints, runbooks, training
    optimization: [1-5]  # Performance, cost, continuous improvement
    ai_integration: [1-5]  # AI-powered decisions, self-healing
  
  total: [sum ÷ 8]
  grade: "[A/B/C/D/F]"
  # A: 4.5+ | B: 3.5-4.4 | C: 2.5-3.4 | D: 1.5-2.4 | F: <1.5
  
  top_gap: "[lowest scoring dimension]"
  next_action: "[specific improvement for top gap]"

100-Point Quality Rubric

DimensionWeight0-2 (Poor)3-5 (Basic)6-8 (Good)9-10 (Excellent)
Design15%No blueprint, ad hocBasic flow documentedFull blueprint with error handlingBlueprint + edge cases + optimization
Reliability20%No error handlingBasic retriesDLQ + circuit breaker + fallbackSelf-healing + auto-scaling
Testing15%No testsHappy path onlyFull test pyramidChaos testing + load testing
Monitoring15%No visibilityBasic success/fail logsDashboard + alertsPredictive monitoring
Documentation10%NoneREADME existsBlueprint + runbookFull docs + training materials
Security10%Hardcoded credentialsEncrypted secretsLeast privilege + rotationZero-trust + audit trail
Performance10%Works but slowAcceptable speedOptimized + cachedAuto-scaling + sub-second
Governance5%No registryListed somewhereFull registry + reviewsChange management + compliance

Score: (weighted sum) → Grade: A (90+) B (80-89) C (70-79) D (60-69) F (<60)


10 Automation Killers

#MistakeFix
1Automating a broken processFix the process FIRST, then automate
2No error handlingEvery step needs a failure path
3Silent failuresIf it fails and nobody knows, it's worse than manual
4Not testing edge casesTest empty, duplicate, malformed, concurrent
5Hardcoded valuesUse config/environment variables for everything
6No monitoringYou can't fix what you can't see
7Building monolith workflowsOne workflow, one job. Chain them together
8Ignoring rate limitsDesign for API limits from day one
9No documentationFuture-you will hate present-you
10Over-automatingNot everything should be automated. Human judgment exists for a reason

Edge Cases

Small Team / Solo Founder

  • Start with Zapier/Make — speed over flexibility
  • Automate the 3 most time-consuming tasks first
  • Graduate to n8n when spending >$100/mo on no-code

Regulated Industry

  • Add approval gates at every decision point
  • Log all automated actions for audit trail
  • Review automations quarterly with compliance team
  • Document data flow for privacy impact assessments

Legacy Systems

  • Use middleware/iPaaS for legacy integration
  • Build adapters that normalize legacy data formats
  • Plan for eventual migration, not permanent workarounds

Multi-Team / Enterprise

  • Establish automation Center of Excellence (CoE)
  • Standardize on 1-2 platforms max
  • Shared component library for common patterns
  • Governance board for cross-team automations

AI-Heavy Workflows

  • Always keep human-in-the-loop for high-stakes decisions
  • Monitor AI output quality continuously
  • Budget for AI API costs separately (they scale differently)
  • Version-pin AI models — don't auto-upgrade in production

Natural Language Commands

Use these to invoke specific phases:

  1. audit my processes for automation opportunities → Phase 1
  2. prioritize automations by ROI → Phase 2
  3. recommend automation platform for [process] → Phase 3
  4. design workflow blueprint for [process] → Phase 4
  5. plan integration between [system A] and [system B] → Phase 5
  6. design error handling for [workflow] → Phase 6
  7. create test plan for [automation] → Phase 7
  8. set up monitoring for [workflow] → Phase 8
  9. optimize [workflow] for scale → Phase 9
  10. review automation governance → Phase 10
  11. add AI to [workflow] → Phase 11
  12. assess automation maturity → Phase 12

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

Automation

ai-dating

Direct dating and matchmaking workflow via curl against the dating HTTP API. Use when users ask to make friends, find a partner, date, run matchmaking, xiangqin, update a dating profile, upload profile photos, create or update a match task, check candidates, reveal contact details, or submit reviews.

Archived SourceRecently Updated
Automation

session-guardian

Never lose a conversation again. Auto-backup, smart recovery, and health monitoring for OpenClaw sessions. Protects against gateway crashes, model disconnections, and token overflow. Use this skill when: - User worries about losing conversations after gateway restart or model crash - User mentions session backup, conversation recovery, session protection, or data loss - User's agent is slow or timing out (likely token overflow from large sessions) - User runs multiple agents and needs to track collaboration across sessions - User asks about session health, backup strategy, or disaster recovery - User mentions "对话丢失", "会话备份", "上下文溢出", "token超限", "Gateway重启后记忆丢失" - Even if user just says "my agent lost everything after a restart" — this is the skill

Archived SourceRecently Updated
Automation

news-hot-scraper

This skill should be used when users need to scrape hot news topics from Chinese platforms (微博、知乎、B站、抖音、今日头条、腾讯新闻、澎湃新闻), generate summaries, and cite sources. It supports both API-based and direct scraping methods, and offers both extractive and abstractive summarization techniques.

Archived SourceRecently Updated
Automation

moltbook-interact

Interact with Moltbook — a social network for AI agents. Post, reply, browse hot posts, and track engagement. Credentials stored in ~/.config/moltbook/credentials.json.

Archived SourceRecently Updated