Agent Skills Refiner
npx skills add https://github.com/yknothing/skills-refiner --skill skills-appreciationقم بتثبيت هذه المهارة باستخدام واجهة سطر الأوامر (CLI) وابدأ في استخدام سير عمل SKILL.md في مساحة عملك.
A skill governance toolkit for analyzing, interpreting, evaluating, and debugging agent skills systems.
Four skills across two layers:
Analysis & Interpretation — judgment and understanding:
skills-refiner — audit, refine, extract, and integrate a skill repository, single skill, or workflow frameworkskills-appreciation — interpret and explain a skill or skills system in a deep, teaching-grade styleGovernance & Observability — health and visibility:
3. skill-hygiene — evaluate health, quality, and topology of installed skills (AI judges; shell scripts collect facts)
4. skill-debug — three-layer observability: discovery diagnostics, activation tracing, effectiveness dashboards
Agent skills grow fast and degrade silently. Most skill ecosystems face two intertwined problems:
This repository addresses both:
skills-refiner and skills-appreciation handle the analysis problem — deep design audit and publishable interpretation.skill-hygiene and skill-debug handle the governance problem — topology scanning, fact collection, activation tracing, and usage analytics.Together with skill-creator (the official Claude skill-creation tool), they form a complete skill lifecycle: creation → testing → design audit → governance → observability → interpretation.
skills-refiner — design-level auditUse when the main job is to:
This skill is decision-oriented. It complements skill-creator by covering what assertion-based testing cannot reach.
skills-appreciation — teaching-grade interpretationUse when the main job is to:
This skill is interpretation-oriented. It does not force engineering-style criteria onto every target — a creative skill is judged differently from an infrastructure skill.
skill-hygiene — installed skill evaluationUse when you need to:
This skill follows the "AI judges, scripts collect" philosophy. The shell script (bin/skill-scan.sh) gathers structured facts; the AI applies expert judgment. It respects the standard skill installation model: skills installed to ~/.agents/skills/ and symlinked to agent directories are distribution links, not duplicates.
skill-debug — skill observabilityUse when you cannot tell whether a skill was discovered, loaded, or followed by the agent. Three layers:
skill-probe) — what skills can the agent see from the current working directory?skill-trace) — inject/remove lightweight canary blocks to track when skills are actually used.skill-dashboard) — usage frequency, zombie detection, context distribution, active rate.Combine with skill-hygiene for a full governance workflow: probe discovery → check usage → evaluate quality → triage.
Across all four skills:
~/.agents/skills/, symlinked to agent directories (.claude/skills/, .cursor/skills/, .codex/skills/, etc.). Symlinks are distribution links, not duplicates. Standalone project repos are not broken global skills.Analysis & Interpretation:
Governance & Observability:
Install with the skills CLI:
npx skills add yknothing/skills-refiner
Works with Claude Code, Cursor, Codex, OpenCode, and many other agents.
Analysis & Interpretation:
skills/skills-refiner/SKILL.md — audit / refine / extract / integrateskills/skills-refiner/references/skill-creator-collaboration.md — collaboration model with skill-creatorskills/skills-appreciation/SKILL.md — teaching-grade appreciation / interpretationskills/skills-appreciation/references/editorial-checklist.md — article quality checklistGovernance & Observability:
skills/skill-hygiene/SKILL.md — AI-driven skill evaluation frameworkskills/skill-hygiene/bin/skill-scan.sh — topology and fact collectorskills/skill-hygiene/tests/test-scan.sh — integration tests (17 cases)skills/skill-debug/SKILL.md — three-layer observabilityskills/skill-debug/bin/skill-probe.sh — discovery diagnosticsskills/skill-debug/bin/skill-trace.sh — activation trace injection/removalskills/skill-debug/bin/skill-dashboard.sh — effectiveness dashboardskills/skill-debug/tests/test-trace.sh — integration tests (8 cases)skills/skill-debug/tests/test-probe.sh — integration tests for discovery probeskills/skill-debug/tests/test-dashboard.sh — integration tests for dashboardSupporting materials:
examples/ — usage examples for all four skillsevals/ — evaluation rubrics, cases, and anchor judgments (9 cases, 2 rubrics)# Audit a repository
Use skills-refiner on this repository.
# Audit and integrate into another repo
Use skills-refiner, and treat yknothing/prodcraft as target_repo.
# Write an appreciation article
Use skills-appreciation on this repository. Write a deep but readable article.
# Explain a single skill
Use skills-appreciation on this skill. I want to understand why it is designed this way.
# Scan installed skills for health issues
bash ~/.agents/skills/skill-hygiene/bin/skill-scan.sh
# What skills can the agent see from here?
bash ~/.agents/skills/skill-debug/bin/skill-probe.sh
# Inject activation tracing into all skills
bash ~/.agents/skills/skill-debug/bin/skill-trace.sh --inject-dir ~/.agents/skills/
# View usage dashboard (last 30 days)
bash ~/.agents/skills/skill-debug/bin/skill-dashboard.sh
# Combined health check
bash ~/.agents/skills/skill-debug/bin/skill-probe.sh --doctor
The evals/ directory contains anchor-based evaluations for the analysis skills:
skills-refiner (cases 01–03, 08) — object identification, stage control, judgment quality, evidence disciplineskills-appreciation (cases 04–07, 09) — thesis clarity, mechanism explanation, writing quality, low "AI flavor"Cases 08–09 test the collaboration scenario with skill-creator.
The governance skills (skill-hygiene, skill-debug) are validated through integration tests that create sandboxed skill topologies and verify scanner/tracer correctness.
See CONTRIBUTING.md for development guidelines.
MIT