alicloud-skill-creator

安装量: 78
排名: #10024

安装

npx skills add https://github.com/cinience/alicloud-skills --skill alicloud-skill-creator

Category: tool Alibaba Cloud Skill Creator Repository-specific skill engineering workflow for alicloud-skills . Use this skill when Creating a new skill under skills/ . Importing an external skill and adapting it to this repository. Updating skill trigger quality ( name and description in frontmatter). Adding or fixing smoke tests under tests/ . Running structured benchmark loops before merge. Do not use this skill when The user only needs to execute an existing product skill. The task is purely application code under apps/ with no skill changes. Repository constraints (must enforce) Skills live under skills//// . Skill folder names use kebab-case and should start with alicloud- . Every skill must include SKILL.md frontmatter with name and description . skills//SKILL.md content must stay English-only. Smoke tests must be in tests///-test/SKILL.md . Generated evidence goes to output// only. If skill inventory changes, refresh README index with scripts/update_skill_index.sh . Standard deliverable layout skills//// ├── SKILL.md ├── agents/openai.yaml ├── references/ │ └── sources.md └── scripts/ (optional) tests///-test/ └── SKILL.md Workflow Capture intent Confirm domain/subdomain and target skill name. Confirm whether this is new creation, migration, or refactor. Confirm expected outputs and success criteria. Implement skill changes For new skills: scaffold structure and draft SKILL.md + agents/openai.yaml . For migration from external repo: copy full source tree first, then adapt. Keep adaptation minimal but explicit: Replace environment-specific instructions that do not match this repo. Add repository validation and output discipline sections. Keep reusable bundled resources ( scripts/ , references/ , assets/ ). Add smoke test Create or update tests//-test/SKILL.md . Keep it minimal, reproducible, and low-risk. Include exact pass criteria and evidence location. Validate locally Run script compile validation for the skill: python3 tests/common/compile_skill_scripts.py \ --skill-path skills/ < domain

/ < subdomain

/ < skill-name

\ --output output/ < skill-name

-test/compile-check.json Refresh skill index when inventory changed: scripts/update_skill_index.sh Confirm index presence: rg -n "" README.md README.zh-CN.md README.zh-TW.md Optional broader checks: make test make build-cli Benchmark loop (optional, for major skills) If the user asks for quantitative skill evaluation, reuse bundled tooling: scripts/run_eval.py scripts/aggregate_benchmark.py eval-viewer/generate_review.py Prefer placing benchmark artifacts in a sibling workspace directory and keep per-iteration outputs. Definition of done Skill path and naming follow repository conventions. Frontmatter is complete and trigger description is explicit. Test skill exists and has objective pass criteria. Validation artifacts are saved under output/ . README skill index is refreshed if inventory changed. References references/schemas.md references/sources.md

返回排行榜