Agent skill
code-implementation-ashcastelinocs124-job-applying
Stars
163
Forks
31
Install this agent skill to your Project
npx add-skill https://github.com/majiayu000/claude-skill-registry/tree/main/skills/development/code-implementation-ashcastelinocs124-job-applying
SKILL.md
Code Implementation Skill
Description: Deliver production-quality code for a requested task: plan → implement → verify → code review. Usage: /code-implementation "task description"
Trigger this skill when:
- User asks to implement/build/add code or a feature
- User requests a function/module/service to be written
- Bug fixes that require non-trivial coding (after root cause is known)
Skip for: Pure analysis, docs-only tasks, minor tweaks that don't need a plan
Execution Workflow
Phase 0: Architecture Context (if applicable)
When invoked from /system-arch:
- Review the approved architecture decision
- Extract specific implementation requirements:
- Components to create/modify
- Interface contracts and APIs
- Data flow requirements
- Success metrics from architecture plan
- Ensure test strategy aligns with architectural requirements
- Note any architectural constraints that affect implementation
Phase 1: Understand & Bound
- Capture requirements and constraints (inputs/outputs, interfaces, performance, edge cases)
- Identify affected components/files
- Note non-functional needs (tests, types, logging, telemetry)
Phase 2: Plan (checklist-driven)
Produce a concise plan before coding:
- Scope & assumptions
- Files to touch (including test files)
- Steps to implement (ordered)
- Test cases required (unit, integration, edge cases)
- Tests/verification commands
- Rollback notes
Use this template:
markdown
## Implementation Plan
- Scope/assumptions: ...
- Files: [...]
- Test Files: [...] (MANDATORY: list all test files to create/modify)
- Test Cases Required:
* Unit tests: [...]
* Integration tests: [...]
* Edge cases: [...]
- Steps:
1) ...
2) Write tests for [component] (BEFORE implementation)
3) Implement [component]
4) ...
- Tests/verify: `...`
- Rollback: ...
Only start coding after the plan is written. Keep the plan updated if scope changes.
CRITICAL: Test-Driven Approach
- Write test cases BEFORE implementing each component when possible
- Every new function/module MUST have corresponding tests
- Aim for >80% code coverage on new code
Phase 3: Implement
- Follow the plan steps in order; keep changes minimal and focused
- Prefer existing patterns; match project style
- Test-first workflow:
- Write test case for a component/function
- Run test (expect it to fail - "red")
- Implement the component/function
- Run test again (expect it to pass - "green")
- Refactor if needed while keeping tests green
- Run
lsp_diagnosticson changed files after each logical unit - Ensure test files are created/updated for ALL new code
Test Coverage Requirements:
- Unit tests for all new functions/methods
- Integration tests for cross-component interactions
- Edge case tests (null, empty, boundary values)
- Error handling tests (exceptions, failures)
Phase 4: Verify (COMPREHENSIVE)
- Run ALL planned tests/commands
- Check test coverage metrics (if available)
- Verify tests cover:
- Happy path scenarios
- Error conditions
- Edge cases
- Integration points
- If ANY failures occur, fix and re-run until ALL tests green
- Document test results in summary
Phase 5: Code Review (MANDATORY)
- Invoke the code-reviewer agent (see AGENTS.md) after implementation & self-verification
- Address review findings; rerun verification if code changes
Phase 6: Summarize
- Summarize changes, tests run, and status vs plan
- Call out any deviations or TODOs
Phase 7: Explain (only if user asks)
- If the user requests an explanation of the new code, use the Explain skill (/explain) or the tutor agent to provide a concise walkthrough (what it does, how it works, key flows)
Quality Guidelines
ALWAYS:
- Plan before you code; keep it short but specific
- Update the plan if the scope shifts
- Keep diffs tight; avoid drive-by refactors
- Add/maintain tests; keep
lsp_diagnosticsclean - Use existing patterns and conventions
- Run code-reviewer agent before considering the task done
NEVER:
- Start coding without a plan
- Skip writing tests for new code
- Write implementation before tests (when TDD is feasible)
- Mark tests as TODO or skip tests
- Ignore failing tests or diagnostics
- Ship code with <80% test coverage
- Write tests after the fact without running them
- Introduce type suppressions (
# type: ignore,as any,@ts-ignore) - Mix unrelated refactors with the requested change
Verification Checklist (tick before closing)
- Plan created with test cases identified
- Plan followed/updated throughout implementation
- Test files created/modified for ALL new code
- Unit tests written and passing
- Integration tests written and passing
- Edge case tests written and passing
- Error handling tests written and passing
- Test coverage >80% on new code (if measurable)
- All tests/commands executed and passing
- lsp_diagnostics clean on changed files
- code-reviewer agent run and feedback addressed
- Summary delivered to user with test results
Didn't find tool you were looking for?