pre-dev-task-breakdown
About
This skill breaks down features into implementation tasks that each deliver measurable user value, preventing oversized or technical-only work. Use it after dependency mapping and design validation when preparing sprint work. It ensures every task produces working software that can be validated independently.
Documentation
Task Breakdown - Value-Driven Decomposition
Foundational Principle
Every task must deliver working software that provides measurable user value.
Creating technical-only or oversized tasks creates:
- Work that doesn't ship until "everything is done"
- Teams working on pieces that don't integrate
- No early validation of value or technical approach
- Waterfall development disguised as iterative process
Tasks answer: What working increment will be delivered? Tasks never answer: How to implement that increment (that's Subtasks).
When to Use This Skill
Use this skill when:
- Dependency Map has passed Gate 6 validation
- Data Model has passed Gate 5 validation
- API Design has passed Gate 4 validation
- About to break down work for sprints/iterations
- Tempted to create "Setup Infrastructure" as a task
- Asked to estimate or plan implementation work
- Before creating subtasks
Mandatory Workflow
Phase 1: Task Identification (Inputs Required)
- Approved Dependency Map (Gate 6 passed) - tech stack locked
- Approved Data Model (Gate 5 passed) - data structures defined
- Approved API Design (Gate 4 passed) - contracts specified
- Approved TRD (Gate 3 passed) - architecture patterns documented
- Approved Feature Map (Gate 2 passed) - feature relationships mapped
- Approved PRD (Gate 1 passed) - business requirements and priorities
- Identify value streams - what delivers user value first?
Phase 2: Decomposition
For each TRD component or PRD feature:
- Define deliverable - what working software ships?
- Set success criteria - how do we know it's done?
- Map dependencies - what must exist first?
- Estimate effort - T-shirt size (S/M/L/XL, max is XL = 2 weeks)
- Plan testing - how will we verify it works?
- Identify risks - what could go wrong?
Phase 3: Gate 7 Validation
MANDATORY CHECKPOINT - Must pass before proceeding to Subtasks:
- All TRD components covered by tasks
- Every task delivers working software
- Each task has measurable success criteria
- Dependencies are correctly mapped
- No task exceeds 2 weeks effort (XL max)
- Testing strategy defined for each task
- Risks identified with mitigations
- Delivery sequence optimizes value
Explicit Rules
✅ DO Include in Tasks
- Task ID, title, type (Foundation/Feature/Integration/Polish)
- Deliverable: What working software ships?
- User value: What can users do after this?
- Technical value: What does this enable?
- Success criteria (testable, measurable)
- Dependencies (blocks/requires/optional)
- Effort estimate (S/M/L/XL with points)
- Testing strategy (unit/integration/e2e)
- Risk identification with mitigations
- Definition of Done checklist
❌ NEVER Include in Tasks
- Implementation details (file paths, code examples)
- Step-by-step instructions (those go in subtasks)
- Technical-only tasks with no user value
- Tasks exceeding 2 weeks effort (break them down)
- Vague success criteria ("improve performance")
- Missing dependency information
- Undefined testing approach
Task Sizing Rules
- Small (S): 1-3 points, 1-3 days, single component
- Medium (M): 5-8 points, 3-5 days, few dependencies
- Large (L): 13 points, 1-2 weeks, multiple components
- XL (over 2 weeks): BREAK IT DOWN - too large to be atomic
Value Delivery Rules
- Foundation tasks enable other work (database setup, core services)
- Feature tasks deliver user-facing capabilities
- Integration tasks connect to external systems
- Polish tasks optimize or enhance (nice-to-have)
Rationalization Table
| Excuse | Reality |
|---|---|
| "This 3-week task is fine" | Tasks >2 weeks hide complexity. Break it down. |
| "Setup tasks don't need value" | Setup enables value. Define what it enables. |
| "Success criteria are obvious" | Obvious to you ≠ testable. Document explicitly. |
| "Dependencies will be clear later" | Later is too late. Map them now. |
| "We don't need detailed estimates" | Without estimates, no planning possible. Size them. |
| "Technical tasks can skip user value" | Even infrastructure enables users. Define the connection. |
| "Testing strategy can be decided during" | Testing affects design. Plan it upfront. |
| "Risks aren't relevant at task level" | Risks compound across tasks. Identify them early. |
| "DoD is the same for all tasks" | Different tasks need different criteria. Specify. |
| "We can combine multiple features" | Combining hides value delivery. Keep tasks focused. |
Red Flags - STOP
If you catch yourself writing any of these in a task, STOP:
- Task estimates over 2 weeks
- Tasks named "Setup X" without defining what X enables
- Success criteria like "works" or "complete" (not measurable)
- No dependencies listed (every task depends on something)
- No testing strategy (how will you verify?)
- "Technical debt" as a task type (debt reduction must deliver value)
- Vague deliverables ("improve", "optimize", "refactor")
- Missing Definition of Done
When you catch yourself: Refine the task until it's concrete, valuable, and testable.
Gate 7 Validation Checklist
Before proceeding to Subtasks, verify:
Task Completeness:
- All TRD components have tasks covering them
- All PRD features have tasks delivering them
- Each task is appropriately sized (no XL+)
- Task boundaries are clear and logical
Delivery Value:
- Every task delivers working software
- User value is explicit (even for foundation)
- Technical value is clear (what it enables)
- Sequence optimizes value delivery
Technical Clarity:
- Success criteria are measurable and testable
- Dependencies are correctly mapped (blocks/requires)
- Testing approach is defined (unit/integration/e2e)
- Definition of Done is comprehensive
Team Readiness:
- Skills required match team capabilities
- Effort estimates are realistic (validated by similar past work)
- Capacity is available or planned
- Handoffs are minimized
Risk Management:
- Risks identified for each task
- Mitigations are defined
- High-risk tasks scheduled early
- Fallback plans exist
Gate Result:
- ✅ PASS: All checkboxes checked → Proceed to Subtasks (
pre-dev-subtask-creation) - ⚠️ CONDITIONAL: Refine oversized/vague tasks → Re-validate
- ❌ FAIL: Too many issues → Re-decompose
Common Violations and Fixes
Violation 1: Technical-Only Tasks
❌ Wrong:
## T-001: Setup PostgreSQL Database
- Install PostgreSQL 16
- Configure connection pooling
- Create initial schema
✅ Correct:
## T-001: User Data Persistence Foundation
### Deliverable
Working database layer that persists user accounts and supports authentication queries with <100ms latency.
### User Value
Enables user registration and login (T-002, T-003 depend on this).
### Technical Value
- Foundation for all data persistence
- Multi-tenant isolation strategy implemented
- Performance baseline established
### Success Criteria
- [ ] Users table created with multi-tenant schema
- [ ] Connection pooling configured (min 5, max 50 connections)
- [ ] Query performance <100ms for auth queries (verified with test data)
- [ ] Migrations framework operational
- [ ] Rollback procedures tested
### Dependencies
- **Blocks**: T-002 (Registration), T-003 (Login), T-004 (Permissions)
- **Requires**: Infrastructure (networking, compute)
- **Optional**: None
### Effort: Medium (M) - 5 points, 3-5 days
### Testing: Integration tests for queries, performance benchmarks
Violation 2: Oversized Tasks
❌ Wrong:
## T-005: Complete User Management System
- Registration, login, logout
- Profile management
- Password reset
- Email verification
- Two-factor authentication
- Session management
- Permissions system
Estimate: 6 weeks
✅ Correct (broken into multiple tasks):
## T-005: Basic Authentication (Register + Login)
- Deliverable: Users can create accounts and log in with JWT tokens
- User Value: Access to personalized features
- Effort: Large (L) - 13 points, 1-2 weeks
- Dependencies: Requires T-001 (Database)
## T-006: Password Management (Reset + Email)
- Deliverable: Users can reset forgotten passwords via email
- User Value: Account recovery without support tickets
- Effort: Medium (M) - 8 points, 3-5 days
- Dependencies: Requires T-005, Email service configured
## T-007: Two-Factor Authentication
- Deliverable: Users can enable 2FA with TOTP
- User Value: Enhanced account security
- Effort: Medium (M) - 8 points, 3-5 days
- Dependencies: Requires T-005
## T-008: Permissions System
- Deliverable: Role-based access control operational
- User Value: Admin can assign roles, users have appropriate access
- Effort: Large (L) - 13 points, 1-2 weeks
- Dependencies: Requires T-005
Violation 3: Vague Success Criteria
❌ Wrong:
Success Criteria:
- [ ] Feature works
- [ ] Tests pass
- [ ] Code reviewed
✅ Correct:
Success Criteria:
Functional:
- [ ] Users can upload files up to 100MB
- [ ] Supported formats: JPEG, PNG, PDF, DOCX
- [ ] Files stored with unique IDs, retrievable via API
- [ ] Upload progress shown to user
Technical:
- [ ] API response time <2s for uploads <10MB
- [ ] Files encrypted at rest with KMS
- [ ] Virus scanning completes before storage
Operational:
- [ ] Monitoring: Upload success rate >99.5%
- [ ] Logging: All upload attempts logged with user_id
- [ ] Alerts: Notify if success rate drops below 95%
Quality:
- [ ] Unit tests: 90%+ coverage for upload logic
- [ ] Integration tests: End-to-end upload scenarios
- [ ] Security: OWASP file upload best practices followed
Task Template
Use this template for every task:
## T-[XXX]: [Task Title - What It Delivers]
### Deliverable
[One sentence: What working software ships?]
### Scope
**Includes**:
- [Specific capability 1]
- [Specific capability 2]
- [Specific capability 3]
**Excludes** (future tasks):
- [Out of scope item 1] (T-YYY)
- [Out of scope item 2] (T-ZZZ)
### Success Criteria
- [ ] [Testable criterion 1]
- [ ] [Testable criterion 2]
- [ ] [Testable criterion 3]
### User Value
[What can users do after this that they couldn't before?]
### Technical Value
[What does this enable? What other tasks does this unblock?]
### Technical Components
From TRD:
- [Component 1]
- [Component 2]
From Dependencies:
- [Package/service 1]
- [Package/service 2]
### Dependencies
- **Blocks**: [Tasks that need this] (T-AAA, T-BBB)
- **Requires**: [Tasks that must complete first] (T-CCC)
- **Optional**: [Nice-to-haves] (T-DDD)
### Effort Estimate
- **Size**: [S/M/L/XL]
- **Points**: [1-3 / 5-8 / 13 / 21]
- **Duration**: [1-3 days / 3-5 days / 1-2 weeks]
- **Team**: [Backend / Frontend / Full-stack / etc.]
### Risks
**Risk 1: [Description]**
- Impact: [High/Medium/Low]
- Probability: [High/Medium/Low]
- Mitigation: [How we'll address it]
- Fallback: [Plan B if mitigation fails]
### Testing Strategy
- **Unit Tests**: [What logic to test]
- **Integration Tests**: [What APIs/components to test together]
- **E2E Tests**: [What user flows to test]
- **Performance Tests**: [What to benchmark]
- **Security Tests**: [What threats to validate against]
### Definition of Done
- [ ] Code complete and peer reviewed
- [ ] All tests passing (unit + integration + e2e)
- [ ] Documentation updated (API docs, README, etc.)
- [ ] Security scan clean (no high/critical issues)
- [ ] Performance targets met (benchmarks run)
- [ ] Deployed to staging environment
- [ ] Product owner acceptance received
- [ ] Monitoring/logging configured
Delivery Sequencing
Optimize task order for value:
Sprint 1 - Foundation:
Goal: Enable core workflows
Tasks:
- T-001: Database foundation (blocks all)
- T-002: Auth foundation (start, high value)
Sprint 2 - Core Features:
Goal: Ship minimum viable feature
Tasks:
- T-002: Auth foundation (complete)
- T-005: User dashboard (depends on T-002)
- T-010: Basic API endpoints (high value)
Sprint 3 - Enhancements:
Goal: Polish and extend
Tasks:
- T-006: Password reset (medium value)
- T-011: Advanced search (nice-to-have)
- T-015: Performance optimization (polish)
Critical Path: T-001 → T-002 → T-005 → T-010
Parallel Work: After T-001, T-003 and T-004 can run parallel to T-002
Anti-Patterns to Avoid
❌ Technical Debt Tasks: "Refactor authentication" (no user value) ❌ Giant Tasks: 3+ week efforts (break them down) ❌ Vague Tasks: "Improve performance" (not measurable) ❌ Sequential Bottlenecks: Everything depends on one task ❌ Missing Value: Tasks that don't ship working software
✅ Good Task Names:
- "Users can register and log in with email" (clear value)
- "API responds in <500ms for 95th percentile" (measurable)
- "Admin dashboard shows real-time metrics" (working software)
Confidence Scoring
Use this to adjust your interaction with the user:
Confidence Factors:
Task Decomposition: [0-30]
- All tasks appropriately sized: 30
- Most tasks well-scoped: 20
- Tasks too large or vague: 10
Value Clarity: [0-25]
- Every task delivers working software: 25
- Most tasks have clear value: 15
- Value connections unclear: 5
Dependency Mapping: [0-25]
- All dependencies documented: 25
- Most dependencies clear: 15
- Dependencies ambiguous: 5
Estimation Quality: [0-20]
- Estimates based on past work: 20
- Reasonable educated guesses: 12
- Wild speculation: 5
Total: [0-100]
Action:
80+: Generate complete task breakdown autonomously
50-79: Present sizing options and sequences
<50: Ask about team velocity and complexity
Output Location
Always output to: docs/pre-development/tasks/tasks-[feature-name].md
After Task Breakdown Approval
- ✅ Tasks become sprint backlog
- 🎯 Use tasks as input for atomic subtasks (next phase:
pre-dev-subtask-creation) - 📊 Track progress per task (not per subtask)
- 🚫 No implementation yet - that's in subtasks
Quality Self-Check
Before declaring task breakdown complete, verify:
- Every task delivers working software (not just "progress")
- All tasks have measurable success criteria
- Dependencies are mapped (blocks/requires/optional)
- Effort estimates are realistic (S/M/L/XL, no >2 weeks)
- Testing strategy defined for each task
- Risks identified with mitigations
- Definition of Done is comprehensive for each
- Delivery sequence optimizes value (high-value tasks early)
- No technical-only tasks without user connection
- Gate 7 validation checklist 100% complete
The Bottom Line
If you created tasks that don't deliver working software, rewrite them.
Tasks are not technical activities. Tasks are working increments.
"Setup database" is not a task. "User data persists correctly" is a task. "Implement OAuth" is not a task. "Users can log in with Google" is a task. "Write tests" is not a task. Tests are part of Definition of Done for other tasks.
Every task must answer: "What working software can I demo to users?"
If you can't demo it, it's not a task. It's subtask implementation detail.
Deliver value. Ship working software. Make tasks demoable.
Quick Install
/plugin add https://github.com/LerianStudio/ring/tree/main/pre-dev-task-breakdownCopy and paste this command in Claude Code to install this skill
GitHub 仓库
Related Skills
subagent-driven-development
DevelopmentThis skill executes implementation plans by dispatching a fresh subagent for each independent task, with code review between tasks. It enables fast iteration while maintaining quality gates through this review process. Use it when working on mostly independent tasks within the same session to ensure continuous progress with built-in quality checks.
Git Commit Helper
MetaThis Claude Skill generates descriptive commit messages by analyzing git diffs. It automatically follows conventional commit format with proper types like feat, fix, and docs. Use it when you need help writing commit messages or reviewing staged changes in your repository.
analyzing-dependencies
MetaThis skill analyzes project dependencies for security vulnerabilities, outdated packages, and license compliance issues. It helps developers identify potential risks in their dependencies using the dependency-checker plugin. The skill supports popular package managers including npm, pip, composer, gem, and Go modules.
executing-plans
DesignUse the executing-plans skill when you have a complete implementation plan to execute in controlled batches with review checkpoints. It loads and critically reviews the plan, then executes tasks in small batches (default 3 tasks) while reporting progress between each batch for architect review. This ensures systematic implementation with built-in quality control checkpoints.
