skillby adolfoaranaes12
risk-profile
Assess implementation risks using Probability × Impact (P×I) scoring methodology across 6 categories (Technical, Security, Performance, Data, Business, Operational). Scores risks (P×I: 1-9), develops mitigations for high-risk items (score ≥6), and prioritizes test scenarios (P0/P1/P2). Use before implementation or during quality review to identify and mitigate potential issues.
Installs: 0
Used in: 1 repos
Updated: 1d ago
$
npx ai-builder add skill adolfoaranaes12/risk-profileInstalls to .claude/skills/risk-profile/
# Risk Profile Assessment
The **risk-profile** skill assesses implementation risks before or during development using the Probability × Impact (P×I) scoring methodology. This skill identifies potential issues early, enables risk-based test prioritization, and informs quality gate decisions. By systematically evaluating risks across 6 categories (Technical, Security, Performance, Data, Business, Operational), the skill produces a risk matrix with mitigation strategies and test priorities.
The P×I methodology scores each risk on a 1-9 scale (Probability 1-3 × Impact 1-3), enabling objective risk measurement and comparison. Critical risks (score ≥7) require immediate mitigation, high risks (score 6) require testing validation, medium risks (3-5) need monitoring, and low risks (1-2) require standard precautions. This scoring directly integrates with the quality gate: risks ≥9 trigger FAIL, risks ≥6 trigger CONCERNS, providing clear quality thresholds.
The skill is most powerful when used pre-implementation (after task spec creation, before coding begins), allowing developers to design mitigations into the implementation rather than retrofitting fixes later. The risk profile generates test priorities (P0/P1/P2) based on risk scores, ensuring the highest-risk areas receive comprehensive testing. The output integrates with test-design and quality-gate skills for comprehensive quality assessment.
## When to Use This Skill
**This skill should be used when:**
- Risks need to be assessed before starting implementation (recommended timing)
- Potential issues need identification early in development cycle
- Test scenarios need to be prioritized by risk level
- Quality gate decisions need to be informed with risk data
- Mitigation strategies need to be developed for high-risk areas
**This skill is particularly valuable:**
- After task spec creation, before implementation begins (optimal timing)
- For complex or high-risk features (external APIs, security, data migrations)
- When planning comprehensive test strategy
- During quality review to validate risk mitigation
**This skill should NOT be used when:**
- Task is simple CRUD with no external dependencies (low value)
- Bug fix has clear root cause and straightforward solution
- Well-established patterns with no unknowns (minimal risk)
## Prerequisites
Before running risk-profile, ensure you have:
1. **Task specification file** with clear objective, acceptance criteria, and context
2. **Project configuration** (.claude/config.yaml) with risk threshold setting
3. **Understanding of implementation approach** (what will be built, how, with what technologies)
**Optimal timing:**
- **Pre-implementation:** Task status "Draft" or "Approved" (best time for risk assessment)
- **During development:** Task status "InProgress" (validate mitigation effectiveness)
- **Post-implementation:** Task status "Review" (inform quality gate)
## Sequential Risk Assessment Process
This skill executes through 7 sequential steps (Step 0-6). Each step must complete successfully before proceeding. The process systematically identifies risks, scores them with P×I methodology, develops mitigations, prioritizes tests, and generates a comprehensive risk profile report.
### Step 0: Load Configuration and Task Context
**Purpose:** Load project configuration and task specification to understand what will be implemented and establish the risk assessment context.
**Actions:**
1. Load project configuration from `.claude/config.yaml`:
- Extract `quality.riskScoreThreshold` (default: 6)
- Extract `quality.qualityLocation` (default: .claude/quality)
2. Read task specification file:
- Load objective and acceptance criteria
- Load context (data models, APIs, components, constraints)
- Load task breakdown (implementation steps)
3. Determine assessment mode from task status:
- Pre-implementation: Status "Draft" or "Approved"
- During development: Status "InProgress"
- Post-implementation: Status "Review"
4. Understand implementation scope and complexity
**Halt If:**
- Configuration file missing or invalid
- Task file not found or unreadable
- Task too vague to assess risks
**Output:** Configuration loaded with risk threshold, task spec loaded with ID/title, assessment mode determined, implementation scope analyzed (tasks/systems/patterns)
**See:** `references/templates.md#step-0-output` for complete format
---
### Step 1: Identify Risk Areas
**Purpose:** Brainstorm potential risks across all 6 risk categories by analyzing task complexity, technical context, acceptance criteria, and known issues.
**Actions:**
1. Analyze task complexity:
- Count tasks/subtasks (>10 = higher complexity)
- Count systems involved (>3 = higher integration risk)
- Identify new patterns vs. established patterns
- Identify unknown vs. familiar technologies
2. Review technical context:
- Data models involved
- External APIs/services
- Authentication/authorization requirements
- Database operations (migrations, complex queries)
- UI components (if applicable)
3. Check acceptance criteria for risk signals:
- Security requirements mentioned?
- Performance targets specified?
- Data migration needed?
- Complex business logic?
- User-facing changes (impact scope)?
4. Brainstorm potential risks in each category:
- **Technical:** Integration challenges, unknown APIs, complexity
- **Security:** Auth vulnerabilities, injection risks, data exposure
- **Performance:** Response time, scalability, N+1 queries, resource usage
- **Data:** Integrity issues, migration complexity, data loss potential
- **Business:** User impact scope, revenue implications, compliance
- **Operational:** Deployment complexity, monitoring gaps, rollback difficulty
5. Document 10-20 potential risks with initial categorization
**Output:** Risk areas identified with count per category (Technical/Security/Performance/Data/Business/Operational), complexity indicators (task count, system count, pattern type)
**See:** `references/templates.md#step-1-output` for complete format with examples
---
### Step 2: Score Each Risk (P×I)
**Purpose:** Systematically score each identified risk using Probability × Impact methodology to enable objective risk comparison and prioritization.
**Actions:**
For each identified risk:
1. Assess Probability (P: 1-3):
- **1 (Low):** Unlikely to occur (<20% chance) - good patterns, known approaches
- **2 (Medium):** May occur (20-60% chance) - some unknowns, moderate complexity
- **3 (High):** Likely to occur (>60% chance) - complex, many unknowns, new territory
2. Assess Impact (I: 1-3):
- **1 (Low):** Minor inconvenience, easy fix, low impact
- **2 (Medium):** Significant issue, moderate effort to fix, notable impact
- **3 (High):** Critical failure, major effort to fix, security/data loss/business impact
3. Calculate Risk Score:
- Risk Score = P × I (1-9 scale)
4. Document reasoning:
- Why this probability? (evidence, similar experiences, complexity factors)
- Why this impact? (user impact, business impact, fix difficulty)
5. Sort risks by score (highest first for reporting)
**Output:** Risks scored with P×I methodology, score distribution (critical/high/medium/low counts), highest score, quality gate impact prediction
**See:** `references/templates.md#step-2-output` for complete format and scoring examples
---
### Step 3: Develop Mitigation Strategies
**Purpose:** Create actionable mitigation strategies for all high-risk items (score ≥6) with concrete prevention, detection, and recovery actions.
**Actions:**
For each high-risk item (prioritize critical risks first):
1. Identify mitigation approach:
- **Prevention:** How to prevent risk from occurring? (design, patterns, validation)
- **Detection:** How to detect if risk occurs? (tests, monitoring, logging)
- **Recovery:** How to recover if risk occurs? (rollback, fallback, manual fix)
2. Specify concrete actions:
- What specific code/design changes?
- What tests to write? (test files, scenarios)
- What monitoring to add? (metrics, alerts)
- What documentation needed?
3. Assign to appropriate phase:
- **During Implementation:** Handle when coding (architectural decisions, validation)
- **Testing:** Validate through tests (unit, integration, E2E)
- **Deployment:** Address in deployment process (migrations, feature flags)
- **Monitoring:** Detect in production (alerts, dashboards)
4. Estimate effort:
- Minimal (<1 hour)
- Moderate (1-4 hours)
- Significant (>4 hours)
**Output:** Mitigation strategies developed for all high-risk items, critical/high risks mitigation counts, total effort estimate, phase breakdown (implementation/testing/deployment/monitoring)
**See:** `references/templates.md#step-3-output` for complete format and mitigation examples
---
### Step 4: Prioritize Test Scenarios
**Purpose:** Map risks to test priorities (P0/P1/P2) and identify must-have test scenarios for high-risk areas to ensure comprehensive validation.
**Actions:**
1. Map risks to test priorities:
- **P0 (Critical):** Risks with score ≥7 (must have before merge)
- **P1 (High):** Risks with score 5-6 (should have before merge)
- **P2 (Medium):** Risks with score 3-4 (nice to have)
- **P3 (Low):** Risks with score 1-2 (standard testing)
2. Identify must-have tests for high-risk areas:
- Security risks → Security test scenarios (injection, auth bypass, data exposure)
- Performance risks → Performance test scenarios (load tests, query analysis)
- Data risks → Data integrity test scenarios (race conditions, migrations)
- Integration risks → Integration test scenarios (external API failures)
3. Specify test scenarios for each P0/P1 risk:
- Describe test scenario (what to test, how to test)
- Specify test level (unit/integration/E2E)
- Assign priority (P0/P1/P2)
- Define expected outcome
4. Document risk-test mapping:
- Which tests validate which risks?
- What test coverage is needed?
- What scenarios would expose the risk?
**Output:** Test scenarios prioritized by risk level, P0/P1/P2 test counts, risk-test mapping complete
**See:** `references/templates.md#step-4-output` for complete format and test examples
---
### Step 5: Generate Risk Profile Report
**Purpose:** Create comprehensive risk profile report documenting all risks, scores, mitigations, and test priorities for reference during implementation and quality review.
**Actions:**
1. Load risk profile template from `.claude/templates/risk-profile.md`
2. Populate risk matrix:
- List all risks sorted by score (highest first)
- Include: #, Category, Risk, P, I, Score, Mitigation summary
3. Create high-risk summary section:
- Risks with score ≥6 with detailed mitigations
- Concrete actions, phase assignment, effort estimates
4. Document test prioritization:
- P0/P1/P2 test scenarios with risk mapping
- Test files, scenarios, expected outcomes
5. Add quality gate impact prediction:
- Will any risks trigger FAIL (score ≥9)?
- Will any risks trigger CONCERNS (score ≥6)?
- What's needed for PASS? (mitigation + testing)
6. Generate file path: `{qualityLocation}/assessments/{taskId}-risk-{YYYYMMDD}.md`
7. Write risk profile file with all sections
**Output:** Risk profile report generated at path, total risks documented, critical/high risks detailed, test priorities documented (P0/P1/P2 counts), quality gate impact prediction
**See:** `references/templates.md#step-5-output` and `#complete-risk-profile-report-template` for complete format
---
### Step 6: Present Summary to User
**Purpose:** Provide concise summary with key risk metrics, critical risks highlighted, mitigation strategies, test priorities, and clear next steps.
**Actions:**
1. Display formatted summary:
- Task metadata (ID, title, assessment date)
- Risk summary (total, critical, high, medium, low counts)
- Critical risk(s) highlighted with mitigation (if any)
- High-risk areas with mitigation summaries
- Test priorities (P0/P1 scenarios)
- Quality gate impact prediction
- Recommendation for next steps
2. Highlight critical risks (score ≥7) requiring immediate attention
3. Provide implementation guidance:
- Address critical risks first
- Implement high-risk mitigations during development
- Write P0/P1 tests to validate
4. Emit telemetry event with all metrics
**Output:** Formatted summary with task metadata, risk counts by severity, critical risks highlighted with mitigations, high-risk areas summarized, P0/P1 test priorities, quality gate impact prediction, path to PASS recommendations, next steps
**See:** `references/templates.md#step-6-user-facing-summary` for complete formatted output examples
**Execution Complete.**
---
## Risk Scoring Methodology
**Probability (P):** 1-3 scale measuring likelihood (1=<20% unlikely, 2=20-60% may occur, 3=>60% likely)
**Impact (I):** 1-3 scale measuring severity (1=minor/easy fix, 2=significant issue/moderate fix, 3=critical failure/major fix)
**Risk Score:** P × I (1-9 scale) | 9=critical immediate mitigation, 6-8=high mitigation+testing, 3-5=medium monitor, 1-2=low standard precautions
**Quality Gate Rules:** Score ≥9 → FAIL | Score ≥6 → CONCERNS | Score <6 → No auto-impact
**See:** `references/templates.md#probability-assessment-guidelines` and `#impact-assessment-guidelines` for detailed scoring criteria and examples
---
## Integration with Other Skills
**Before:** Planning skills (create-task-spec, breakdown-epic) create task spec → Task approved → Ready for risk assessment
**After - Pre-implementation:** Developer aware of risks, mitigations inform implementation approach, test priorities guide test writing
**Handoff to test-design:** Risk profile with P0/P1/P2 priorities → test-design creates detailed scenarios for high-risk areas
**Handoff to quality-gate:** Risk profile informs gate decision | Critical risks (≥7) checked for mitigation | High risks (≥6) checked for test coverage
**See:** `references/templates.md#integration-examples` for complete workflows with data flow
---
## Best Practices
Assess early (after spec, before code) | Be honest about probability (consider team experience) | Consider real impact (data loss, security, business) | Actionable mitigations (specific, phase-assigned, effort-estimated) | Risk-driven testing (high risk = high priority tests) | Continuous reassessment (if requirements change or new risks discovered)
---
## Reference Files
Detailed documentation in `references/`:
- **templates.md**: All output formats (Step 0-6), risk scoring examples, complete risk profile report template, risk category details, probability/impact guidelines, mitigation strategies, test prioritization examples, integration workflows, JSON output format
- **risk-categories.md**: Risk category definitions and examples (currently placeholder - see templates.md)
- **risk-scoring.md**: P×I methodology details (currently placeholder - see templates.md)
- **mitigation-strategies.md**: Mitigation patterns (currently placeholder - see templates.md)
- **risk-examples.md**: Risk profile examples (currently placeholder - see templates.md)
---
*Risk Profile Assessment skill - Version 2.0 - Minimal V2 Architecture*Quick Install
$
npx ai-builder add skill adolfoaranaes12/risk-profileDetails
- Type
- skill
- Author
- adolfoaranaes12
- Slug
- adolfoaranaes12/risk-profile
- Created
- 4d ago