docs: update all documentation and add AI tooling configs
- Rewrite README.md with current architecture, features and stack - Update docs/API.md with all current endpoints (corporate, BI, client 360) - Update docs/ARCHITECTURE.md with cache, modular queries, services, ETL - Update docs/GUIA-USUARIO.md for all roles (admin, corporate, agente) - Add docs/INDEX.md documentation index - Add PROJETO.md comprehensive project reference - Add BI-CCC-Implementation-Guide.md - Include AI agent configs (.claude, .agents, .gemini, _bmad) - Add netbird VPN configuration - Add status report Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
This commit is contained in:
50
.agents/skills/bmad-code-review/steps/step-03-triage.md
Normal file
50
.agents/skills/bmad-code-review/steps/step-03-triage.md
Normal file
@@ -0,0 +1,50 @@
|
||||
---
|
||||
---
|
||||
|
||||
# Step 3: Triage
|
||||
|
||||
## RULES
|
||||
|
||||
- YOU MUST ALWAYS SPEAK OUTPUT in your Agent communication style with the config `{communication_language}`
|
||||
- Be precise. When uncertain between categories, prefer the more conservative classification.
|
||||
|
||||
## INSTRUCTIONS
|
||||
|
||||
1. **Normalize** findings into a common format. Expected input formats:
|
||||
- Adversarial (Blind Hunter): markdown list of descriptions
|
||||
- Edge Case Hunter: JSON array with `location`, `trigger_condition`, `guard_snippet`, `potential_consequence` fields
|
||||
- Acceptance Auditor: markdown list with title, AC/constraint reference, and evidence
|
||||
|
||||
If a layer's output does not match its expected format, attempt best-effort parsing. Note any parsing issues for the user.
|
||||
|
||||
Convert all to a unified list where each finding has:
|
||||
- `id` -- sequential integer
|
||||
- `source` -- `blind`, `edge`, `auditor`, or merged sources (e.g., `blind+edge`)
|
||||
- `title` -- one-line summary
|
||||
- `detail` -- full description
|
||||
- `location` -- file and line reference (if available)
|
||||
|
||||
2. **Deduplicate.** If two or more findings describe the same issue, merge them into one:
|
||||
- Use the most specific finding as the base (prefer edge-case JSON with location over adversarial prose).
|
||||
- Append any unique detail, reasoning, or location references from the other finding(s) into the surviving `detail` field.
|
||||
- Set `source` to the merged sources (e.g., `blind+edge`).
|
||||
|
||||
3. **Classify** each finding into exactly one bucket:
|
||||
- **intent_gap** -- The spec/intent is incomplete; cannot resolve from existing information. Only possible if `{review_mode}` = `"full"`.
|
||||
- **bad_spec** -- The spec should have prevented this; spec is wrong or ambiguous. Only possible if `{review_mode}` = `"full"`.
|
||||
- **patch** -- Code issue that is trivially fixable without human input. Just needs a code change.
|
||||
- **defer** -- Pre-existing issue not caused by the current change. Real but not actionable now.
|
||||
- **reject** -- Noise, false positive, or handled elsewhere.
|
||||
|
||||
If `{review_mode}` = `"no-spec"` and a finding would otherwise be `intent_gap` or `bad_spec`, reclassify it as `patch` (if code-fixable) or `defer` (if not).
|
||||
|
||||
4. **Drop** all `reject` findings. Record the reject count for the summary.
|
||||
|
||||
5. If `{failed_layers}` is non-empty, report which layers failed before announcing results. If zero findings remain after dropping rejects AND `{failed_layers}` is non-empty, warn the user that the review may be incomplete rather than announcing a clean review.
|
||||
|
||||
6. If zero findings remain after dropping rejects and no layers failed, note clean review.
|
||||
|
||||
|
||||
## NEXT
|
||||
|
||||
Read fully and follow `./step-04-present.md`
|
||||
Reference in New Issue
Block a user