fix: resolve agent namespace issues across 11 plugins

Fixed 109 agent namespace issues where external agent references were
missing proper namespace prefixes (plugin-name::agent-name format).

Affected plugins (bumped to v1.2.1):
- tdd-workflows: Fixed 12 references in tdd-cycle, tdd-green, tdd-red
- incident-response: Fixed 29 references in incident-response, smart-fix
- data-engineering: Fixed 12 references in data-driven-feature
- backend-development: Fixed 11 references in feature-development
- framework-migration: Fixed 10 references in legacy-modernize
- git-pr-workflows: Fixed 7 references in git-workflow
- security-scanning: Fixed 7 references in security-hardening
- full-stack-orchestration: Fixed 6 references in full-stack-feature
- application-performance: Fixed 5 references in performance-optimization
- comprehensive-review: Fixed 5 references in full-review
- multi-platform-apps: Fixed 5 references in multi-platform

Changes:
- Added namespace prefixes for all external agent references
- Replaced 14 missing agents with appropriate equivalents
- Updated marketplace.json to bump affected plugin versions

Closes #95
This commit is contained in:
Seth Hobson
2025-10-13 19:15:30 -04:00
parent 9b64821cc9
commit 492adf9353
15 changed files with 126 additions and 126 deletions

View File

@@ -39,7 +39,7 @@ Use Task tool with security and performance agents, incorporating Phase 1 findin
- Context: Incorporates architectural vulnerabilities identified in Phase 1B
### 2B. Performance & Scalability Analysis
- Use Task tool with subagent_type="performance-engineer"
- Use Task tool with subagent_type="application-performance::performance-engineer"
- Prompt: "Conduct performance analysis and scalability assessment for: $ARGUMENTS. Profile code for CPU/memory hotspots, analyze database query performance, review caching strategies, identify N+1 problems, assess connection pooling, and evaluate asynchronous processing patterns. Consider architectural findings from Phase 1: {phase1_architecture_context}. Check for memory leaks, resource contention, and bottlenecks under load."
- Expected output: Performance metrics, bottleneck analysis, optimization recommendations
- Context: Uses architecture insights to identify systemic performance issues
@@ -49,13 +49,13 @@ Use Task tool with security and performance agents, incorporating Phase 1 findin
Use Task tool for test and documentation quality assessment:
### 3A. Test Coverage & Quality Analysis
- Use Task tool with subagent_type="test-automator"
- Use Task tool with subagent_type="unit-testing::test-automator"
- Prompt: "Evaluate testing strategy and implementation for: $ARGUMENTS. Analyze unit test coverage, integration test completeness, end-to-end test scenarios, test pyramid adherence, and test maintainability. Review test quality metrics including assertion density, test isolation, mock usage, and flakiness. Consider security and performance test requirements from Phase 2: {phase2_security_context}, {phase2_performance_context}. Verify TDD practices if --tdd-review flag is set."
- Expected output: Coverage report, test quality metrics, testing gap analysis
- Context: Incorporates security and performance testing requirements from Phase 2
### 3B. Documentation & API Specification Review
- Use Task tool with subagent_type="docs-architect"
- Use Task tool with subagent_type="code-documentation::docs-architect"
- Prompt: "Review documentation completeness and quality for: $ARGUMENTS. Assess inline code documentation, API documentation (OpenAPI/Swagger), architecture decision records (ADRs), README completeness, deployment guides, and runbooks. Verify documentation reflects actual implementation based on all previous phase findings: {phase1_context}, {phase2_context}. Check for outdated documentation, missing examples, and unclear explanations."
- Expected output: Documentation coverage report, inconsistency list, improvement recommendations
- Context: Cross-references all previous findings to ensure documentation accuracy
@@ -65,13 +65,13 @@ Use Task tool for test and documentation quality assessment:
Use Task tool to verify framework-specific and industry best practices:
### 4A. Framework & Language Best Practices
- Use Task tool with subagent_type="framework-specialist"
- Use Task tool with subagent_type="framework-migration::legacy-modernizer"
- Prompt: "Verify adherence to framework and language best practices for: $ARGUMENTS. Check modern JavaScript/TypeScript patterns, React hooks best practices, Python PEP compliance, Java enterprise patterns, Go idiomatic code, or framework-specific conventions (based on --framework flag). Review package management, build configuration, environment handling, and deployment practices. Include all quality issues from previous phases: {all_previous_contexts}."
- Expected output: Best practices compliance report, modernization recommendations
- Context: Synthesizes all previous findings for framework-specific guidance
### 4B. CI/CD & DevOps Practices Review
- Use Task tool with subagent_type="devops-engineer"
- Use Task tool with subagent_type="cicd-automation::deployment-engineer"
- Prompt: "Review CI/CD pipeline and DevOps practices for: $ARGUMENTS. Evaluate build automation, test automation integration, deployment strategies (blue-green, canary), infrastructure as code, monitoring/observability setup, and incident response procedures. Assess pipeline security, artifact management, and rollback capabilities. Consider all issues identified in previous phases that impact deployment: {all_critical_issues}."
- Expected output: Pipeline assessment, DevOps maturity evaluation, automation recommendations
- Context: Focuses on operationalizing fixes for all identified issues