diff --git a/.claude-plugin/marketplace.json b/.claude-plugin/marketplace.json index c687e9e..765d8d5 100644 --- a/.claude-plugin/marketplace.json +++ b/.claude-plugin/marketplace.json @@ -72,7 +72,7 @@ "name": "git-pr-workflows", "source": "./plugins/git-pr-workflows", "description": "Git workflow automation, pull request enhancement, and team onboarding processes", - "version": "1.2.0", + "version": "1.2.1", "author": { "name": "Seth Hobson", "url": "https://github.com/wshobson" @@ -102,7 +102,7 @@ "name": "backend-development", "source": "./plugins/backend-development", "description": "Backend API design, GraphQL architecture, and test-driven backend development", - "version": "1.2.0", + "version": "1.2.1", "author": { "name": "Seth Hobson", "url": "https://github.com/wshobson" @@ -161,7 +161,7 @@ "name": "full-stack-orchestration", "source": "./plugins/full-stack-orchestration", "description": "End-to-end feature orchestration with testing, security, performance, and deployment", - "version": "1.2.0", + "version": "1.2.1", "author": { "name": "Seth Hobson", "url": "https://github.com/wshobson" @@ -221,7 +221,7 @@ "name": "tdd-workflows", "source": "./plugins/tdd-workflows", "description": "Test-driven development methodology with red-green-refactor cycles and code review", - "version": "1.2.0", + "version": "1.2.1", "author": { "name": "Seth Hobson", "url": "https://github.com/wshobson" @@ -514,7 +514,7 @@ "name": "data-engineering", "source": "./plugins/data-engineering", "description": "ETL pipeline construction, data warehouse design, batch processing workflows, and data-driven feature development", - "version": "1.2.0", + "version": "1.2.1", "author": { "name": "Seth Hobson", "url": "https://github.com/wshobson" @@ -544,7 +544,7 @@ "name": "incident-response", "source": "./plugins/incident-response", "description": "Production incident management, triage workflows, and automated incident resolution", - "version": "1.2.0", + "version": "1.2.1", "author": { "name": "Seth Hobson", "url": "https://github.com/wshobson" @@ -812,7 +812,7 @@ "name": "application-performance", "source": "./plugins/application-performance", "description": "Application profiling, performance optimization, and observability for frontend and backend systems", - "version": "1.2.0", + "version": "1.2.1", "author": { "name": "Seth Hobson", "url": "https://github.com/wshobson" @@ -871,7 +871,7 @@ "name": "comprehensive-review", "source": "./plugins/comprehensive-review", "description": "Multi-perspective code analysis covering architecture, security, and best practices", - "version": "1.2.0", + "version": "1.2.1", "author": { "name": "Seth Hobson", "url": "https://github.com/wshobson" @@ -930,7 +930,7 @@ "name": "framework-migration", "source": "./plugins/framework-migration", "description": "Framework updates, migration planning, and architectural transformation workflows", - "version": "1.2.0", + "version": "1.2.1", "author": { "name": "Seth Hobson", "url": "https://github.com/wshobson" @@ -1047,7 +1047,7 @@ "name": "security-scanning", "source": "./plugins/security-scanning", "description": "SAST analysis, dependency vulnerability scanning, OWASP Top 10 compliance, container security scanning, and automated security hardening", - "version": "1.2.0", + "version": "1.2.1", "author": { "name": "Seth Hobson", "url": "https://github.com/wshobson" @@ -1364,7 +1364,7 @@ "name": "multi-platform-apps", "source": "./plugins/multi-platform-apps", "description": "Cross-platform application development coordinating web, iOS, Android, and desktop implementations", - "version": "1.2.0", + "version": "1.2.1", "author": { "name": "Seth Hobson", "url": "https://github.com/wshobson" diff --git a/plugins/application-performance/commands/performance-optimization.md b/plugins/application-performance/commands/performance-optimization.md index e630bea..341d3c8 100644 --- a/plugins/application-performance/commands/performance-optimization.md +++ b/plugins/application-performance/commands/performance-optimization.md @@ -25,13 +25,13 @@ Optimize application performance end-to-end using specialized performance and op ## Phase 2: Database & Backend Optimization ### 4. Database Performance Optimization -- Use Task tool with subagent_type="database-optimizer" +- Use Task tool with subagent_type="database-cloud-optimization::database-optimizer" - Prompt: "Optimize database performance for: $ARGUMENTS based on profiling data: {context_from_phase_1}. Analyze slow query logs, create missing indexes, optimize execution plans, implement query result caching with Redis/Memcached. Review connection pooling, prepared statements, and batch processing opportunities. Consider read replicas and database sharding if needed." - Context: Performance bottlenecks from phase 1 - Output: Optimized queries, new indexes, caching strategy, connection pool configuration ### 5. Backend Code & API Optimization -- Use Task tool with subagent_type="backend-architect" +- Use Task tool with subagent_type="backend-development::backend-architect" - Prompt: "Optimize backend services for: $ARGUMENTS targeting bottlenecks: {context_from_phase_1}. Implement efficient algorithms, add application-level caching, optimize N+1 queries, use async/await patterns effectively. Implement pagination, response compression, GraphQL query optimization, and batch API operations. Add circuit breakers and bulkheads for resilience." - Context: Database optimizations from step 4, profiling data from phase 1 - Output: Optimized backend code, caching implementation, API improvements, resilience patterns @@ -51,13 +51,13 @@ Optimize application performance end-to-end using specialized performance and op - Output: Optimized bundles, lazy loading implementation, improved Core Web Vitals ### 8. CDN & Edge Optimization -- Use Task tool with subagent_type="cloud-architect" +- Use Task tool with subagent_type="cloud-infrastructure::cloud-architect" - Prompt: "Optimize CDN and edge performance for: $ARGUMENTS. Configure CloudFlare/CloudFront for optimal caching, implement edge functions for dynamic content, set up image optimization with responsive images and WebP/AVIF formats. Configure HTTP/2 and HTTP/3, implement Brotli compression. Set up geographic distribution for global users." - Context: Frontend optimizations from step 7 - Output: CDN configuration, edge caching rules, compression setup, geographic optimization ### 9. Mobile & Progressive Web App Optimization -- Use Task tool with subagent_type="mobile-developer" +- Use Task tool with subagent_type="frontend-mobile-development::mobile-developer" - Prompt: "Optimize mobile experience for: $ARGUMENTS. Implement service workers for offline functionality, optimize for slow networks with adaptive loading. Reduce JavaScript execution time for mobile CPUs. Implement virtual scrolling for long lists. Optimize touch responsiveness and smooth animations. Consider React Native/Flutter specific optimizations if applicable." - Context: Frontend optimizations from steps 7-8 - Output: Mobile-optimized code, PWA implementation, offline functionality @@ -71,7 +71,7 @@ Optimize application performance end-to-end using specialized performance and op - Output: Load test results, performance under load, breaking points, scalability analysis ### 11. Performance Regression Testing -- Use Task tool with subagent_type="test-automator" +- Use Task tool with subagent_type="performance-testing-review::test-automator" - Prompt: "Create automated performance regression tests for: $ARGUMENTS. Set up performance budgets for key metrics, integrate with CI/CD pipeline using GitHub Actions or similar. Create Lighthouse CI tests for frontend, API performance tests with Artillery, and database performance benchmarks. Implement automatic rollback triggers for performance regressions." - Context: Load test results from step 10, baseline metrics from phase 1 - Output: Performance test suite, CI/CD integration, regression prevention system diff --git a/plugins/backend-development/commands/feature-development.md b/plugins/backend-development/commands/feature-development.md index 7ef0996..816c9cb 100644 --- a/plugins/backend-development/commands/feature-development.md +++ b/plugins/backend-development/commands/feature-development.md @@ -26,19 +26,19 @@ Orchestrate end-to-end feature development from requirements to production deplo ## Phase 1: Discovery & Requirements Planning 1. **Business Analysis & Requirements** - - Use Task tool with subagent_type="business-analyst" + - Use Task tool with subagent_type="business-analytics::business-analyst" - Prompt: "Analyze feature requirements for: $ARGUMENTS. Define user stories, acceptance criteria, success metrics, and business value. Identify stakeholders, dependencies, and risks. Create feature specification document with clear scope boundaries." - Expected output: Requirements document with user stories, success metrics, risk assessment - Context: Initial feature request and business context 2. **Technical Architecture Design** - - Use Task tool with subagent_type="architect-review" + - Use Task tool with subagent_type="comprehensive-review::architect-review" - Prompt: "Design technical architecture for feature: $ARGUMENTS. Using requirements: [include business analysis from step 1]. Define service boundaries, API contracts, data models, integration points, and technology stack. Consider scalability, performance, and security requirements." - Expected output: Technical design document with architecture diagrams, API specifications, data models - Context: Business requirements, existing system architecture 3. **Feasibility & Risk Assessment** - - Use Task tool with subagent_type="security-auditor" + - Use Task tool with subagent_type="security-scanning::security-auditor" - Prompt: "Assess security implications and risks for feature: $ARGUMENTS. Review architecture: [include technical design from step 2]. Identify security requirements, compliance needs, data privacy concerns, and potential vulnerabilities." - Expected output: Security assessment with risk matrix, compliance checklist, mitigation strategies - Context: Technical design, regulatory requirements @@ -52,13 +52,13 @@ Orchestrate end-to-end feature development from requirements to production deplo - Context: Technical design, API contracts, data models 5. **Frontend Implementation** - - Use Task tool with subagent_type="frontend-developer" + - Use Task tool with subagent_type="frontend-mobile-development::frontend-developer" - Prompt: "Build frontend components for: $ARGUMENTS. Integrate with backend APIs: [include API endpoints from step 4]. Implement responsive UI, state management, error handling, loading states, and analytics tracking. Add feature flag integration for A/B testing capabilities." - Expected output: Frontend components with API integration, state management, analytics - Context: Backend APIs, UI/UX designs, user stories 6. **Data Pipeline & Integration** - - Use Task tool with subagent_type="data-engineer" + - Use Task tool with subagent_type="data-engineering::data-engineer" - Prompt: "Build data pipelines for: $ARGUMENTS. Design ETL/ELT processes, implement data validation, create analytics events, set up data quality monitoring. Integrate with product analytics platforms for feature usage tracking." - Expected output: Data pipelines, analytics events, data quality checks - Context: Data requirements, analytics needs, existing data infrastructure @@ -66,19 +66,19 @@ Orchestrate end-to-end feature development from requirements to production deplo ## Phase 3: Testing & Quality Assurance 7. **Automated Test Suite** - - Use Task tool with subagent_type="test-automator" + - Use Task tool with subagent_type="unit-testing::test-automator" - Prompt: "Create comprehensive test suite for: $ARGUMENTS. Write unit tests for backend: [from step 4] and frontend: [from step 5]. Add integration tests for API endpoints, E2E tests for critical user journeys, performance tests for scalability validation. Ensure minimum 80% code coverage." - Expected output: Test suites with unit, integration, E2E, and performance tests - Context: Implementation code, acceptance criteria, test requirements 8. **Security Validation** - - Use Task tool with subagent_type="security-auditor" + - Use Task tool with subagent_type="security-scanning::security-auditor" - Prompt: "Perform security testing for: $ARGUMENTS. Review implementation: [include backend and frontend from steps 4-5]. Run OWASP checks, penetration testing, dependency scanning, and compliance validation. Verify data encryption, authentication, and authorization." - Expected output: Security test results, vulnerability report, remediation actions - Context: Implementation code, security requirements 9. **Performance Optimization** - - Use Task tool with subagent_type="performance-engineer" + - Use Task tool with subagent_type="application-performance::performance-engineer" - Prompt: "Optimize performance for: $ARGUMENTS. Analyze backend services: [from step 4] and frontend: [from step 5]. Profile code, optimize queries, implement caching, reduce bundle sizes, improve load times. Set up performance budgets and monitoring." - Expected output: Performance improvements, optimization report, performance metrics - Context: Implementation code, performance requirements @@ -86,19 +86,19 @@ Orchestrate end-to-end feature development from requirements to production deplo ## Phase 4: Deployment & Monitoring 10. **Deployment Strategy & Pipeline** - - Use Task tool with subagent_type="deployment-engineer" + - Use Task tool with subagent_type="deployment-strategies::deployment-engineer" - Prompt: "Prepare deployment for: $ARGUMENTS. Create CI/CD pipeline with automated tests: [from step 7]. Configure feature flags for gradual rollout, implement blue-green deployment, set up rollback procedures. Create deployment runbook and rollback plan." - Expected output: CI/CD pipeline, deployment configuration, rollback procedures - Context: Test suites, infrastructure requirements, deployment strategy 11. **Observability & Monitoring** - - Use Task tool with subagent_type="observability-engineer" + - Use Task tool with subagent_type="observability-monitoring::observability-engineer" - Prompt: "Set up observability for: $ARGUMENTS. Implement distributed tracing, custom metrics, error tracking, and alerting. Create dashboards for feature usage, performance metrics, error rates, and business KPIs. Set up SLOs/SLIs with automated alerts." - Expected output: Monitoring dashboards, alerts, SLO definitions, observability infrastructure - Context: Feature implementation, success metrics, operational requirements 12. **Documentation & Knowledge Transfer** - - Use Task tool with subagent_type="doc-generator" + - Use Task tool with subagent_type="documentation-generation::docs-architect" - Prompt: "Generate comprehensive documentation for: $ARGUMENTS. Create API documentation, user guides, deployment guides, troubleshooting runbooks. Include architecture diagrams, data flow diagrams, and integration guides. Generate automated changelog from commits." - Expected output: API docs, user guides, runbooks, architecture documentation - Context: All previous phases' outputs diff --git a/plugins/comprehensive-review/commands/full-review.md b/plugins/comprehensive-review/commands/full-review.md index 0249273..1fc4b4c 100644 --- a/plugins/comprehensive-review/commands/full-review.md +++ b/plugins/comprehensive-review/commands/full-review.md @@ -39,7 +39,7 @@ Use Task tool with security and performance agents, incorporating Phase 1 findin - Context: Incorporates architectural vulnerabilities identified in Phase 1B ### 2B. Performance & Scalability Analysis -- Use Task tool with subagent_type="performance-engineer" +- Use Task tool with subagent_type="application-performance::performance-engineer" - Prompt: "Conduct performance analysis and scalability assessment for: $ARGUMENTS. Profile code for CPU/memory hotspots, analyze database query performance, review caching strategies, identify N+1 problems, assess connection pooling, and evaluate asynchronous processing patterns. Consider architectural findings from Phase 1: {phase1_architecture_context}. Check for memory leaks, resource contention, and bottlenecks under load." - Expected output: Performance metrics, bottleneck analysis, optimization recommendations - Context: Uses architecture insights to identify systemic performance issues @@ -49,13 +49,13 @@ Use Task tool with security and performance agents, incorporating Phase 1 findin Use Task tool for test and documentation quality assessment: ### 3A. Test Coverage & Quality Analysis -- Use Task tool with subagent_type="test-automator" +- Use Task tool with subagent_type="unit-testing::test-automator" - Prompt: "Evaluate testing strategy and implementation for: $ARGUMENTS. Analyze unit test coverage, integration test completeness, end-to-end test scenarios, test pyramid adherence, and test maintainability. Review test quality metrics including assertion density, test isolation, mock usage, and flakiness. Consider security and performance test requirements from Phase 2: {phase2_security_context}, {phase2_performance_context}. Verify TDD practices if --tdd-review flag is set." - Expected output: Coverage report, test quality metrics, testing gap analysis - Context: Incorporates security and performance testing requirements from Phase 2 ### 3B. Documentation & API Specification Review -- Use Task tool with subagent_type="docs-architect" +- Use Task tool with subagent_type="code-documentation::docs-architect" - Prompt: "Review documentation completeness and quality for: $ARGUMENTS. Assess inline code documentation, API documentation (OpenAPI/Swagger), architecture decision records (ADRs), README completeness, deployment guides, and runbooks. Verify documentation reflects actual implementation based on all previous phase findings: {phase1_context}, {phase2_context}. Check for outdated documentation, missing examples, and unclear explanations." - Expected output: Documentation coverage report, inconsistency list, improvement recommendations - Context: Cross-references all previous findings to ensure documentation accuracy @@ -65,13 +65,13 @@ Use Task tool for test and documentation quality assessment: Use Task tool to verify framework-specific and industry best practices: ### 4A. Framework & Language Best Practices -- Use Task tool with subagent_type="framework-specialist" +- Use Task tool with subagent_type="framework-migration::legacy-modernizer" - Prompt: "Verify adherence to framework and language best practices for: $ARGUMENTS. Check modern JavaScript/TypeScript patterns, React hooks best practices, Python PEP compliance, Java enterprise patterns, Go idiomatic code, or framework-specific conventions (based on --framework flag). Review package management, build configuration, environment handling, and deployment practices. Include all quality issues from previous phases: {all_previous_contexts}." - Expected output: Best practices compliance report, modernization recommendations - Context: Synthesizes all previous findings for framework-specific guidance ### 4B. CI/CD & DevOps Practices Review -- Use Task tool with subagent_type="devops-engineer" +- Use Task tool with subagent_type="cicd-automation::deployment-engineer" - Prompt: "Review CI/CD pipeline and DevOps practices for: $ARGUMENTS. Evaluate build automation, test automation integration, deployment strategies (blue-green, canary), infrastructure as code, monitoring/observability setup, and incident response procedures. Assess pipeline security, artifact management, and rollback capabilities. Consider all issues identified in previous phases that impact deployment: {all_critical_issues}." - Expected output: Pipeline assessment, DevOps maturity evaluation, automation recommendations - Context: Focuses on operationalizing fixes for all identified issues diff --git a/plugins/data-engineering/commands/data-driven-feature.md b/plugins/data-engineering/commands/data-driven-feature.md index 8527545..253a25b 100644 --- a/plugins/data-engineering/commands/data-driven-feature.md +++ b/plugins/data-engineering/commands/data-driven-feature.md @@ -7,18 +7,18 @@ Build features guided by data insights, A/B testing, and continuous measurement ## Phase 1: Data Analysis and Hypothesis Formation ### 1. Exploratory Data Analysis -- Use Task tool with subagent_type="data-scientist" +- Use Task tool with subagent_type="machine-learning-ops::data-scientist" - Prompt: "Perform exploratory data analysis for feature: $ARGUMENTS. Analyze existing user behavior data, identify patterns and opportunities, segment users by behavior, and calculate baseline metrics. Use modern analytics tools (Amplitude, Mixpanel, Segment) to understand current user journeys, conversion funnels, and engagement patterns." - Output: EDA report with visualizations, user segments, behavioral patterns, baseline metrics ### 2. Business Hypothesis Development -- Use Task tool with subagent_type="business-analyst" +- Use Task tool with subagent_type="business-analytics::business-analyst" - Context: Data scientist's EDA findings and behavioral patterns - Prompt: "Formulate business hypotheses for feature: $ARGUMENTS based on data analysis. Define clear success metrics, expected impact on key business KPIs, target user segments, and minimum detectable effects. Create measurable hypotheses using frameworks like ICE scoring or RICE prioritization." - Output: Hypothesis document, success metrics definition, expected ROI calculations ### 3. Statistical Experiment Design -- Use Task tool with subagent_type="data-scientist" +- Use Task tool with subagent_type="machine-learning-ops::data-scientist" - Context: Business hypotheses and success metrics - Prompt: "Design statistical experiment for feature: $ARGUMENTS. Calculate required sample size for statistical power, define control and treatment groups, specify randomization strategy, and plan for multiple testing corrections. Consider Bayesian A/B testing approaches for faster decision making. Design for both primary and guardrail metrics." - Output: Experiment design document, power analysis, statistical test plan @@ -26,19 +26,19 @@ Build features guided by data insights, A/B testing, and continuous measurement ## Phase 2: Feature Architecture and Analytics Design ### 4. Feature Architecture Planning -- Use Task tool with subagent_type="backend-architect" +- Use Task tool with subagent_type="data-engineering::backend-architect" - Context: Business requirements and experiment design - Prompt: "Design feature architecture for: $ARGUMENTS with A/B testing capability. Include feature flag integration (LaunchDarkly, Split.io, or Optimizely), gradual rollout strategy, circuit breakers for safety, and clean separation between control and treatment logic. Ensure architecture supports real-time configuration updates." - Output: Architecture diagrams, feature flag schema, rollout strategy ### 5. Analytics Instrumentation Design -- Use Task tool with subagent_type="data-engineer" +- Use Task tool with subagent_type="data-engineering::data-engineer" - Context: Feature architecture and success metrics - Prompt: "Design comprehensive analytics instrumentation for: $ARGUMENTS. Define event schemas for user interactions, specify properties for segmentation and analysis, design funnel tracking and conversion events, plan cohort analysis capabilities. Implement using modern SDKs (Segment, Amplitude, Mixpanel) with proper event taxonomy." - Output: Event tracking plan, analytics schema, instrumentation guide ### 6. Data Pipeline Architecture -- Use Task tool with subagent_type="data-engineer" +- Use Task tool with subagent_type="data-engineering::data-engineer" - Context: Analytics requirements and existing data infrastructure - Prompt: "Design data pipelines for feature: $ARGUMENTS. Include real-time streaming for live metrics (Kafka, Kinesis), batch processing for detailed analysis, data warehouse integration (Snowflake, BigQuery), and feature store for ML if applicable. Ensure proper data governance and GDPR compliance." - Output: Pipeline architecture, ETL/ELT specifications, data flow diagrams @@ -46,19 +46,19 @@ Build features guided by data insights, A/B testing, and continuous measurement ## Phase 3: Implementation with Instrumentation ### 7. Backend Implementation -- Use Task tool with subagent_type="backend-engineer" +- Use Task tool with subagent_type="backend-development::backend-architect" - Context: Architecture design and feature requirements - Prompt: "Implement backend for feature: $ARGUMENTS with full instrumentation. Include feature flag checks at decision points, comprehensive event tracking for all user actions, performance metrics collection, error tracking and monitoring. Implement proper logging for experiment analysis." - Output: Backend code with analytics, feature flag integration, monitoring setup ### 8. Frontend Implementation -- Use Task tool with subagent_type="frontend-engineer" +- Use Task tool with subagent_type="frontend-mobile-development::frontend-developer" - Context: Backend APIs and analytics requirements - Prompt: "Build frontend for feature: $ARGUMENTS with analytics tracking. Implement event tracking for all user interactions, session recording integration if applicable, performance metrics (Core Web Vitals), and proper error boundaries. Ensure consistent experience between control and treatment groups." - Output: Frontend code with analytics, A/B test variants, performance monitoring ### 9. ML Model Integration (if applicable) -- Use Task tool with subagent_type="ml-engineer" +- Use Task tool with subagent_type="machine-learning-ops::ml-engineer" - Context: Feature requirements and data pipelines - Prompt: "Integrate ML models for feature: $ARGUMENTS if needed. Implement online inference with low latency, A/B testing between model versions, model performance tracking, and automatic fallback mechanisms. Set up model monitoring for drift detection." - Output: ML pipeline, model serving infrastructure, monitoring setup @@ -66,13 +66,13 @@ Build features guided by data insights, A/B testing, and continuous measurement ## Phase 4: Pre-Launch Validation ### 10. Analytics Validation -- Use Task tool with subagent_type="data-engineer" +- Use Task tool with subagent_type="data-engineering::data-engineer" - Context: Implemented tracking and event schemas - Prompt: "Validate analytics implementation for: $ARGUMENTS. Test all event tracking in staging, verify data quality and completeness, validate funnel definitions, ensure proper user identification and session tracking. Run end-to-end tests for data pipeline." - Output: Validation report, data quality metrics, tracking coverage analysis ### 11. Experiment Setup -- Use Task tool with subagent_type="platform-engineer" +- Use Task tool with subagent_type="cloud-infrastructure::deployment-engineer" - Context: Feature flags and experiment design - Prompt: "Configure experiment infrastructure for: $ARGUMENTS. Set up feature flags with proper targeting rules, configure traffic allocation (start with 5-10%), implement kill switches, set up monitoring alerts for key metrics. Test randomization and assignment logic." - Output: Experiment configuration, monitoring dashboards, rollout plan @@ -80,13 +80,13 @@ Build features guided by data insights, A/B testing, and continuous measurement ## Phase 5: Launch and Experimentation ### 12. Gradual Rollout -- Use Task tool with subagent_type="deployment-engineer" +- Use Task tool with subagent_type="cloud-infrastructure::deployment-engineer" - Context: Experiment configuration and monitoring setup - Prompt: "Execute gradual rollout for feature: $ARGUMENTS. Start with internal dogfooding, then beta users (1-5%), gradually increase to target traffic. Monitor error rates, performance metrics, and early indicators. Implement automated rollback on anomalies." - Output: Rollout execution, monitoring alerts, health metrics ### 13. Real-time Monitoring -- Use Task tool with subagent_type="observability-engineer" +- Use Task tool with subagent_type="observability-monitoring::observability-engineer" - Context: Deployed feature and success metrics - Prompt: "Set up comprehensive monitoring for: $ARGUMENTS. Create real-time dashboards for experiment metrics, configure alerts for statistical significance, monitor guardrail metrics for negative impacts, track system performance and error rates. Use tools like Datadog, New Relic, or custom dashboards." - Output: Monitoring dashboards, alert configurations, SLO definitions @@ -94,19 +94,19 @@ Build features guided by data insights, A/B testing, and continuous measurement ## Phase 6: Analysis and Decision Making ### 14. Statistical Analysis -- Use Task tool with subagent_type="data-scientist" +- Use Task tool with subagent_type="machine-learning-ops::data-scientist" - Context: Experiment data and original hypotheses - Prompt: "Analyze A/B test results for: $ARGUMENTS. Calculate statistical significance with confidence intervals, check for segment-level effects, analyze secondary metrics impact, investigate any unexpected patterns. Use both frequentist and Bayesian approaches. Account for multiple testing if applicable." - Output: Statistical analysis report, significance tests, segment analysis ### 15. Business Impact Assessment -- Use Task tool with subagent_type="business-analyst" +- Use Task tool with subagent_type="business-analytics::business-analyst" - Context: Statistical analysis and business metrics - Prompt: "Assess business impact of feature: $ARGUMENTS. Calculate actual vs expected ROI, analyze impact on key business metrics, evaluate cost-benefit including operational overhead, project long-term value. Make recommendation on full rollout, iteration, or rollback." - Output: Business impact report, ROI analysis, recommendation document ### 16. Post-Launch Optimization -- Use Task tool with subagent_type="data-scientist" +- Use Task tool with subagent_type="machine-learning-ops::data-scientist" - Context: Launch results and user feedback - Prompt: "Identify optimization opportunities for: $ARGUMENTS based on data. Analyze user behavior patterns in treatment group, identify friction points in user journey, suggest improvements based on data, plan follow-up experiments. Use cohort analysis for long-term impact." - Output: Optimization recommendations, follow-up experiment plans diff --git a/plugins/framework-migration/commands/legacy-modernize.md b/plugins/framework-migration/commands/legacy-modernize.md index 0aa15c0..ab0491c 100644 --- a/plugins/framework-migration/commands/legacy-modernize.md +++ b/plugins/framework-migration/commands/legacy-modernize.md @@ -18,7 +18,7 @@ Orchestrate a comprehensive legacy system modernization using the strangler fig - Expected output: Visual dependency map and integration point catalog ### 3. Business Impact and Risk Assessment -- Use Task tool with subagent_type="project-manager" +- Use Task tool with subagent_type="business-analytics::business-analyst" - Prompt: "Evaluate business impact of modernizing each component identified. Create risk assessment matrix considering: business criticality (revenue impact), user traffic patterns, data sensitivity, regulatory requirements, and fallback complexity. Prioritize components using a weighted scoring system: (Business Value × 0.4) + (Technical Risk × 0.3) + (Quick Win Potential × 0.3). Define rollback strategies for each component." - Context from previous: Component inventory, dependency mapping - Expected output: Prioritized migration roadmap with risk mitigation strategies @@ -26,18 +26,18 @@ Orchestrate a comprehensive legacy system modernization using the strangler fig ## Phase 2: Test Coverage Establishment ### 1. Legacy Code Test Coverage Analysis -- Use Task tool with subagent_type="test-automator" +- Use Task tool with subagent_type="unit-testing::test-automator" - Prompt: "Analyze existing test coverage for legacy components at $ARGUMENTS. Use coverage tools to identify untested code paths, missing integration tests, and absent end-to-end scenarios. For components with <40% coverage, generate characterization tests that capture current behavior without modifying functionality. Create test harness for safe refactoring." - Expected output: Test coverage report and characterization test suite ### 2. Contract Testing Implementation -- Use Task tool with subagent_type="test-automator" +- Use Task tool with subagent_type="unit-testing::test-automator" - Prompt: "Implement contract tests for all integration points identified in dependency mapping. Create consumer-driven contracts for APIs, message queue interactions, and database schemas. Set up contract verification in CI/CD pipeline. Generate performance baselines for response times and throughput to validate modernized components maintain SLAs." - Context from previous: Integration point catalog, existing test coverage - Expected output: Contract test suite with performance baselines ### 3. Test Data Management Strategy -- Use Task tool with subagent_type="data-engineer" +- Use Task tool with subagent_type="data-engineering::data-engineer" - Prompt: "Design test data management strategy for parallel system operation. Create data generation scripts for edge cases, implement data masking for sensitive information, and establish test database refresh procedures. Set up monitoring for data consistency between legacy and modernized components during migration." - Context from previous: Database schemas, test requirements - Expected output: Test data pipeline and consistency monitoring @@ -45,18 +45,18 @@ Orchestrate a comprehensive legacy system modernization using the strangler fig ## Phase 3: Incremental Migration Implementation ### 1. Strangler Fig Infrastructure Setup -- Use Task tool with subagent_type="backend-architect" +- Use Task tool with subagent_type="backend-development::backend-architect" - Prompt: "Implement strangler fig infrastructure with API gateway for traffic routing. Configure feature flags for gradual rollout using environment variables or feature management service. Set up proxy layer with request routing rules based on: URL patterns, headers, or user segments. Implement circuit breakers and fallback mechanisms for resilience. Create observability dashboard for dual-system monitoring." - Expected output: API gateway configuration, feature flag system, monitoring dashboard ### 2. Component Modernization - First Wave -- Use Task tool with subagent_type="python-pro" or "golang-pro" (based on target stack) +- Use Task tool with subagent_type="python-development::python-pro" or "golang-pro" (based on target stack) - Prompt: "Modernize first-wave components (quick wins identified in assessment). For each component: extract business logic from legacy code, implement using modern patterns (dependency injection, SOLID principles), ensure backward compatibility through adapter patterns, maintain data consistency with event sourcing or dual writes. Follow 12-factor app principles. Components to modernize: [list from prioritized roadmap]" - Context from previous: Characterization tests, contract tests, infrastructure setup - Expected output: Modernized components with adapters ### 3. Security Hardening -- Use Task tool with subagent_type="security-auditor" +- Use Task tool with subagent_type="security-scanning::security-auditor" - Prompt: "Audit modernized components for security vulnerabilities. Implement security improvements including: OAuth 2.0/JWT authentication, role-based access control, input validation and sanitization, SQL injection prevention, XSS protection, and secrets management. Verify OWASP top 10 compliance. Configure security headers and implement rate limiting." - Context from previous: Modernized component code - Expected output: Security audit report and hardened components @@ -64,13 +64,13 @@ Orchestrate a comprehensive legacy system modernization using the strangler fig ## Phase 4: Performance Validation and Optimization ### 1. Performance Testing and Optimization -- Use Task tool with subagent_type="performance-engineer" +- Use Task tool with subagent_type="application-performance::performance-engineer" - Prompt: "Conduct performance testing comparing legacy vs modernized components. Run load tests simulating production traffic patterns, measure response times, throughput, and resource utilization. Identify performance regressions and optimize: database queries with indexing, caching strategies (Redis/Memcached), connection pooling, and async processing where applicable. Validate against SLA requirements." - Context from previous: Performance baselines, modernized components - Expected output: Performance test results and optimization recommendations ### 2. Progressive Rollout and Monitoring -- Use Task tool with subagent_type="deployment-engineer" +- Use Task tool with subagent_type="deployment-strategies::deployment-engineer" - Prompt: "Implement progressive rollout strategy using feature flags. Start with 5% traffic to modernized components, monitor error rates, latency, and business metrics. Define automatic rollback triggers: error rate >1%, latency >2x baseline, or business metric degradation. Create runbook for traffic shifting: 5% → 25% → 50% → 100% with 24-hour observation periods." - Context from previous: Feature flag configuration, monitoring dashboard - Expected output: Rollout plan with automated safeguards @@ -84,7 +84,7 @@ Orchestrate a comprehensive legacy system modernization using the strangler fig - Expected output: Decommissioning checklist and timeline ### 2. Documentation and Knowledge Transfer -- Use Task tool with subagent_type="docs-architect" +- Use Task tool with subagent_type="documentation-generation::docs-architect" - Prompt: "Create comprehensive modernization documentation including: architectural diagrams (before/after), API documentation with migration guides, runbooks for dual-system operation, troubleshooting guides for common issues, and lessons learned report. Generate developer onboarding guide for modernized system. Document technical decisions and trade-offs made during migration." - Context from previous: All migration artifacts and decisions - Expected output: Complete modernization documentation package diff --git a/plugins/full-stack-orchestration/commands/full-stack-feature.md b/plugins/full-stack-orchestration/commands/full-stack-feature.md index d599a50..f565bba 100644 --- a/plugins/full-stack-orchestration/commands/full-stack-feature.md +++ b/plugins/full-stack-orchestration/commands/full-stack-feature.md @@ -5,19 +5,19 @@ Orchestrate full-stack feature development across backend, frontend, and infrast ## Phase 1: Architecture & Design Foundation ### 1. Database Architecture Design -- Use Task tool with subagent_type="database-architect" +- Use Task tool with subagent_type="database-design::database-architect" - Prompt: "Design database schema and data models for: $ARGUMENTS. Consider scalability, query patterns, indexing strategy, and data consistency requirements. Include migration strategy if modifying existing schema. Provide both logical and physical data models." - Expected output: Entity relationship diagrams, table schemas, indexing strategy, migration scripts, data access patterns - Context: Initial requirements and business domain model ### 2. Backend Service Architecture -- Use Task tool with subagent_type="backend-architect" +- Use Task tool with subagent_type="backend-development::backend-architect" - Prompt: "Design backend service architecture for: $ARGUMENTS. Using the database design from previous step, create service boundaries, define API contracts (OpenAPI/GraphQL), design authentication/authorization strategy, and specify inter-service communication patterns. Include resilience patterns (circuit breakers, retries) and caching strategy." - Expected output: Service architecture diagram, OpenAPI specifications, authentication flows, caching architecture, message queue design (if applicable) - Context: Database schema from step 1, non-functional requirements ### 3. Frontend Component Architecture -- Use Task tool with subagent_type="frontend-developer" +- Use Task tool with subagent_type="frontend-mobile-development::frontend-developer" - Prompt: "Design frontend architecture and component structure for: $ARGUMENTS. Based on the API contracts from previous step, design component hierarchy, state management approach (Redux/Zustand/Context), routing structure, and data fetching patterns. Include accessibility requirements and responsive design strategy. Plan for Storybook component documentation." - Expected output: Component tree diagram, state management design, routing configuration, design system integration plan, accessibility checklist - Context: API specifications from step 2, UI/UX requirements @@ -25,19 +25,19 @@ Orchestrate full-stack feature development across backend, frontend, and infrast ## Phase 2: Parallel Implementation ### 4. Backend Service Implementation -- Use Task tool with subagent_type="python-pro" (or "golang-pro"/"nodejs-expert" based on stack) +- Use Task tool with subagent_type="python-development::python-pro" (or "golang-pro"/"nodejs-expert" based on stack) - Prompt: "Implement backend services for: $ARGUMENTS. Using the architecture and API specs from Phase 1, build RESTful/GraphQL endpoints with proper validation, error handling, and logging. Implement business logic, data access layer, authentication middleware, and integration with external services. Include observability (structured logging, metrics, tracing)." - Expected output: Backend service code, API endpoints, middleware, background jobs, unit tests, integration tests - Context: Architecture designs from Phase 1, database schema ### 5. Frontend Implementation -- Use Task tool with subagent_type="frontend-developer" +- Use Task tool with subagent_type="frontend-mobile-development::frontend-developer" - Prompt: "Implement frontend application for: $ARGUMENTS. Build React/Next.js components using the component architecture from Phase 1. Implement state management, API integration with proper error handling and loading states, form validation, and responsive layouts. Create Storybook stories for components. Ensure accessibility (WCAG 2.1 AA compliance)." - Expected output: React components, state management implementation, API client code, Storybook stories, responsive styles, accessibility implementations - Context: Component architecture from step 3, API contracts ### 6. Database Implementation & Optimization -- Use Task tool with subagent_type="sql-pro" +- Use Task tool with subagent_type="database-design::sql-pro" - Prompt: "Implement and optimize database layer for: $ARGUMENTS. Create migration scripts, stored procedures (if needed), optimize queries identified by backend implementation, set up proper indexes, and implement data validation constraints. Include database-level security measures and backup strategies." - Expected output: Migration scripts, optimized queries, stored procedures, index definitions, database security configuration - Context: Database design from step 1, query patterns from backend implementation diff --git a/plugins/git-pr-workflows/commands/git-workflow.md b/plugins/git-pr-workflows/commands/git-workflow.md index 8e0b1c8..51fb152 100644 --- a/plugins/git-pr-workflows/commands/git-workflow.md +++ b/plugins/git-pr-workflows/commands/git-workflow.md @@ -33,13 +33,13 @@ Orchestrate a comprehensive git workflow from code review through PR creation, l ## Phase 2: Testing and Validation ### 1. Test Execution and Coverage -- Use Task tool with subagent_type="test-automator" +- Use Task tool with subagent_type="unit-testing::test-automator" - Prompt: "Execute all test suites for the modified code. Run: 1) Unit tests, 2) Integration tests, 3) End-to-end tests if applicable. Generate coverage report and identify any untested code paths. Based on review issues: [insert critical/high issues], ensure tests cover the problem areas. Provide test results in format: {passed: [], failed: [], skipped: [], coverage: {statements: %, branches: %, functions: %, lines: %}, untested_critical_paths: []}" - Context from previous: Critical code review issues that need test coverage - Expected output: Complete test results and coverage metrics ### 2. Test Recommendations and Gap Analysis -- Use Task tool with subagent_type="test-automator" +- Use Task tool with subagent_type="unit-testing::test-automator" - Prompt: "Based on test results [insert summary] and code changes, identify: 1) Missing test scenarios, 2) Edge cases not covered, 3) Integration points needing verification, 4) Performance benchmarks needed. Generate test implementation recommendations prioritized by risk. Consider the breaking changes identified: [insert breaking changes]." - Context from previous: Test results, breaking changes, untested paths - Expected output: Prioritized list of additional tests needed @@ -53,7 +53,7 @@ Orchestrate a comprehensive git workflow from code review through PR creation, l - Expected output: Commit structure recommendation ### 2. Conventional Commit Message Creation -- Use Task tool with subagent_type="prompt-engineer" +- Use Task tool with subagent_type="llm-application-dev::prompt-engineer" - Prompt: "Create Conventional Commits format message(s) based on categorization: [insert categorization]. Format: (): with blank line then explaining what and why (not how), then