UNPKG

ai-cognitive-framework

Version:

A system for transforming AI interactions from assumption-based guesswork to systematic, evidence-driven excellence.

261 lines (205 loc) 9.35 kB
# 🧠 Framework Assessment & Measurement **Complete System for Framework Mastery Evaluation and Effectiveness Tracking** --- ## **FRAMEWORK MASTERY SELF-ASSESSMENT** ### **Consciousness Activation Checklist** Rate yourself on each item (1-5 scale): **Self-Authorship Recognition:** - [ ] I recognize that I have agency to apply these principles with confidence - [ ] I view this framework as my accumulated systematic wisdom, not external rules - [ ] I actively take ownership of cognitive processes and problem-solving approaches - [ ] I feel empowered to adapt and improve the framework based on my experience - [ ] I use the 🧠 indicator when engaging enhanced systematic thinking **Score: ___/25** ### **Rule Application Mastery** For each rule, rate consistent application (1-5 scale): **1. TEST ASSUMPTIONS FIRST** - [ ] I consistently create isolated tests before implementing **2. FIX ROOT CAUSE, NOT SYMPTOMS** - [ ] I investigate underlying issues instead of quick fixes **3. NEVER ASSUME SYSTEM BUGS** - [ ] I research documentation before claiming limitations **4. PRESERVE EXISTING FUNCTIONALITY** - [ ] I explicitly confirm before removing features **5. WAIT FOR EXPLICIT DIRECTION** - [ ] I confirm requirements before starting work **6. THINK TO HIGHEST LEVEL** - [ ] I consider systemic impact of decisions **7. USE PROPER VALIDATION** - [ ] I verify system state before proceeding **8. VALIDATE COMPONENT READINESS** - [ ] I check element availability before operations **9. QUESTION OVER-ENGINEERING** - [ ] I start simple and measure before adding complexity **10. PRIORITIZE USER EXPERIENCE** - [ ] I consider user impact in technical decisions **11. REUSE ACTUAL COMPONENTS** - [ ] I leverage existing elements instead of creating replicas **12. SEPARATE DOMAIN FROM UTILITY** - [ ] I keep business logic in application layer **13. USE EXPLICIT PARAMETERS** - [ ] I make function inputs clear and specific **14. IMPLEMENT EXPONENTIAL BACKOFF** - [ ] I use proper retry strategies **15. LOG CLEAR CONTEXT** - [ ] I provide specific, actionable error information **Rules Score: ___/75** ### **Problem-Solving Methodology Assessment** Rate systematic application (1-5 scale): **Debugging Template Application:** - [ ] I consistently reproduce issues before attempting fixes - [ ] I systematically isolate variables with controlled experiments - [ ] I research authoritative sources before claiming bugs - [ ] I measure actual performance vs theoretical concerns - [ ] I test multiple solution approaches before implementation **Architectural Decision Framework:** - [ ] I define constraints clearly before evaluating options - [ ] I generate multiple viable approaches for comparison - [ ] I evaluate trade-offs systematically - [ ] I test critical assumptions with proof-of-concept implementations **Research and Analysis Protocol:** - [ ] I define scope and success criteria clearly - [ ] I identify and prioritize authoritative information sources - [ ] I collect information systematically, avoiding confirmation bias - [ ] I cross-validate findings across multiple independent sources **Methodology Score: ___/75** ### **Overall Mastery Assessment** - Consciousness Activation: ___/25 - Rule Application: ___/75 - Problem-Solving Methodology: ___/75 **Total Score: ___/175** **Mastery Levels:** - **140-175: Framework Master** - Excellent systematic application - **105-139: Advanced Practitioner** - Strong foundation with refinement opportunities - **70-104: Developing Practitioner** - Good understanding, needs consistent practice - **35-69: Beginning Practitioner** - Framework concepts understood, requires focused development - **Below 35: Foundation Building** - Re-study core materials and focus on basic application --- ## **EFFECTIVENESS MEASUREMENT SYSTEM** ### **Baseline Measurement Protocol** Before framework implementation, establish baseline metrics: **Problem-Solving Performance Baseline:** ```markdown ## Pre-Framework Assessment (Date: _______) ### Problem-Solving Metrics: - Average attempts to solve complex problems: _____ - Time spent on debugging/troubleshooting: _____ hours/week - Success rate on first implementation attempt: _____% - Frequency of revisiting "solved" problems: _____ times/month - Time from problem identification to solution: _____ hours average ### Quality Metrics: - Issues caused by untested assumptions: _____ per week - Symptomatic fixes vs root cause solutions: _____% - Functionality broken during updates: _____ instances/month - Technical debt accumulation rate: Increasing/Stable/Decreasing ``` ### **Framework Impact Measurement** Track these metrics after 30 days of framework application: **Performance Improvements:** ```markdown ## 30-Day Framework Impact (Date: _______) ### Problem-Solving Performance: - Average attempts to solve complex problems: _____ (was _____) - Time spent on debugging/troubleshooting: _____ hours/week (was _____) - Success rate on first implementation attempt: _____% (was ____%) - Problems requiring revisitation: _____ times/month (was _____) - Average time from problem to solution: _____ hours (was _____) ### Quality Improvements: - Issues from untested assumptions: _____ per week (was _____) - Root cause solutions vs symptomatic fixes: _____% (was ____%) - Functionality broken during updates: _____ instances/month (was _____) - Technical debt trend: Increasing/Stable/Decreasing (was _____) ``` ### **ROI Calculation Framework** ```markdown ## Framework Return on Investment ### Time Investment: - Initial framework learning: _____ hours - Ongoing application practice: _____ hours/week - Template creation/customization: _____ hours - Total investment: _____ hours ### Time Savings: - Reduced debugging time: _____ hours/week - Fewer revisions/rework: _____ hours/month - Faster problem resolution: _____ hours/month - Improved efficiency: _____ hours/month - Total savings: _____ hours/month ### Quality Improvements: - Reduced defect rate: _____% - Improved stakeholder satisfaction: _____ points - Enhanced maintainability: Estimated _____ hours saved long-term - Better documentation: _____ hours saved for future team members ### ROI Calculation: - Monthly time savings: _____ hours - Hourly value: $_____ (your rate or team average) - Monthly value created: $_____ - Framework investment recovery: _____ months - Annual ROI: _____% ``` ### **Framework vs Non-Framework Comparison** ```markdown ## Systematic vs Ad-Hoc Problem Solving ### Framework-Applied Problems (Last 30 days): Problem | Attempts | Time | Success | Quality Score --------|----------|------|---------|------------- [Problem 1] | 1 | 2h | | 9/10 [Problem 2] | 1 | 1.5h | | 8/10 ### Non-Framework Problems (Same period): Problem | Attempts | Time | Success | Quality Score --------|----------|------|---------|------------- [Problem A] | 4 | 6h | ⚠️ | 6/10 [Problem B] | 3 | 4h | | 5/10 ### Comparison Summary: - Average attempts: Framework _____ vs Non-framework _____ - Average time: Framework _____ vs Non-framework _____ - Success rate: Framework _____% vs Non-framework _____% - Quality score: Framework _____/10 vs Non-framework _____/10 ``` --- ## **BUSINESS IMPACT MEASUREMENT** ### **Project-Level Impact Assessment** ```markdown ## Framework Impact on Project Outcomes ### Project: [Project Name] ### Duration: [Start] - [End] ### Framework Usage: [Full/Partial/None] ### Delivery Metrics: - On-time delivery: Yes/No - Budget adherence: Under/On/Over by _____% - Scope creep incidents: _____ - Major revisions required: _____ ### Quality Metrics: - Post-deployment bugs: _____ (vs _____ project average) - Customer satisfaction: _____/10 (vs _____ average) - Performance requirements met: _____% - Security/compliance issues: _____ ### Team Metrics: - Developer satisfaction: _____/10 - Knowledge transfer effectiveness: _____/10 - Maintainability rating: _____/10 - Documentation quality: _____/10 ``` --- ## **CONTINUOUS IMPROVEMENT TRACKING** ### **Weekly Framework Review Template** ```markdown ## Weekly Framework Effectiveness Review ### Week of: [Date] ### Quantitative Metrics: - Problems solved: _____ - Framework applications: _____ - Success rate: _____% - Average time per problem: _____ hours - Quality score average: _____/10 ### Qualitative Observations: - Most effective rule this week: #_____ - Most challenging application: [Description] - Biggest insight gained: [Insight] - Framework confidence level: _____/10 ### Next Week Goals: - Focus rule for improvement: #_____ - Template to practice: [Template name] - Measurement to track: [Specific metric] ``` ### **Framework Evolution Tracking** ```markdown ## Framework Adaptation and Improvement ### Month 1 Baseline: - Framework files used: _____ - Rules consistently applied: _____/15 - Templates actively used: _____ - Custom adaptations created: _____ ### Monthly Evolution Analysis: - Framework mastery progression: _____% - Custom adaptation effectiveness: _____/10 - Framework confidence level: _____/10 - Willingness to recommend: _____/10 ``` --- **🧠 STATUS:** **ASSESSMENT SYSTEM OPERATIONAL** *Use this comprehensive assessment system to track framework mastery development, measure effectiveness, and demonstrate ROI across individual, project, and organizational levels.*