UNPKG

@astermind/astermind-pro

Version:

Astermind Pro - Premium ML Toolkit with Advanced RAG, Reranking, Summarization, and Information Flow Analysis

580 lines (430 loc) 17.4 kB
# Astermind Pro - Premium Features This document lists all premium features extracted from the `astermind-kelm-elm-demo` that are **not included** in the base `@astermind/astermind-elm` package. --- ## Overview The demo showcases **AsterMind Omega**, an advanced RAG (Retrieval-Augmented Generation) system built on top of Astermind ELM. These premium features extend the base ELM capabilities with: - Advanced retrieval and reranking - Sophisticated summarization - Information flow analysis - Production-grade numerical methods - Advanced text processing --- ## 1. Omega RAG System ### 1.1 Omega (`Omega.ts`) **Advanced RAG answer composition with recursive compression** **Features:** - Recursive sentence compression using online ridge regression - Multi-round summarization with weighted sentence selection - Query-aligned sentence scoring with cosine similarity - Lexical bonus for overlapping query terms - Personality modes (neutral, teacher, scientist) - Deterministic, context-locked summarization **Not in base ELM:** Base ELM has no RAG or summarization capabilities. --- ### 1.2 OmegaRR (`OmegaRR.ts`) **Production-grade reranking system with engineered features** **Features:** - **Rich Feature Engineering:** - TF-IDF and BM25 sparse similarity - Heading-query match scores - Jaccard token overlap - Code block detection flags - Structural signals (Go code, links, etc.) - Random projection dense hints - Length heuristics - Prior score integration - **Weak Supervision:** - Automatic label generation from heuristics - Per-query ridge model training - Relevance probability estimation - **MMR (Maximal Marginal Relevance) Filtering:** - Diversity-aware selection - Character budget constraints - Coverage optimization - **Feature Exposure:** - Optional feature vector export - Feature name mapping - Diagnostic utilities **Not in base ELM:** Base ELM has no reranking, feature engineering, or MMR capabilities. --- ### 1.3 OmegaSumDet (`OmegaSumDet.ts`) **Deterministic, intent-aware summarization** **Features:** - **Intent Detection:** - Function, variable, constant detection - Concurrency pattern recognition - Loop detection - **Code-Aware Processing:** - Atomic code block handling - Intent-aware code gating - Code relevance scoring - Query-aligned code inclusion - **Advanced Text Processing:** - Stemmed Dice coefficient for heading alignment - Stopword-aware tokenization - Jaccard deduplication - Section diversity capping - **Deterministic Scoring:** - Normalized feature weights - Explicit tie-breakers - Stable ordering guarantees - Context-locked (no leakage) - **Output Shaping:** - Character budget management - Bullet point formatting - Citation generation - Footer with sources **Not in base ELM:** Base ELM has no summarization, intent detection, or code-aware processing. --- ## 2. Advanced Numerical Methods ### 2.1 KRR - Kernel Ridge Regression (`krr.ts`) **Production-grade ridge regression solver** **Features:** - Cholesky decomposition with adaptive jitter - Conjugate Gradient (CG) fallback for ill-conditioned systems - Symmetry enforcement - Matrix validation (NaN/Inf detection) - Abort signal support for cancellation - Comprehensive diagnostics **Not in base ELM:** Base ELM uses simpler ridge solving without these production features. --- ### 2.2 RFF - Random Fourier Features (`rff.ts`) **RBF kernel approximation via random projections** **Features:** - Deterministic RFF construction - Box-Muller Gaussian sampling - L2 normalization for stability - Efficient cosine/sine feature mapping **Not in base ELM:** Base ELM's KernelELM uses exact kernels or Nyström, but not RFF. --- ### 2.3 OnlineRidge (`online_ridge.ts`) **Online ridge regression with rank-1 updates** **Features:** - Incremental updates via Sherman-Morrison formula - Efficient inverse maintenance - Multi-output support (stacked heads) - No full retraining required **Not in base ELM:** Base OnlineELM uses RLS but not this specific online ridge implementation. --- ### 2.4 Advanced Math Utilities (`math.ts`) **Production-grade numerical operations** **Features:** - Robust vector operations (dot, add, scale, Hadamard) - In-place variants for GC efficiency - Safe exp/log/sigmoid with overflow guards - Stable softmax via log-sum-exp trick - Hyperbolic distance proxy - Vector normalization and clamping - Statistical functions (mean, variance, standardization) - Top-K selection - Formatting utilities **Not in base ELM:** Base ELM has basic math but not these production-grade utilities. --- ## 3. Information Flow Analysis ### 3.1 Transfer Entropy (`infoflow/TransferEntropy.ts`) **Information-theoretic causal analysis** **Features:** - Streaming Transfer Entropy (TE) estimation - Linear-Gaussian approximation - Configurable lag windows - Ridge-regularized regression - Bits/nats reporting - Multi-variable monitoring - InfoFlow graph construction **Use Cases:** - Monitor information flow between RAG components - Query → Score influence tracking - Feature → Relevance analysis - Kept chunks → Summary grounding verification **Not in base ELM:** Base ELM has no information flow or causal analysis capabilities. --- ### 3.2 Transfer Entropy PWS (`infoflow/TransferEntropyPWS.ts`) **Phase-Weighted Stacking variant with importance sampling** **Features:** - **Importance Sampling:** - Rare event detection via tail quantile thresholds - Tail boost weighting for important samples - Time decay for recency weighting - **Path-Weight Sampling (PWS):** - Jittered context histories - Multiple perturbed path averaging - Kernel Density Estimation (KDE) for conditional entropy - Silverman's rule-of-thumb bandwidth selection - **KDE-Based Estimation:** - Product Gaussian kernels - Joint and marginal density estimation - Ridge floor to avoid log(0) issues - Importance-weighted averaging **Not in base ELM:** Advanced variant with importance sampling and PWS not available in base. --- ### 3.3 TE Controller (`infoflow/TEController.ts`) **Closed-loop adaptive control via Transfer Entropy** **Features:** - **Target Band Management:** - Configurable TE target bands per channel - Query→Score, Feature→Score, Kept→Summary monitoring - Optional loop stability guards - **Adaptive Parameter Tuning:** - EMA (Exponential Moving Average) smoothing - Single-knob adjustments per step - Cooldown periods between adjustments - Hard caps on session adjustments - **Knob Control:** - Alpha (sparse/dense mix) - Sigma (kernel bandwidth) - Ridge (regularization) - ProbThresh (reranker threshold) - MMR Lambda (diversity tradeoff) - BudgetChars (answer length) - **Safety Features:** - Minimum sample requirements - Parameter limits enforcement - Step size constraints - Adjustment history tracking **Use Cases:** - Automatic hyperparameter tuning based on information flow - Maintaining optimal TE ranges for system health - Preventing information leakage or overfitting - Adaptive system optimization **Not in base ELM:** No closed-loop control or TE-based adaptation in base. --- ## 4. Advanced Retrieval **NEW:** All retrieval functionality is now available as **standalone, reusable modules** outside of workers! Use `buildIndex()`, `hybridRetrieve()`, and related functions directly in your applications. ### 4.1 Hybrid Retrieval System **Combining sparse and dense methods** **Features:** - **TF-IDF Sparse Retrieval:** - Heading-weighted tokenization - Custom vocabulary management - Stemming support - **Dense Kernel Retrieval:** - Nyström approximation for scalability - Multiple kernel types (RBF, cosine, polynomial) - Landmark-based projection - **Hybrid Scoring:** - Ridge-regularized combination - Alpha/beta mixing parameters - Keyword bonus integration - Tanh clipping for stability **Not in base ELM:** Base ELM has EmbeddingStore but not this hybrid retrieval pipeline. **Available Modules:** - `buildIndex()` - Build vocabulary, IDF, and Nyström landmarks from documents - `hybridRetrieve()` - Perform hybrid retrieval (sparse + dense + keyword bonus) - `toTfidf()` - Compute TF-IDF vectors - `cosineSparse()` - Sparse vector cosine similarity - `projectToDense()` - Project sparse vectors to dense using Nyström landmarks - `parseMarkdownToSections()` - Parse markdown into hierarchical sections - `flattenSections()` - Flatten sections into chunks --- ### 4.2 Nyström Approximation **Efficient kernel computation for large datasets** **Features:** - Landmark selection strategies - Sparse-to-dense projection - Kernel similarity computation - Normalized feature vectors - Configurable landmark count **Not in base ELM:** Base KernelELM has Nyström but not integrated into a full retrieval system. --- ## 5. Advanced Text Processing ### 5.1 Tree-Aware Markdown Parsing **Hierarchical section extraction** **Features:** - Multi-level heading parsing (##, ###, etc.) - Parent-child relationship tracking - Empty parent backfilling - Rich + plain text retention - Code fence preservation - Link handling **Not in base ELM:** Base ELM has basic text encoding but not markdown parsing. --- ### 5.2 Advanced Tokenization **Production-grade text preprocessing** **Features:** - **Stemming:** - Plural folding (ies → y, etc.) - Suffix removal (ization → ize) - Conservative rule-based stemming - Memoization for performance - **Query Expansion:** - Domain-specific term expansion - Context-aware keyword addition - Go-specific expansions - **Code-Aware Splitting:** - Fenced code block detection - Atomic code block handling - Mixed text/code processing **Not in base ELM:** Base ELM has UniversalEncoder but not this advanced preprocessing. --- ### 5.3 ELM Scorer (`elm_scorer.ts`) **Custom ELM-based relevance scoring** **Features:** - Random hidden layer initialization - GELU activation approximation - Online ridge output layer - Partial fit support - Deterministic seeding **Not in base ELM:** This is a custom implementation, though base ELM could be used similarly. --- ## 6. Auto-Tuning System **NEW:** Auto-tuning is now available as a **standalone function** outside of workers! Use `autoTune()` directly in your applications. ### 6.1 Hyperparameter Optimization **Automated configuration search** **Features:** - Random search with refinement - Jaccard-based evaluation metric - Ridge-aware parameter exploration - Caching for efficiency - Penalty functions for complexity - Real-time progress reporting **Parameters Tuned:** - Alpha (sparse/dense mix) - Beta (keyword bonus) - Sigma (kernel bandwidth) - Kernel type - Vocabulary size - Landmark count - Prefilter size - Top-K selection - Heading weights - Chunk/overlap sizes - Ridge regularization **Not in base ELM:** Base ELM has no auto-tuning capabilities. **Available Functions:** - `autoTune()` - Automated hyperparameter optimization - `sampleQueriesFromCorpus()` - Generate synthetic queries for tuning - `penalty()` - Compute complexity penalty for configurations - `jaccard()` - Calculate Jaccard similarity between index sets --- ## 7. Model Persistence **NEW:** Model serialization is now available as **standalone functions** outside of workers! Use `exportModel()` and `importModel()` directly in your applications. ### 7.1 Serialized Model Format **Complete model export/import** **Features:** - Full state serialization - Version tracking - Optional dense vector storage - Checksum generation - Settings snapshot - Vocabulary + IDF preservation - Nyström landmark storage - Chunk metadata retention **Not in base ELM:** Base ELM has JSON import/export but not this comprehensive format. **Available Functions:** - `exportModel()` - Export complete model state to serialized format - `importModel()` - Import model from serialized format - `quickHash()` - Generate deterministic hash for model verification --- ## 8. Worker Architecture ### 8.1 Web Worker Pipeline **Background processing for RAG** **Features:** - Non-blocking retrieval - Async model loading - Progress reporting - Error handling - Message-based API - State management **Not in base ELM:** Base ELM has ELMWorker but not this full RAG pipeline. --- ## 9. Advanced ELM Variants ### 9.1 Deep ELM Pro (`elm/deep-elm-pro.ts`) **Improved multi-layer ELM with advanced training strategies** **Improvements over Base DeepELM:** | Feature | Base DeepELM | DeepELMPro | |---------|--------------|------------| | **Pretraining** | ❌ None | ✅ Autoencoder pretraining for each layer | | **Training Strategy** | Joint only | ✅ Layer-wise or Joint (configurable) | | **Regularization** | ❌ None | ✅ L1/L2/Elastic Net regularization | | **Batch Normalization** | ❌ None | ✅ Optional batch normalization between layers | | **Dropout** | ❌ None | ✅ Optional dropout with configurable rate | | **Training Flexibility** | Basic | ✅ Advanced with multiple strategies | **Key Features:** - **Autoencoder Pretraining**: Each layer can be pretrained as an autoencoder to learn better feature representations before supervised training - **Layer-wise Training**: Train layers sequentially for more stable learning and better feature extraction - **Regularization**: L1, L2, and Elastic Net regularization to prevent overfitting and improve generalization - **Batch Normalization**: Normalize activations between layers for faster convergence and training stability - **Dropout**: Randomly drop neurons during training to reduce overfitting - **Flexible Training Modes**: Choose between layer-wise sequential training or joint training based on your data **Not in base ELM:** Base DeepELM has basic multi-layer support but lacks these advanced training strategies, regularization, and normalization techniques. **Use Cases:** - Complex pattern recognition requiring deep feature hierarchies - High-dimensional data classification - When base DeepELM overfits or doesn't converge well - Production systems requiring robust, generalizable models --- ### 9.2 Multi-Kernel ELM (`elm/multi-kernel-elm.ts`) **Combines multiple kernel types for improved accuracy** **Features:** - Weighted combination of multiple kernels (RBF, linear) - Automatic kernel weight learning based on validation performance - Multiple kernel types support - Ridge-regularized combination **Not in base ELM:** Base KernelELM uses a single kernel type. Multi-Kernel ELM combines multiple kernels for better performance. --- ### 9.3 Online Kernel ELM (`elm/online-kernel-elm.ts`) **Real-time learning for streaming data** **Features:** - Incremental kernel matrix updates - Sliding window with forgetting mechanisms - Adaptive landmark selection - Real-time prediction **Not in base ELM:** Base KernelELM requires batch training. Online Kernel ELM supports streaming updates. --- ### 9.4 Multi-Task ELM (`elm/multi-task-elm.ts`) **Joint learning across related tasks** **Features:** - Shared feature extraction layer - Task-specific output layers - Task weighting for importance - Joint optimization **Not in base ELM:** Base ELM trains one task at a time. Multi-Task ELM learns multiple related tasks simultaneously. --- ### 9.5 Sparse ELM (`elm/sparse-elm.ts`) **Efficiency and interpretability for high-dimensional data** **Features:** - L1/L2/Elastic net regularization - Weight pruning for sparsity - Feature importance ranking - Interpretable models **Not in base ELM:** Base ELM has no built-in sparsity mechanisms. Sparse ELM provides regularization and feature selection. --- ## Summary: Premium Feature Categories ### Core Premium Features: 1.**Omega RAG System** - Complete RAG pipeline 2.**OmegaRR Reranking** - Production reranking with MMR 3.**OmegaSumDet** - Intent-aware summarization 4.**Transfer Entropy** - Information flow analysis 5.**Hybrid Retrieval** - Sparse + dense combination (**NEW: Standalone modules!**) 6.**Auto-Tuning** - Hyperparameter optimization (**NEW: Standalone function!**) 7.**Advanced Math** - Production-grade numerics 8.**Tree-Aware Parsing** - Hierarchical markdown processing (**NEW: Standalone functions!**) 9.**Model Serialization** - Export/import models (**NEW: Standalone functions!**) 10.**Tokenization & Utilities** - Reusable text processing (**NEW: Standalone functions!**) 11.**Advanced ELM Variants** - 5 premium variants (**NEW!**) ### Advanced Algorithms: - KRR (Kernel Ridge Regression with CG fallback) - RFF (Random Fourier Features) - OnlineRidge (Rank-1 updates) - Nyström approximation - MMR filtering - Weak supervision ### Production Features: - Robust error handling - Matrix validation - Abort signals - Comprehensive diagnostics - Model versioning - Checksums --- ## Integration Notes These premium features are designed to work **on top of** `@astermind/astermind-elm`: - Uses base ELM's `Tokenizer`, `TFIDFVectorizer`, `ELM`, `OnlineELM` - Extends with premium retrieval, reranking, and summarization - Adds information flow monitoring - Provides production-grade numerical methods --- ## License & Usage These premium features are part of **Astermind Pro** and are not included in the MIT-licensed Community Edition (`@astermind/astermind-elm`). --- *Last updated: 2025-01-16*