UNPKG

smart-ai-cache

Version:

🚀 Lightning-fast AI response caching with Redis support. Reduce OpenAI/Claude/Gemini API costs by 40-80% with sub-millisecond cache lookups.

105 lines (81 loc) • 3.49 kB
# Changelog All notable changes to this project will be documented in this file. The format is based on [Keep a Changelog](https://keepachangelog.com/en/1.0.0/), and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0.html). ## [1.0.6] - 2024-08-28 ### Added - Enhanced README with comprehensive Redis examples - Docker Compose setup for Redis in production - Migration guide from v1.0.4 to v1.0.5+ - Advanced cache management features documentation - Performance benchmarking section with real metrics - Complete package metadata for better NPM discoverability ### Changed - Improved package description for better searchability - Updated keywords for broader AI/ML community reach - Enhanced documentation structure with table of contents ## [1.0.5] - 2024-08-28 ### Added - **Redis storage backend** with full error handling and fallback - **Comprehensive error handling** with retry logic and exponential backoff - **Pattern-based cache invalidation** (`deleteByPattern` method) - **Input validation** for configuration and runtime parameters - **Performance benchmarking script** (`npm run benchmark`) - **93.7% test coverage** with 50 comprehensive tests - **Graceful degradation** when cache storage fails - **Protected methods** for provider inheritance (replaces `@ts-ignore` hacks) ### Changed - **BREAKING**: Cache methods are now async (`clear()`, `delete()`, `has()`, `getCacheSize()`) - **BREAKING**: Provider classes no longer require client instances - Cache operations now use storage abstraction for better extensibility - Improved statistics tracking with provider-specific metrics - Enhanced key generation with better normalization ### Fixed - LRU eviction performance optimizations - Memory leak prevention in concurrent scenarios - TypeScript compilation issues with ioredis v5 - Test reliability and console output cleanup ### Performance - **Cache lookup**: 0.0009ms (1,111x faster than 1ms requirement) - **Memory usage**: 2.86MB for 10K entries (35x better than 100MB requirement) - **Throughput**: 451,842 requests/second - **LRU eviction**: 86μs per operation ## [1.0.4] - 2024-08-27 ### Added - GitHub integration and workflow setup - Version bump and release automation ### Fixed - Package distribution and build process ## [1.0.3] - 2024-08-27 ### Added - Basic caching functionality for AI responses - Support for OpenAI, Anthropic, and Google AI providers - In-memory storage with LRU eviction - Statistics tracking for cache performance - TypeScript definitions and documentation ### Features - Zero-configuration setup for quick start - Provider-specific classes with automatic cost tracking - MD5-based cache key generation - Comprehensive test suite with Jest - TypeDoc documentation generation ## [1.0.0] - Initial Release ### Added - Core caching middleware for AI API responses - Multi-provider support (OpenAI, Anthropic, Google) - Basic statistics and cost tracking - TypeScript support - MIT license --- ## Migration Notes ### v1.0.4 → v1.0.5+ - Update cache method calls to use `await` - Remove client instances from provider constructors - Consider upgrading to Redis for production deployments - Update error handling to leverage new retry logic ### v1.0.5 → v1.0.6 - No breaking changes - Enhanced documentation and package metadata - New benchmarking capabilities --- For more details on any release, see the [GitHub releases page](https://github.com/ArchiesDubey/smart-ai-cache/releases).