This release introduces significant improvements to configuration flexibility, performance, and stability.
**Key Changes:**
* **Smart Configuration:**
* Added `summary_model_max_context` to allow independent context limits for the summary model (e.g., using `gemini-flash` with 1M context to summarize `gpt-4` history).
* Implemented auto-detection of base model settings for custom models, ensuring correct threshold application.
* **Performance & Refactoring:**
* Optimized `model_thresholds` parsing with caching to reduce overhead.
* Refactored `inlet` and `outlet` logic to remove redundant code and improve maintainability.
* Replaced all `print` statements with proper `logging` calls for better production monitoring.
* **Bug Fixes & Modernization:**
* Fixed `datetime.utcnow()` deprecation warnings by switching to timezone-aware `datetime.now(timezone.utc)`.
* Corrected type annotations and improved error handling for `JSONResponse` objects from LLM backends.
* Removed hard truncation in summary generation to allow full context usage.
**Files Updated:**
* Plugin source code (English & Chinese)
* Documentation and READMEs
* Version bumped to 1.2.1