# Change Log All notable changes to the "Xendi AI" extension will be documented in this file. ## [1.2.1] - 2025-10-03 ### ๐Ÿงน Packaging Improvements #### Bug Fixes - Fixed "nul" file causing unsafe extraction error - Removed temporary build folders from VSIX package - Cleaned up .vscodeignore to exclude development files #### Package Optimization - Removed `temp_extract/` and `temp_latest/` directories - Excluded `*.log` files - Excluded `IMPROVEMENTS_ROADMAP.md` (development doc) - Excluded `tsconfig.json.disabled` - Package size reduced from 2.8 MB to 2.73 MB --- ## [1.2.0] - 2025-10-03 ### ๐ŸŒ Multi-Provider Support - Major Release #### โœจ 8 AI Provider Support - **LM Studio** - Run models locally on your machine (existing) - **Ollama** - Simple local AI with easy model management (existing) - **OpenAI** - GPT-4, GPT-4 Turbo, GPT-3.5 Turbo (NEW) - **OpenRouter** - Access to 100+ models including GPT-4, Claude, Gemini (NEW) - **Anthropic** - Claude 3.5 Sonnet, Claude 3 Opus (NEW) - **Cohere** - Command R+, Command R (NEW) - **Groq** - Ultra-fast inference with Llama, Mixtral (NEW) - **Mistral** - Mistral Large, Mistral Medium (NEW) #### โšก Productivity Features - **Real-time Streaming**: Streaming responses for all cloud providers (except Ollama) - **Keyboard Shortcuts**: 15+ shortcuts for maximum productivity - `Ctrl+Enter` - Send message - `Ctrl+K` - New chat - `Ctrl+L` - Clear chat - `Ctrl+/` - Focus input - `Ctrl+Shift+C` - Copy last response - `Escape` - Cancel request - `Ctrl+โ†‘/โ†“` - Navigate message history - `Ctrl+1/2/3/4` - Switch tabs - `Ctrl+Shift+M` - Toggle Chat/Agent mode - **Token Counter**: Live token counting with cost estimation per model - Per-model pricing for GPT-4, Claude, Cohere, etc. - Session statistics tracking - Accurate estimates for code vs text - **Theme Sync**: Automatic light/dark theme sync with VS Code #### ๐ŸŽจ UI/UX Improvements - **Provider Selection**: AI Provider dropdown in Settings tab - **API Key Management**: Secure API key input for cloud providers - **Fixed Minimum Sizes**: Consistent UI sizing (600px body, 400px tabs, 200px chat) - **Better Tab Switching**: Fixed tab navigation bugs - **Null-safe JavaScript**: Comprehensive null checks for all DOM elements #### ๐Ÿ”ง Technical Improvements - **Provider-specific Authentication**: - Bearer tokens for OpenAI, OpenRouter, Groq, Mistral - x-api-key header for Anthropic - Authorization header for Cohere - **Smart Default URLs**: Each provider has pre-configured default endpoints - **Default Models**: Pre-configured models for Anthropic and Cohere - **Streaming Implementation**: Full SSE (Server-Sent Events) parsing for real-time responses - **Optimized Logging**: Structured logging for better debugging #### ๐Ÿ› Bug Fixes - Fixed tab switching not working (ID mismatch resolved) - Fixed null reference errors in addEventListener calls - Fixed TypeScript compilation overwriting extension.js - Fixed streaming cursor animation - Fixed settings not persisting correctly #### ๐Ÿ“š Documentation - Completely overhauled README.md with 8 provider setup guides - Added keyboard shortcuts documentation - Added token counter and cost estimation docs - Updated troubleshooting section - Provider comparison table ### ๐Ÿ”„ Migration Notes - Existing LM Studio and Ollama configurations will continue to work - New users can choose any of the 8 supported providers - API keys are stored securely in VS Code globalState - No breaking changes to existing functionality --- ## [1.1.3] - 2025-10-02 ### ๐Ÿ Python Virtual Environment Support - **Automatic venv Detection**: Agent now detects `venv/`, `.venv/`, and `env/` directories - **Auto-Activation**: Python virtual environments are automatically activated before running commands - **Seamless pip/python**: No need to manually activate venv - just use `pip install` or `python` directly - **Smart Command Wrapping**: Commands are wrapped with `source venv/bin/activate &&` automatically ### ๐Ÿ”ง Developer Experience - **Agent knows about venv**: Updated system prompt to inform agent about automatic venv activation - **Better Python Development**: Agent can now install packages and run Python code without environment issues - **Cross-Project Support**: Works with any Python project structure (venv, .venv, or env) --- ## [1.1.2] - 2025-10-02 ### ๐ŸŽฏ Agent Improvements - **Fixed Topic Switching**: Agent now stays focused on one task until completion - **No More Task Jumping**: Agent won't suggest or start additional work unless requested - **Laser Focus**: Clear instructions to work ONLY on the current user request ### ๐ŸŽจ UI Improvements - **Command Result Dropdowns**: All agent tool outputs now display in collapsible dropdowns - **Tool Icons**: Visual icons for each command type (๐Ÿ” glob, ๐Ÿ“„ readFile, โšก runShellCommand, etc.) - **Cleaner Agent Chat**: Command results are collapsed by default for better readability - **Consistent Design**: Dropdown style matches thinking process display ### ๐Ÿ› Bug Fixes - Fixed agent wandering off-topic during multi-step tasks - Improved command output formatting in agent mode --- ## [1.1.1] - 2025-10-02 ### ๐Ÿš€ Performance Improvements - **Optimized API Requests**: Added connection keep-alive and request timeouts (60s) for faster communication - **Reduced Token Limits**: Decreased max_tokens from 2000 to 1000 for faster response times - **Better Server Communication**: Improved fetch configuration with abort controllers ### ๐ŸŽจ UI Improvements - **Redesigned Statistics Panel**: New clean, professional design matching extension theme - **Removed Emojis**: Statistics now use text labels (FILES, LINES, TOKENS) instead of emoji icons - **Enhanced Styling**: Monospace font for values, uppercase labels, purple accent colors - **Better Layout**: Vertical stat items with clear separation and improved readability ### ๐Ÿ› Bug Fixes - Fixed statistics panel visual consistency with extension design - Improved stats visibility toggle with CSS classes --- ## [1.1.0] - 2025-10-02 ### โœจ New Features #### ๐Ÿง  Thinking Process Display - **AI Reasoning Visualization**: Models that output thinking process (using `` tags) now show expandable dropdown - **Transparent AI Decisions**: See exactly how the AI reasons through problems - **Collapsible Interface**: Keeps chat clean while making reasoning available on-demand #### ๐Ÿ“Š Agent Statistics & Monitoring - **Real-time Stats Display**: See files created/modified, lines added/deleted, tokens used - **Task Plan View**: Visual overview of agent's execution plan and progress - **File Diff Links**: Clickable file paths with diff view (like GitHub Copilot) - **Performance Metrics**: Track token usage and operation efficiency #### ๐Ÿ”„ Ollama Support - **Multiple LLM Backends**: Now supports both LM Studio and Ollama - **Server Type Selection**: Easy switching between LM Studio (`localhost:1234`) and Ollama (`localhost:11434`) - **Auto-detection**: Automatically handles different API formats #### ๐Ÿ’ฌ Enhanced Conversation Management - **Auto-generated Titles**: Titles automatically created from first message - **Timestamp Display**: Shows relative time ("5 Min.", "2 days") or full date/time - **Context Preservation**: Full conversation history maintained across all messages - **Delete with Confirmation**: Safe conversation deletion with German/English prompts - **Quick Actions**: "New Conversation" button in chat for instant access #### ๐ŸŽจ Markdown Rendering - **Full Markdown Support**: Renders code blocks, lists, headers, bold, italic - **Syntax Highlighting**: Language-specific code block rendering - **Clean Formatting**: Professional display of AI responses #### ๐ŸŒ Context Reset Commands - `vergiss alles` (German) - `alles vergessen` (German) - `reset context` (English) - `clear context` (English) - `forget everything` (English) ### ๐Ÿ› Bug Fixes - Fixed conversation context not persisting between messages - Fixed agent losing context during multi-step tasks - Improved error handling for network issues ### ๐Ÿ”ง Technical Improvements - Better API compatibility (LM Studio + Ollama) - Optimized conversation history handling - Enhanced markdown rendering engine - Improved statistics tracking --- ## [1.0.0] - 2025-10-02 ### ๐ŸŽ‰ Initial Release **Xendi AI - Autonomous AI Coding Assistant for VS Code** ### โœจ Features #### ๐Ÿค– Agent Mode - **Autonomous task completion** - Works like GitHub Copilot/Claude - **Multi-step workflow** - Explores, analyzes, fixes, and tests automatically - **Smart tool execution** - Automatically reads files, finds bugs, implements fixes - **Continuous iteration** - Keeps working until task is 100% complete - **Workspace awareness** - Knows project structure and context #### ๐Ÿ’ฌ Chat Mode - Interactive AI assistant for coding questions - Code explanations and suggestions - Real-time code generation #### ๐Ÿ› ๏ธ Powerful Tools - **File Operations**: Read, write, search, replace - **Code Analysis**: Pattern search, bug detection - **Shell Commands**: Run tests, builds, install packages - **Web Integration**: Google search, web content fetching - **Memory**: Save and recall information #### ๐ŸŒ Universal Language Support - JavaScript/TypeScript - Python - Java, C++, C# - Go, Rust - PHP, Ruby, Swift, Kotlin - And many more... #### ๐Ÿ”’ Privacy & Security - 100% local execution with LM Studio - No cloud services - No telemetry or tracking - Your code stays on your machine #### ๐ŸŽจ User Experience - Clean, intuitive interface - Multilingual UI (English/German) - Comprehensive logging system - Auto-approve tools for seamless workflow - Model selection and switching ### ๐Ÿ”ง Technical Details - **LM Studio Integration**: Seamless connection to local LLM server - **Workspace Tools**: Full VSCode workspace API integration - **Conversation History**: Persistent chat and agent conversations - **Error Handling**: Robust error detection and recovery - **Performance**: Optimized file operations with caching ### ๐Ÿ“ Configuration - Customizable LM Studio server URL - Language selection (English/German) - Auto-approve settings for tools - Model temperature and token settings ### ๐ŸŽฏ Best For - **Debugging**: Find and fix bugs automatically - **Refactoring**: Improve code structure and quality - **Feature Development**: Build new features end-to-end - **Learning**: Understand code and get explanations - **Testing**: Run and fix test failures ### ๐Ÿ™ Credits - Built with LM Studio API - Inspired by GitHub Copilot and Claude - Community feedback and testing --- **Developer**: Robin Oliver Lucas (RL-Dev) **License**: MIT **Website**: https://rl-dev.de --- For support and feedback: - GitHub Issues: https://github.com/RL-Dev/xendi-ai/issues - Website: https://rl-dev.de