Files
XendiAI-Plugin/CHANGELOG.md
2025-10-05 09:16:08 +00:00

287 lines
11 KiB
Markdown

# Change Log
All notable changes to the "Xendi AI" extension will be documented in this file.
## [1.2.1] - 2025-10-03
### 🧹 Packaging Improvements
#### Bug Fixes
- Fixed "nul" file causing unsafe extraction error
- Removed temporary build folders from VSIX package
- Cleaned up .vscodeignore to exclude development files
#### Package Optimization
- Removed `temp_extract/` and `temp_latest/` directories
- Excluded `*.log` files
- Excluded `IMPROVEMENTS_ROADMAP.md` (development doc)
- Excluded `tsconfig.json.disabled`
- Package size reduced from 2.8 MB to 2.73 MB
---
## [1.2.0] - 2025-10-03
### 🌐 Multi-Provider Support - Major Release
#### ✨ 8 AI Provider Support
- **LM Studio** - Run models locally on your machine (existing)
- **Ollama** - Simple local AI with easy model management (existing)
- **OpenAI** - GPT-4, GPT-4 Turbo, GPT-3.5 Turbo (NEW)
- **OpenRouter** - Access to 100+ models including GPT-4, Claude, Gemini (NEW)
- **Anthropic** - Claude 3.5 Sonnet, Claude 3 Opus (NEW)
- **Cohere** - Command R+, Command R (NEW)
- **Groq** - Ultra-fast inference with Llama, Mixtral (NEW)
- **Mistral** - Mistral Large, Mistral Medium (NEW)
#### ⚡ Productivity Features
- **Real-time Streaming**: Streaming responses for all cloud providers (except Ollama)
- **Keyboard Shortcuts**: 15+ shortcuts for maximum productivity
- `Ctrl+Enter` - Send message
- `Ctrl+K` - New chat
- `Ctrl+L` - Clear chat
- `Ctrl+/` - Focus input
- `Ctrl+Shift+C` - Copy last response
- `Escape` - Cancel request
- `Ctrl+↑/↓` - Navigate message history
- `Ctrl+1/2/3/4` - Switch tabs
- `Ctrl+Shift+M` - Toggle Chat/Agent mode
- **Token Counter**: Live token counting with cost estimation per model
- Per-model pricing for GPT-4, Claude, Cohere, etc.
- Session statistics tracking
- Accurate estimates for code vs text
- **Theme Sync**: Automatic light/dark theme sync with VS Code
#### 🎨 UI/UX Improvements
- **Provider Selection**: AI Provider dropdown in Settings tab
- **API Key Management**: Secure API key input for cloud providers
- **Fixed Minimum Sizes**: Consistent UI sizing (600px body, 400px tabs, 200px chat)
- **Better Tab Switching**: Fixed tab navigation bugs
- **Null-safe JavaScript**: Comprehensive null checks for all DOM elements
#### 🔧 Technical Improvements
- **Provider-specific Authentication**:
- Bearer tokens for OpenAI, OpenRouter, Groq, Mistral
- x-api-key header for Anthropic
- Authorization header for Cohere
- **Smart Default URLs**: Each provider has pre-configured default endpoints
- **Default Models**: Pre-configured models for Anthropic and Cohere
- **Streaming Implementation**: Full SSE (Server-Sent Events) parsing for real-time responses
- **Optimized Logging**: Structured logging for better debugging
#### 🐛 Bug Fixes
- Fixed tab switching not working (ID mismatch resolved)
- Fixed null reference errors in addEventListener calls
- Fixed TypeScript compilation overwriting extension.js
- Fixed streaming cursor animation
- Fixed settings not persisting correctly
#### 📚 Documentation
- Completely overhauled README.md with 8 provider setup guides
- Added keyboard shortcuts documentation
- Added token counter and cost estimation docs
- Updated troubleshooting section
- Provider comparison table
### 🔄 Migration Notes
- Existing LM Studio and Ollama configurations will continue to work
- New users can choose any of the 8 supported providers
- API keys are stored securely in VS Code globalState
- No breaking changes to existing functionality
---
## [1.1.3] - 2025-10-02
### 🐍 Python Virtual Environment Support
- **Automatic venv Detection**: Agent now detects `venv/`, `.venv/`, and `env/` directories
- **Auto-Activation**: Python virtual environments are automatically activated before running commands
- **Seamless pip/python**: No need to manually activate venv - just use `pip install` or `python` directly
- **Smart Command Wrapping**: Commands are wrapped with `source venv/bin/activate &&` automatically
### 🔧 Developer Experience
- **Agent knows about venv**: Updated system prompt to inform agent about automatic venv activation
- **Better Python Development**: Agent can now install packages and run Python code without environment issues
- **Cross-Project Support**: Works with any Python project structure (venv, .venv, or env)
---
## [1.1.2] - 2025-10-02
### 🎯 Agent Improvements
- **Fixed Topic Switching**: Agent now stays focused on one task until completion
- **No More Task Jumping**: Agent won't suggest or start additional work unless requested
- **Laser Focus**: Clear instructions to work ONLY on the current user request
### 🎨 UI Improvements
- **Command Result Dropdowns**: All agent tool outputs now display in collapsible dropdowns
- **Tool Icons**: Visual icons for each command type (🔍 glob, 📄 readFile, ⚡ runShellCommand, etc.)
- **Cleaner Agent Chat**: Command results are collapsed by default for better readability
- **Consistent Design**: Dropdown style matches thinking process display
### 🐛 Bug Fixes
- Fixed agent wandering off-topic during multi-step tasks
- Improved command output formatting in agent mode
---
## [1.1.1] - 2025-10-02
### 🚀 Performance Improvements
- **Optimized API Requests**: Added connection keep-alive and request timeouts (60s) for faster communication
- **Reduced Token Limits**: Decreased max_tokens from 2000 to 1000 for faster response times
- **Better Server Communication**: Improved fetch configuration with abort controllers
### 🎨 UI Improvements
- **Redesigned Statistics Panel**: New clean, professional design matching extension theme
- **Removed Emojis**: Statistics now use text labels (FILES, LINES, TOKENS) instead of emoji icons
- **Enhanced Styling**: Monospace font for values, uppercase labels, purple accent colors
- **Better Layout**: Vertical stat items with clear separation and improved readability
### 🐛 Bug Fixes
- Fixed statistics panel visual consistency with extension design
- Improved stats visibility toggle with CSS classes
---
## [1.1.0] - 2025-10-02
### ✨ New Features
#### 🧠 Thinking Process Display
- **AI Reasoning Visualization**: Models that output thinking process (using `<think>` tags) now show expandable dropdown
- **Transparent AI Decisions**: See exactly how the AI reasons through problems
- **Collapsible Interface**: Keeps chat clean while making reasoning available on-demand
#### 📊 Agent Statistics & Monitoring
- **Real-time Stats Display**: See files created/modified, lines added/deleted, tokens used
- **Task Plan View**: Visual overview of agent's execution plan and progress
- **File Diff Links**: Clickable file paths with diff view (like GitHub Copilot)
- **Performance Metrics**: Track token usage and operation efficiency
#### 🔄 Ollama Support
- **Multiple LLM Backends**: Now supports both LM Studio and Ollama
- **Server Type Selection**: Easy switching between LM Studio (`localhost:1234`) and Ollama (`localhost:11434`)
- **Auto-detection**: Automatically handles different API formats
#### 💬 Enhanced Conversation Management
- **Auto-generated Titles**: Titles automatically created from first message
- **Timestamp Display**: Shows relative time ("5 Min.", "2 days") or full date/time
- **Context Preservation**: Full conversation history maintained across all messages
- **Delete with Confirmation**: Safe conversation deletion with German/English prompts
- **Quick Actions**: "New Conversation" button in chat for instant access
#### 🎨 Markdown Rendering
- **Full Markdown Support**: Renders code blocks, lists, headers, bold, italic
- **Syntax Highlighting**: Language-specific code block rendering
- **Clean Formatting**: Professional display of AI responses
#### 🌐 Context Reset Commands
- `vergiss alles` (German)
- `alles vergessen` (German)
- `reset context` (English)
- `clear context` (English)
- `forget everything` (English)
### 🐛 Bug Fixes
- Fixed conversation context not persisting between messages
- Fixed agent losing context during multi-step tasks
- Improved error handling for network issues
### 🔧 Technical Improvements
- Better API compatibility (LM Studio + Ollama)
- Optimized conversation history handling
- Enhanced markdown rendering engine
- Improved statistics tracking
---
## [1.0.0] - 2025-10-02
### 🎉 Initial Release
**Xendi AI - Autonomous AI Coding Assistant for VS Code**
### ✨ Features
#### 🤖 Agent Mode
- **Autonomous task completion** - Works like GitHub Copilot/Claude
- **Multi-step workflow** - Explores, analyzes, fixes, and tests automatically
- **Smart tool execution** - Automatically reads files, finds bugs, implements fixes
- **Continuous iteration** - Keeps working until task is 100% complete
- **Workspace awareness** - Knows project structure and context
#### 💬 Chat Mode
- Interactive AI assistant for coding questions
- Code explanations and suggestions
- Real-time code generation
#### 🛠️ Powerful Tools
- **File Operations**: Read, write, search, replace
- **Code Analysis**: Pattern search, bug detection
- **Shell Commands**: Run tests, builds, install packages
- **Web Integration**: Google search, web content fetching
- **Memory**: Save and recall information
#### 🌍 Universal Language Support
- JavaScript/TypeScript
- Python
- Java, C++, C#
- Go, Rust
- PHP, Ruby, Swift, Kotlin
- And many more...
#### 🔒 Privacy & Security
- 100% local execution with LM Studio
- No cloud services
- No telemetry or tracking
- Your code stays on your machine
#### 🎨 User Experience
- Clean, intuitive interface
- Multilingual UI (English/German)
- Comprehensive logging system
- Auto-approve tools for seamless workflow
- Model selection and switching
### 🔧 Technical Details
- **LM Studio Integration**: Seamless connection to local LLM server
- **Workspace Tools**: Full VSCode workspace API integration
- **Conversation History**: Persistent chat and agent conversations
- **Error Handling**: Robust error detection and recovery
- **Performance**: Optimized file operations with caching
### 📝 Configuration
- Customizable LM Studio server URL
- Language selection (English/German)
- Auto-approve settings for tools
- Model temperature and token settings
### 🎯 Best For
- **Debugging**: Find and fix bugs automatically
- **Refactoring**: Improve code structure and quality
- **Feature Development**: Build new features end-to-end
- **Learning**: Understand code and get explanations
- **Testing**: Run and fix test failures
### 🙏 Credits
- Built with LM Studio API
- Inspired by GitHub Copilot and Claude
- Community feedback and testing
---
**Developer**: Robin Oliver Lucas (RL-Dev)
**License**: MIT
**Website**: https://rl-dev.de
---
For support and feedback:
- GitHub Issues: https://github.com/RL-Dev/xendi-ai/issues
- Website: https://rl-dev.de