Skip to content

v1.0.0 - Stable Oasis

Choose a tag to compare

@dzivkovi dzivkovi released this 28 Sep 23:47
· 4 commits to main since this release

πŸŽ‰ First Stable Release - Production Ready!

✨ What's New

This is the first stable release of vLLM HuggingFace Bridge, a fork of the original HuggingFace VS Code extension optimized for enterprise and air-gapped deployments.

πŸš€ Key Features

  • Air-Gapped Operation: Complete offline functionality with local vLLM/TGI servers
  • Dual Mode Support: Seamlessly switch between local and HuggingFace cloud models
  • Enterprise Ready: Production-tested in secure environments
  • Optimized Package: 92KB (91% smaller than original)
  • Smart Token Management: Automatic allocation for small context models (2048+)
  • Enhanced Observability: Clear logging showing which model handles each request

πŸ“¦ Installation

Quick Install (Recommended)

  1. Download the VSIX file below
  2. In VS Code: Ctrl+Shift+P β†’ Extensions: Install from VSIX...
  3. Select the downloaded file and restart VS Code

Command Line

```bash
wget https://github.com/dzivkovi/vllm-huggingface-bridge/releases/download/v1.0.0/vllm-huggingface-bridge-1.0.0.vsix
code --install-extension vllm-huggingface-bridge-1.0.0.vsix
```

πŸ”§ Configuration

For Local vLLM/TGI Servers

```json
{
"huggingface.localEndpoint": "http://localhost:8000"
}
```

For HuggingFace Cloud

  1. Open GitHub Copilot Chat
  2. Click model picker β†’ "Manage Models..."
  3. Select "Hugging Face" provider
  4. Add your HF token

πŸ§ͺ Tested With

  • VS Code 1.104+
  • vLLM 0.5+
  • TGI 2.0+
  • Various GPUs: RTX 4060, A100, H100
  • Models: DeepSeek-Coder, Qwen, Mistral, Llama

πŸ“Š Technical Improvements

  • All 42 tests passing
  • ESLint configured for pragmatic development
  • Package size reduced from 954KB to 92KB
  • Enhanced logging for debugging
  • Fixed VS Code API compatibility issues

πŸ™ Credits

Built on the excellent foundation of HuggingFace VS Code Chat.

πŸ“ License

MIT License - See LICENSE.txt for details