Skip to content

Popular repositories Loading

  1. rmbg-1.4 rmbg-1.4 Public template

    State-of-the-art background removal model, designed to effectively separate foreground from background. <metadata> gpu: T4 | collections: ["HF Transformers"] </metadata>

    Python 21 12

  2. triton-co-pilot triton-co-pilot Public

    Generate Glue Code in seconds to simplify your Nvidia Triton Inference Server Deployments

    Python 20 3

  3. smaug-72b smaug-72b Public

    Smaug-72B topped the Hugging Face LLM leaderboard and it’s the first model with an average score of 80, making it the world’s best open-source foundation model. <metadata> gpu: A100 | collections: …

    Python 17 5

  4. qwq-32b-preview qwq-32b-preview Public template

    A 32B experimental reasoning model for advanced text generation and robust instruction following. <metadata> gpu: A100 | collections: ["vLLM"] </metadata>

    Python 17 7

  5. whisper-large-v3 whisper-large-v3 Public template

    State‑of‑the‑art speech recognition model for English, delivering transcription accuracy across diverse audio scenarios. <metadata> gpu: T4 | collections: ["CTranslate2"] </metadata>

    Python 16 15

  6. deepseek-r1-distill-qwen-32b deepseek-r1-distill-qwen-32b Public template

    A distilled DeepSeek-R1 variant built on Qwen2.5-32B, fine-tuned with curated data for enhanced performance and efficiency. <metadata> gpu: A100 | collections: ["vLLM"] </metadata>

    Python 16 34

Repositories

Showing 10 of 165 repositories
  • kokoro Public

    82M parameters lightweight text-to-speech (TTS) model that delivers high-quality voice synthesis. <metadata> gpu: A10 | collections: ["HF Transformers"] </metadata>

    inferless/kokoro’s past year of commit activity
    Python 0 0 0 0 Updated May 17, 2025
  • qwen3-14b Public

    14B model with hybrid approach to problem-solving with two distinct modes: "thinking mode," which enables step-by-step reasoning and "non-thinking mode," designed for rapid, general-purpose responses. <metadata> gpu: A100 | collections: ["vLLM"] </metadata>

    inferless/qwen3-14b’s past year of commit activity
    Python 0 0 0 0 Updated May 15, 2025
  • qwen2.5-omni-7b Public template

    An advanced end-to-end multimodal which can processes text, images, audio, and video inputs, generating real-time text and natural speech responses. <metadata> gpu: A100 | collections: ["HF Transformers"] </metadata>

    inferless/qwen2.5-omni-7b’s past year of commit activity
    Python 0 0 0 0 Updated May 12, 2025
  • qwen3-8b Public template

    Qwen3-8B is a language model that supports seamless switching between “thinking” mode-for advanced math, coding, and logical inference-and “non-thinking” mode for fast, natural conversation. <metadata> gpu: A100 | collections: ["HF Transformers"] </metadata>

    inferless/qwen3-8b’s past year of commit activity
    Python 0 1 0 0 Updated May 12, 2025
  • inferless/MCP-Google-Map-Agent’s past year of commit activity
    Python 0 0 0 0 Updated Apr 30, 2025
  • phi-4-multimodal-instruct Public template

    State‑of‑the‑art multimodal foundation model developed by Microsoft Research which seamlessly fuses robust language understanding with advanced visual and audio analysis. <metadata> gpu: A100 | collections: ["HF Transformers"] </metadata>

    inferless/phi-4-multimodal-instruct’s past year of commit activity
    Python 0 4 0 0 Updated Apr 27, 2025
  • stable-diffusion-3.5-large Public

    8B model, excels in producing high-quality, detailed images up to 1 megapixel in resolution. <metadata> gpu: A100 | collections: ["Diffusers"] </metadata>

    inferless/stable-diffusion-3.5-large’s past year of commit activity
    Python 0 0 0 0 Updated Apr 21, 2025
  • phi-4-GGUF Public template

    A 14B model optimized in GGUF format for efficient inference, designed to excel in complex reasoning tasks. <metadata> gpu: A100 | collections: ["llama.cpp","GGUF"] </metadata>

    inferless/phi-4-GGUF’s past year of commit activity
    Python 0 7 0 0 Updated Apr 19, 2025
  • tinyllama-1-1b-chat-v1-0 Public template

    A chat model fine-tuned on TinyLlama, a compact 1.1B Llama model pretrained on 3 trillion tokens. <metadata> gpu: T4 | collections: ["vLLM"] </metadata>

    inferless/tinyllama-1-1b-chat-v1-0’s past year of commit activity
    Python 1 3 0 0 Updated Apr 18, 2025
  • llama-2-13b-chat-hf Public template

    A 13B model fine-tuned with reinforcement learning from human feedback, part of Meta’s Llama 2 family for dialogue tasks. <metadata> gpu: A100 | collections: ["HF Transformers"] </metadata>

    inferless/llama-2-13b-chat-hf’s past year of commit activity
    Python 0 1 0 0 Updated Apr 18, 2025