Skip to content
Change the repository type filter

All

    Repositories list

    • qwen2.5-vl-7b-instruct

      Public template
      Vision-Language model that integrates advanced image, video, and text understanding. <metadata> gpu: A100 | collections: ["vLLM"] </metadata>
      Python
      0000Updated Feb 28, 2025Feb 28, 2025
    • Uses Stable Diffusion to generate images and automatically uploads them to an S3 bucket. <metadata> gpu: A100 | collections: ["S3 Storage", "Complex Outputs"] </metadata>
      Python
      0001Updated Feb 28, 2025Feb 28, 2025
    • vicuna-7b-1.1

      Public template
      Open-source chatbot fine-tuned from LLaMA on 70K ShareGPT conversations, optimized for research and conversational tasks. <metadata> gpu: A100 | collections: ["HF Transformers"] </metadata>
      Python
      2000Updated Feb 28, 2025Feb 28, 2025
    • A chat model fine-tuned on TinyLlama, a compact 1.1B Llama model pretrained on 3 trillion tokens. <metadata> gpu: T4 | collections: ["vLLM"] </metadata>
      Python
      2100Updated Feb 28, 2025Feb 28, 2025
    • rmbg-1.4

      Public template
      State-of-the-art background removal model, designed to effectively separate foreground from background. <metadata> gpu: T4 | collections: ["HF Transformers"] </metadata>
      Python
      101900Updated Feb 28, 2025Feb 28, 2025
    • A State-Of-The-Art coder LLM, tailored for instruction-based tasks, particularly in code generation, reasoning, and repair. <metadata> gpu: A100 | collections: ["vLLM"] </metadata>
      Python
      2100Updated Feb 28, 2025Feb 28, 2025
    • A state-of-the-art model that segments and labels audio recordings by accurately distinguishing different speakers. <metadata> gpu: T4 | collections: ["HF Transformers"] </metadata>
      Python
      1200Updated Feb 28, 2025Feb 28, 2025
    • playground-v2.5

      Public template
      Generate highly aesthetic 1024x1024 images with superior quality, flexible aspect ratios, and outstanding human preference alignment. <metadata> gpu: T4 | collections: ["Diffusers"] </metadata>
      Python
      3700Updated Feb 28, 2025Feb 28, 2025
    • phi-3.5-moe-instruct

      Public template
      An instruction-tuned variant of Phi-3.5, delivering efficient, context-aware responses across diverse language tasks. <metadata> gpu: A100 | collections: ["HF Transformers"] </metadata>
      Python
      1000Updated Feb 28, 2025Feb 28, 2025
    • openchat-3.5

      Public template
      A fine-tuned chat model with C-RLFT - a strategy inspired by offline reinforcement learning, optimized for natural, context-aware conversations, excelling in instruction following and text generation tasks. <metadata> gpu: A100 | collections: ["vLLM"] </metadata>
      Python
      3000Updated Feb 28, 2025Feb 28, 2025
    • A GPTQ‑quantized 7B language model based on Mistral, fine‑tuned for robust, efficient conversational and text generation tasks. <metadata> gpu: A100 | collections: ["vLLM","GPTQ"] </metadata>
      Python
      2000Updated Feb 28, 2025Feb 28, 2025
    • mixtral-8x7b-v0.1

      Public template
      A GPTQ-quantized variant of the Mixtral 8x7B model, fine-tuned for efficient text generation and conversational applications. <metadata> gpu: A100 | collections: ["vLLM","GPTQ"] </metadata>
      Python
      4200Updated Feb 28, 2025Feb 28, 2025
    • An 7B model with a 32k token context window and optimized attention mechanisms for superior dialogue and reasoning. <metadata> gpu: A100 | collections: ["vLLM"] </metadata>
      Python
      0000Updated Feb 28, 2025Feb 28, 2025
    • 7B model fine-tuned for precise instruction following and robust contextual understanding. <metadata> gpu: A100 | collections: ["vLLM"] </metadata>
      Python
      0000Updated Feb 28, 2025Feb 28, 2025
    • llama-3.2-3b-instruct

      Public template
      3B compact instruction-tuned model generate detailed responses across a range of tasks. <metadata> gpu: A100 | collections: ["vLLM"] </metadata>
      Python
      2000Updated Feb 28, 2025Feb 28, 2025
    • 11B multimodal model integrating vision and text for image reasoning, captioning, and Q&A. <metadata> gpu: A100 | collections: ["HF Transformers"] </metadata>
      Python
      11700Updated Feb 28, 2025Feb 28, 2025
    • llama-3.1-8b-instruct

      Public template
      An 8B multilingual instruction model fine-tuned with RLHF for chat completion, supporting up to 128k tokens. <metadata> gpu: A100 | collections: ["HF Transformers"] </metadata>
      Python
      5000Updated Feb 28, 2025Feb 28, 2025
    • flux.1-schnell

      Public template
      12B text-to-image mode that upscales and generates high-quality images from text prompts using advanced diffusion techniques. <metadata> gpu: A100 | collections: ["Diffusers","Variable Inputs"] </metadata>
      Python
      9200Updated Feb 28, 2025Feb 28, 2025
    • A GPTQ‑quantized version of Eric Hartford’s Dolphin 2.5 Mixtral 8x7B model, fine‑tuned for coding and conversational tasks. <metadata> gpu: A100 | collections: ["vLLM","GPTQ"] </metadata>
      Python
      10500Updated Feb 28, 2025Feb 28, 2025
    • A 6.7B model fine-tuned on 2 billion tokens of instruction data, designed for code generation and completion tasks. <metadata> gpu: A100 | collections: ["vLLM"] </metadata>
      Python
      3100Updated Feb 28, 2025Feb 28, 2025
    • animagine-xl-3.0

      Public template
      High-quality image generation from text prompts, with improved hand anatomy and concept understanding. <metadata> gpu: A10 | collections: ["Diffusers"] </metadata>
      Python
      11300Updated Feb 28, 2025Feb 28, 2025
    • animagine-xl-3.1

      Public template
      Generates high-quality anime images with improved hand anatomy and new aesthetic tags for enhanced image creation. <metadata> gpu: A10 | collections: ["Diffusers"] </metadata>
      Python
      8300Updated Feb 28, 2025Feb 28, 2025
    • An 8B-parameter, instruction-tuned variant of Meta's Llama-3.1 model, optimized in GGUF format for efficient inference. <metadata> gpu: A100 | collections: ["Using NFS Volumes", "lama.cpp"] </metadata>
      Python
      5000Updated Feb 25, 2025Feb 25, 2025
    • phi-4-GGUF

      Public template
      A 14B model optimized in GGUF format for efficient inference, designed to excel in complex reasoning tasks. <metadata> gpu: A100 | collections: ["llama.cpp"] </metadata>
      Python
      2000Updated Feb 24, 2025Feb 24, 2025
    • mistral-7b

      Public template
      A 7B autoregressive language model by Mistral AI, optimized for efficient text generation and robust reasoning. <metadata> gpu: A100 | collections: ["vLLM"] </metadata>
      Python
      11300Updated Feb 24, 2025Feb 24, 2025
    • gemma-2-9b-it

      Public template
      Instruct-tuned model for instruction following, delivering coherent, high-quality responses across a broad spectrum of tasks. <metadata> gpu: A10 | collections: ["HF Transformers"] </metadata>
      Python
      1000Updated Feb 23, 2025Feb 23, 2025
    • A distilled and cost-effective variant of SDXL that delivers high-quality text-to-image generation with accelerated inference speed. <metadata> gpu: T4 | collections: ["Diffusers"] </metadata>
      Python
      10300Updated Feb 21, 2025Feb 21, 2025
    • DeciLM-7B

      Public
      Python
      1000Updated Feb 21, 2025Feb 21, 2025
    • qwq-32b-preview

      Public template
      A 32B experimental reasoning model for advanced text generation and robust instruction following. <metadata> gpu: A100 | collections: ["vLLM"] </metadata>
      Python
      2200Updated Feb 20, 2025Feb 20, 2025
    • whisper-large-v3-turbo

      Public template
      A turbocharged variant of Whisper large‑v3 for English speech recognition, optimized for lower latency. <metadata> gpu: T4 | collections: ["HF Transformers","Complex Outputs"] </metadata>
      Python
      4000Updated Feb 20, 2025Feb 20, 2025