Skip to content
Change the repository type filter

All

    Repositories list

    • A Flexible Framework for Experiencing Cutting-edge LLM Inference Optimizations
      Python
      Apache License 2.0
      799100Updated Mar 4, 2025Mar 4, 2025
    • gpustack

      Public
      Manage GPU clusters for running LLMs
      Python
      Apache License 2.0
      181000Updated Mar 4, 2025Mar 4, 2025
    • ollama

      Public
      Get up and running with Llama 3, Mistral, Gemma, and other large language models.
      Go
      MIT License
      11k000Updated Feb 27, 2025Feb 27, 2025
    • llama-box

      Public
      LLM inference server implementation based on llama.cpp.
      C++
      MIT License
      15000Updated Feb 16, 2025Feb 16, 2025
    • Stable Diffusion and Flux in pure C/C++
      C++
      MIT License
      349000Updated Feb 14, 2025Feb 14, 2025
    • llama.cpp

      Public
      LLM inference in C/C++
      C++
      MIT License
      11k300Updated Feb 13, 2025Feb 13, 2025
    • exo

      Public
      Run your own AI cluster at home with everyday devices 📱💻 🖥️⌚
      Python
      GNU General Public License v3.0
      1.6k000Updated Nov 28, 2024Nov 28, 2024
    • Python bindings for llama.cpp
      Python
      MIT License
      1.1k000Updated Nov 26, 2024Nov 26, 2024
    • fastfetch

      Public
      Like neofetch, but much faster because written mostly in C.
      C
      MIT License
      478000Updated Nov 19, 2024Nov 19, 2024
    • vllm

      Public
      A high-throughput and memory-efficient inference and serving engine for LLMs
      Python
      Apache License 2.0
      6k000Updated Oct 16, 2024Oct 16, 2024
    • k8sgpt

      Public
      Giving Kubernetes Superpowers to everyone
      Go
      Apache License 2.0
      750000Updated Sep 24, 2024Sep 24, 2024
    • Automatic SRE Superpowers within your Kubernetes cluster
      Go
      Apache License 2.0
      98000Updated Jul 31, 2024Jul 31, 2024
    • llm.c

      Public
      LLM training in simple, raw C/CUDA
      Cuda
      MIT License
      3k000Updated Jul 22, 2024Jul 22, 2024
    • A proxy that allows you to host ollama images in your local environment
      Go
      MIT License
      2000Updated Jul 2, 2024Jul 2, 2024
    • LLM Benchmark for Throughput via Ollama (Local LLMs)
      Python
      MIT License
      28000Updated Jun 11, 2024Jun 11, 2024
    • makllama

      Public
      MaK(Mac+Kubernetes)llama - Running and orchestrating large language models (LLMs) on Kubernetes with macOS nodes.
      Go
      Apache License 2.0
      33600Updated May 22, 2024May 22, 2024
    • An open and reliable container runtime
      Go
      Apache License 2.0
      3.5k100Updated May 22, 2024May 22, 2024
    • .github

      Public
      0000Updated May 21, 2024May 21, 2024
    • cri

      Public
      Go
      17100Updated May 21, 2024May 21, 2024