Best Open-source Model Serving & Inference Platforms tools

Explore curated open-source tools in the Model Serving & Inference Platforms category. Compare technologies, see alternatives, and find the right solution for your workflow.

10+ projects · Page 1 of 1

SGLang logo

SGLang

High‑performance serving framework for LLMs and vision‑language models.

Stars
24,203
License
Apache-2.0
Last commit
2 hours ago
PythonActive
Triton Inference Server logo

Triton Inference Server

Unified AI model serving across clouds, edge, and GPUs

Stars
10,408
License
BSD-3-Clause
Last commit
12 hours ago
PythonActive
KServe logo

KServe

Unified AI inference platform for generative and predictive workloads on Kubernetes

Stars
5,172
License
Apache-2.0
Last commit
16 hours ago
GoActive
BentoML logo

BentoML

Unified Python framework for building high‑performance AI inference APIs

Stars
8,492
License
Apache-2.0
Last commit
1 day ago
PythonActive
NanoFlow logo

NanoFlow

High‑throughput LLM serving with intra‑device parallelism and asynchronous CPU scheduling

Stars
948
License
Last commit
4 months ago
Jupyter NotebookStable
FEDML logo

FEDML

Unified ML library for scalable training, serving, and federated learning.

Stars
4,009
License
Apache-2.0
Last commit
4 months ago
PythonStable