Best Open-source Model Serving & Inference Platforms tools

Explore curated open-source tools in the Model Serving & Inference Platforms category. Compare technologies, see alternatives, and find the right solution for your workflow.

10+ projects · Page 1 of 1

SGLang logo

SGLang

High‑performance serving framework for LLMs and vision‑language models.

Stars
22,618
License
Apache-2.0
Last commit
10 hours ago
PythonActive
Triton Inference Server logo

Triton Inference Server

Unified AI model serving across clouds, edge, and GPUs

Stars
10,255
License
BSD-3-Clause
Last commit
18 hours ago
PythonActive
KServe logo

KServe

Unified AI inference platform for generative and predictive workloads on Kubernetes

Stars
5,021
License
Apache-2.0
Last commit
21 hours ago
GoActive
BentoML logo

BentoML

Unified Python framework for building high‑performance AI inference APIs

Stars
8,377
License
Apache-2.0
Last commit
9 days ago
PythonActive
NanoFlow logo

NanoFlow

High‑throughput LLM serving with intra‑device parallelism and asynchronous CPU scheduling

Stars
939
License
Last commit
2 months ago
Jupyter NotebookActive
FEDML logo

FEDML

Unified ML library for scalable training, serving, and federated learning.

Stars
3,992
License
Apache-2.0
Last commit
2 months ago
PythonActive