Best Open-source Model Serving & Inference Platforms tools

Explore curated open-source tools in the Model Serving & Inference Platforms category. Compare technologies, see alternatives, and find the right solution for your workflow.

10+ projects · Page 1 of 1

SGLang logo

SGLang

High‑performance serving framework for LLMs and vision‑language models.

Stars
25,430
License
Apache-2.0
Last commit
17 days ago
PythonActive
Triton Inference Server logo

Triton Inference Server

Unified AI model serving across clouds, edge, and GPUs

Stars
10,516
License
BSD-3-Clause
Last commit
18 days ago
PythonActive
KServe logo

KServe

Unified AI inference platform for generative and predictive workloads on Kubernetes

Stars
5,294
License
Apache-2.0
Last commit
19 days ago
GoActive
BentoML logo

BentoML

Unified Python framework for building high‑performance AI inference APIs

Stars
8,554
License
Apache-2.0
Last commit
20 days ago
PythonActive
NanoFlow logo

NanoFlow

High‑throughput LLM serving with intra‑device parallelism and asynchronous CPU scheduling

Stars
952
License
Last commit
24 days ago
Jupyter NotebookActive
FEDML logo

FEDML

Unified ML library for scalable training, serving, and federated learning.

Stars
4,029
License
Apache-2.0
Last commit
5 months ago
PythonStable