Best Open-source Model Serving & Inference Platforms tools

Explore curated open-source tools in the Model Serving & Inference Platforms category. Compare technologies, see alternatives, and find the right solution for your workflow.

10+ projects · Page 1 of 1

SGLang logo

SGLang

High‑performance serving framework for LLMs and vision‑language models.

Stars
20,742
License
Apache-2.0
Last commit
3 days ago
PythonActive
Triton Inference Server logo

Triton Inference Server

Unified AI model serving across clouds, edge, and GPUs

Stars
10,085
License
BSD-3-Clause
Last commit
3 days ago
PythonActive
BentoML logo

BentoML

Unified Python framework for building high‑performance AI inference APIs

Stars
8,272
License
Apache-2.0
Last commit
3 days ago
PythonActive
KServe logo

KServe

Unified AI inference platform for generative and predictive workloads on Kubernetes

Stars
4,855
License
Apache-2.0
Last commit
4 days ago
ShellActive
NanoFlow logo

NanoFlow

High‑throughput LLM serving with intra‑device parallelism and asynchronous CPU scheduling

Stars
918
License
Unknown
Last commit
1 month ago
Jupyter NotebookActive
FEDML logo

FEDML

Unified ML library for scalable training, serving, and federated learning.

Stars
3,983
License
Apache-2.0
Last commit
1 month ago
PythonActive