
LLaMA-Factory
Zero-code fine-tuning platform for diverse large language models
- Stars
- 68,012
- License
- Apache-2.0
- Last commit
- 1 day ago
Train and fine-tune models with distributed jobs, schedulers and adapters.
Model training and fine-tuning platforms provide the infrastructure to run large-scale machine-learning jobs, often across multiple GPUs or nodes. They support adapters such as LoRA, parameter-efficient fine-tuning (PEFT), and distributed schedulers to reduce compute cost and accelerate iteration. Both open-source projects (e.g., LLaMA-Factory, Unsloth, PEFT) and SaaS offerings (e.g., Amazon SageMaker JumpStart, Anyscale) exist, giving teams options that balance flexibility, community support, and managed services. Choosing a platform depends on factors like hardware availability, workflow complexity, and integration needs.

Zero-code fine-tuning platform for diverse large language models

Accelerate LLM fine‑tuning with up to 2× speed and 70% less VRAM

Efficiently fine-tune large models with minimal parameters

Fine‑tune LLMs fast with flexible, scalable open‑source framework

End-to-end platform for building, training, and deploying foundation models

Rapid, lightweight fine-tuning for Stable Diffusion using LoRA
Zero-code fine-tuning platform for diverse large language models
Axolotl streamlines LLM and multimodal model fine‑tuning, offering LoRA, QLoRA, QAT, DPO, and multi‑GPU/Node support via simple YAML configs and Docker/PyPI deployment.
Ability to orchestrate jobs across multiple GPUs, nodes, or cloud instances, with support for common schedulers and resource managers.
Native integration of LoRA, adapters, and other parameter-efficient fine-tuning methods to reduce training time and memory usage.
Clear onboarding, example pipelines, and API references that help users move from data preparation to model deployment.
Active open-source contributions, plugin architecture, and compatibility with popular frameworks such as Hugging Face Transformers.
Open-source licenses versus SaaS subscription models, including hidden costs like compute, storage, and support.
Most tools in this category support these baseline capabilities.
ML hub with curated foundation models, pretrained algorithms, and solution templates you can deploy and fine-tune in SageMaker
Ray-powered platform for scalable LLM training and inference.
Enterprise AI platform providing LLMs (Command, Aya) plus Embed/Rerank for retrieval
High-performance inference and fine-tuning platform for open and proprietary models.
API-first platform to run, fine-tune, and deploy AI models without managing infrastructure
AI acceleration cloud for fast inference, fine-tuning, and training via a simple API
Anyscale offers serverless endpoints and managed Ray clusters to serve, fine-tune, and evaluate models with autoscaling, GPUs, and enterprise controls.
Frequently replaced when teams want private deployments and lower TCO.
Apply low-rank adapters to large language models to specialize them on domain-specific data while keeping GPU memory requirements low.
Run parallel training jobs across a cluster to explore learning rates, batch sizes, and other hyperparameters efficiently.
Capture checkpoints, metrics, and configuration metadata for reproducibility and downstream deployment.
Combine PyTorch, TensorFlow, or JAX components within a single training workflow using a common orchestration layer.
Leverage cloud-native services to offload infrastructure management while still using adapters and custom datasets.
What is a model training and fine-tuning platform?
It is software that manages the end-to-end workflow for training or adapting machine-learning models, handling data loading, resource allocation, training loops, and checkpoint management.
How does fine-tuning differ from training from scratch?
Fine-tuning starts from a pre-trained model and adjusts only a subset of parameters (often via adapters), whereas training from scratch learns all weights from random initialization.
What are LoRA adapters and why are they useful?
LoRA (Low-Rank Adaptation) adds small trainable matrices to existing layers, enabling efficient fine-tuning with far fewer trainable parameters and lower memory consumption.
When should I choose an open-source platform over a SaaS solution?
Open-source is preferable when you need full control over the stack, have on-prem hardware, or want to customize the workflow. SaaS is better for rapid setup, managed scaling, and reduced operational overhead.
What hardware is required for large-scale fine-tuning?
At minimum, a GPU with 16 GB VRAM for moderate models; larger LLMs typically need multiple GPUs or cloud instances with high-speed interconnects (NVLink, InfiniBand).
How do platforms handle distributed training jobs?
They provide schedulers that split the workload across nodes, synchronize gradients, and manage fault tolerance, often integrating with Kubernetes, Slurm, or cloud-native services.