Oumi streamlines model fine-tuning and performance iteration by providing multiple training methods and flexible configuration options. This allows you to experiment efficiently while retaining full control over your setup.Documentation Index
Fetch the complete documentation index at: https://docs.oumi.ai/llms.txt
Use this file to discover all available pages before exploring further.
HOW TO RUN TRAINING JOBS
You can either selectSupervised Fine-Tuning to train a model using labeled examples, or On-policy Distillation to train a student model using a teacher model for knowledge distillation.
SUPERVISED FINE-TUNING (SFT)
To start an SFT job, initiate a training run from the Models page.- Click on
Train New Model. - In the Builder, select
Supervised Fine-Tuning. - Select the base model to fine-tune. Oumi offers a broad range of commonly used models.
- Choose your training dataset, and optionally select validation and test datasets. You can use uploaded datasets, synthesized data, or merged datasets.
- Select a training method. Oumi supports full fine-tuning (FFT) and parameter-efficient fine-tuning (PEFT), including LoRA.
- Adjust advanced hyperparameters (e.g.,
maximum steps,learning rate) if needed. - Review your configuration, (optionally) save it as a reusable recipe, and launch the training job.
ON-POLICY DISTILLATION
To start an on-policy distillation job, initiate a training run from the Models page.- Click on
Train New Model. - In the Builder, select
On-Policy Distillation. - Leave
Training MethodonOn-Policy Distillation. - Choose your
Base ModelandTeacher Model. - Select your
Training Dataset. - Configure advanced settings (e.g.,
Training Settings,Distillation Settings,Parameter-Efficient Settings) if needed.
Please see On-Policy Distillation for more information regarding configuration optionss and settings.
CHECKING JOB STATUS
After a training job launches, it will appear on the Activity log page with a status ofRunning.
When training completes, you can access your model from the Model page.