Config file
The configuration file controls the training pipeline through four main sections: base, tuning, evaluation, and synthgen. Each section handles a specific aspect of the model training process.
File format
The configuration file supports two formats depending on how you interact with distil labs:
- Webapp: Use JSON format (
.jsonfile) - API: Use YAML format (
.yamlfile)
Both formats are functionally equivalent—choose based on your workflow. Examples in this documentation show YAML, but the JSON equivalent is straightforward:
Configuration structure
Base configuration
General parameters relevant to the overall task.
Supported student models
Supported teacher models
Tuning configuration
Parameters controlling the finetuning of the student model.
Evaluation configuration
Parameters used in teacher evaluation.
Synthetic generation configuration
Parameters for fine-grained control over synthetic data generation.
Example configuration
Minimal configuration
Full configuration example
Model-specific notes
DeepSeek R1
When using deepseek.r1 as the teacher model, the recommended temperature range is 0.5 to 0.7. Configurations with temperatures outside this range will raise a validation error.
GPT OSS 120B Thinking
The openai.gpt-oss-120b-thinking model uses a medium reasoning effort setting by default for enhanced chain-of-thought capabilities.
