core.training_args
core.training_args
extra axolotl specific training args
Classes
Name | Description |
---|---|
AxolotlCPOConfig | CPO config for CPO training |
AxolotlKTOConfig | KTO config for KTO training |
AxolotlORPOConfig | ORPO config for ORPO training |
AxolotlPRMConfig | PRM config for PRM training |
AxolotlRewardConfig | Reward config for Reward training |
AxolotlTrainingArguments | Training arguments for Causal trainer |
AxolotlCPOConfig
=None) core.training_args.AxolotlCPOConfig(simpo_gamma
CPO config for CPO training
AxolotlKTOConfig
core.training_args.AxolotlKTOConfig()
KTO config for KTO training
AxolotlORPOConfig
core.training_args.AxolotlORPOConfig()
ORPO config for ORPO training
AxolotlPRMConfig
core.training_args.AxolotlPRMConfig()
PRM config for PRM training
AxolotlRewardConfig
core.training_args.AxolotlRewardConfig()
Reward config for Reward training
AxolotlTrainingArguments
core.training_args.AxolotlTrainingArguments()
Training arguments for Causal trainer
This code is duplicated due to HF TrainingArguments not setting output_dir with a default value so it can’t be used as a mixin.