Enum FineTuningMethodType
- Namespace
- AiDotNet.Interfaces
- Assembly
- AiDotNet.dll
Specific fine-tuning method types.
public enum FineTuningMethodType
Fields
APO = 15Adversarial/Anchored Preference Optimization.
CPO = 11Contrastive Preference Optimization.
CalDPO = 13Calibrated DPO.
ConstitutionalAI = 24Constitutional AI - principle-based alignment.
DPO = 7Direct Preference Optimization.
GAPO = 28Generative Adversarial Policy Optimization.
GRPO = 3Group Relative Policy Optimization (DeepSeek).
IPO = 8Identity Preference Optimization - addresses DPO overfitting.
KTO = 9Kahneman-Tversky Optimization - uses prospect theory.
KnowledgeDistillation = 27Knowledge Distillation from teacher model.
LCPO = 16Latent Collective Preference Optimization.
NCA = 21Noise Contrastive Alignment.
ORPO = 17Odds Ratio Preference Optimization - combines SFT and alignment.
PPO = 2Proximal Policy Optimization.
PRO = 23Preference Ranking Optimization.
RDPO = 12Robust DPO.
REINFORCE = 6REINFORCE algorithm.
RLAIF = 4Reinforcement Learning from AI Feedback.
RLHF = 1Reinforcement Learning from Human Feedback with PPO.
RLVR = 5Reinforcement Learning with Verifiable Rewards.
RRHF = 19Rank Responses to align Human Feedback.
RSO = 18Rejection Sampling Optimization.
SFT = 0Supervised Fine-Tuning - standard labeled data training.
SLiCHF = 20Sequence Likelihood Calibration with Human Feedback.
SPIN = 26Self-Play Fine-Tuning.
SafeEXO = 25Safe Externalized Optimization.
SafeNCA = 22Safe Noise Contrastive Alignment.
SimPO = 10Simple Preference Optimization - reference-free.
TPO = 14Triple Preference Optimization.
Remarks
This enum provides a complete list of supported fine-tuning methods based on the latest research (2023-2025).