geodesic-research/sfm_baseline_unfiltered_dpo
Text Generation
•
7B
•
Updated
•
22
Here is a selection of models that have undergone DPO. We also share the earlier instruction checkpoints. We recommend using the DPO models.