You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
|[Supervised Fine-Tuning (SFT)](#supervised-fine-tuning-sft)| Task adaptation | Input-output pairs |Moderate| Fine-tunes pre-trained models on specific tasks by providing labeled conversations. |
13
-
|[Vision-Language SFT](#vision-language-sft)| Multimodal tasks | Image-text pairs |High| Extends SFT to handle both images and text, enabling image understanding problems. |
12
+
|[Supervised Fine-Tuning (SFT)](#supervised-fine-tuning-sft)| Task adaptation | Input-output pairs |Low| Fine-tunes pre-trained models on specific tasks by providing labeled conversations. |
13
+
|[Vision-Language SFT](#vision-language-sft)| Multimodal tasks | Image-text pairs |Moderate| Extends SFT to handle both images and text, enabling image understanding problems. |
14
14
|[Pretraining](#pretraining)| Domain adaptation | Raw text | Very High | Trains a language model from scratch or adapts it to a new domain using large amounts of unlabeled text. |
15
-
|[Direct Preference Optimization (DPO)](#direct-preference-optimization-dpo)| Preference learning | Preference pairs |Moderate| Trains a model to align with human preferences by providing pairs of preferred and rejected outputs. |
15
+
|[Direct Preference Optimization (DPO)](#direct-preference-optimization-dpo)| Preference learning | Preference pairs |Low| Trains a model to align with human preferences by providing pairs of preferred and rejected outputs. |
0 commit comments