Preference Datasets for DPO Collection This collection contains a list of curated preference datasets for DPO fine-tuning for intent alignment of LLMs • 7 items • Updated Dec 11, 2024 • 48
argilla/ultrafeedback-binarized-preferences-cleaned Viewer • Updated Dec 11, 2023 • 60.9k • 4.13k • 161
Aratako/Synthetic-Japanese-Roleplay-gpt-4o-mini-39.6k-formatted Viewer • Updated Oct 27, 2024 • 39.6k • 65 • 11