Hugging Face's logo Hugging Face
  • Models
  • Datasets
  • Spaces
  • Docs
  • Enterprise
  • Pricing

  • Log In
  • Sign Up
G's picture
10

G

G-reen
tomar753's profile picture Gargaz's profile picture 21world's profile picture
·

AI & ML interests

SFT, DPO, ORPO, LLMs, text-generation

Organizations

None yet

G-reen 's collections 1

ORPO v DPO v SFT + Training Loss Curves; argilla/dpo-mix-7k
Several trained models to compare the differences between each method. Each model has a complete description of hyperparams with wandb reports.
  • unsloth/mistral-7b-v0.2-bnb-4bit

    Text Generation • 4B • Updated Sep 11, 2024 • 199 • 15
  • argilla/dpo-mix-7k

    Viewer • Updated Jul 16, 2024 • 7.5k • 259 • 168
  • G-reen/EXPERIMENT-DPO-m7b2-1-merged

    Text Generation • 4B • Updated Apr 15, 2024 • 3
  • G-reen/EXPERIMENT-DPO-m7b2-2-merged

    Text Generation • 4B • Updated Apr 15, 2024 • 3
ORPO v DPO v SFT + Training Loss Curves; argilla/dpo-mix-7k
Several trained models to compare the differences between each method. Each model has a complete description of hyperparams with wandb reports.
  • unsloth/mistral-7b-v0.2-bnb-4bit

    Text Generation • 4B • Updated Sep 11, 2024 • 199 • 15
  • argilla/dpo-mix-7k

    Viewer • Updated Jul 16, 2024 • 7.5k • 259 • 168
  • G-reen/EXPERIMENT-DPO-m7b2-1-merged

    Text Generation • 4B • Updated Apr 15, 2024 • 3
  • G-reen/EXPERIMENT-DPO-m7b2-2-merged

    Text Generation • 4B • Updated Apr 15, 2024 • 3
Company
TOS Privacy About Jobs
Website
Models Datasets Spaces Pricing Docs