metadata
base_model: HuggingFaceTB/SmolLM2-135M
library_name: transformers
model_name: smollm-smtalk-v1
tags:
- generated_from_trainer
- trl
- sft
licence: license
license: mit
datasets:
- HuggingFaceTB/smol-smoltalk
pipeline_tag: text-generation
Model Card for smollm-smtalk-v1
This model is a fine-tuned version of HuggingFaceTB/SmolLM2-135M. It has been trained using TRL.
Quick start
from transformers import pipeline
question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
generator = pipeline("text-generation", model="sfarrukh/smollm-smtalk-v1", device="cuda")
output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
print(output["generated_text"])
Training procedure
This model was trained with SFT.
Framework versions
- TRL: 0.13.0
- Transformers: 4.48.1
- Pytorch: 2.5.1+cu121
- Datasets: 3.2.0
- Tokenizers: 0.21.0