---
language:
- en
tags:
- mistral
- lora
- adapter
- fine-tuned
- politics
- conversational
license: mit
datasets:
- rohanrao/joe-biden-tweets
- christianlillelund/joe-biden-2020-dnc-speech
base_model: mistralai/Mistral-7B-Instruct-v0.2
library_name: peft
---
# 🇺🇸 Biden Mistral Adapter 🇺🇸
> *"Look, folks, this adapter, it's about our common purpose, our shared values. That's no joke."*
This LoRA adapter for [Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) has been fine-tuned to emulate Joe Biden's distinctive speaking style, discourse patterns, and policy positions. The model captures the measured cadence, personal anecdotes, and characteristic expressions associated with the current U.S. President.
## ✨ Model Details
| Feature | Description |
|---------|-------------|
| **Base Model** | [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) |
| **Architecture** | LoRA adapter (Low-Rank Adaptation) |
| **LoRA Rank** | 16 |
| **Language** | English |
| **Training Focus** | Biden's communication style, rhetoric, and response patterns |
| **Merged Adapters** | Combines style and identity LoRA weights from:
- nnat03/biden-mistral-adapter (original adapter)
- ./identity-adapters/biden-identity-adapter |
## 🎯 Intended Use
📚 Education |
🔍 Research |
🎭 Creative |
Political discourse analysis |
Rhetoric pattern studies |
Interactive simulations |
## 📊 Training Data
This model was trained on carefully curated datasets that capture authentic speech patterns:
- 📱 [Biden tweets dataset (2007-2020)](https://www.kaggle.com/datasets/rohanrao/joe-biden-tweets) - Extensive collection capturing everyday communication
- 🎤 [Biden 2020 DNC speech dataset](https://www.kaggle.com/datasets/christianlillelund/joe-biden-2020-dnc-speech) - Formal oratorical patterns
These datasets were processed into a specialized instruction format to optimize learning of distinctive speech patterns.
## ⚙️ Technical Specifications
### Training Configuration
```
🧠 Framework: Hugging Face Transformers + PEFT
📊 Optimization: 4-bit quantization
🔧 LoRA Config: r=16, alpha=64, dropout=0.05
🎛️ Target modules: q_proj, k_proj, v_proj, o_proj, gate_proj, up_proj, down_proj
```
### Training Parameters
```
📦 Batch size: 4
🔄 Gradient accumulation: 4
📈 Learning rate: 2e-4
🔁 Epochs: 3
📉 LR scheduler: cosine
⚡ Optimizer: paged_adamw_8bit
🧮 Precision: BF16
```
## ⚠️ Limitations and Biases
- This model mimics a speaking style but doesn't guarantee factual accuracy
- While emulating Biden's rhetoric, it doesn't represent his actual views
- May reproduce biases present in the training data
- Not suitable for production applications without additional fact-checking
## 💻 Usage
Run this code to start using the adapter with the Mistral-7B-Instruct-v0.2 base model:
```python
from transformers import AutoModelForCausalLM, AutoTokenizer, BitsAndBytesConfig
from peft import PeftModel
import torch
# Load base model with 4-bit quantization
base_model_id = "mistralai/Mistral-7B-Instruct-v0.2"
bnb_config = BitsAndBytesConfig(
load_in_4bit=True,
bnb_4bit_compute_dtype=torch.float16,
bnb_4bit_quant_type="nf4",
bnb_4bit_use_double_quant=True,
)
# Load model and tokenizer
model = AutoModelForCausalLM.from_pretrained(
base_model_id,
quantization_config=bnb_config,
device_map="auto",
torch_dtype=torch.float16
)
tokenizer = AutoTokenizer.from_pretrained(base_model_id)
# Apply the adapter
model = PeftModel.from_pretrained(model, "nnat03/biden-mistral-adapter")
# Generate a response
prompt = "What's your vision for America's future?"
input_text = f"[INST] {prompt} [/INST]"
inputs = tokenizer(input_text, return_tensors="pt").to("cuda")
outputs = model.generate(**inputs, max_length=512, temperature=0.7, do_sample=True)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(response.split("[/INST]")[-1].strip())
```
## 📚 Citation
If you use this model in your research, please cite:
```bibtex
@misc{nnat03-biden-mistral-adapter,
author = {nnat03},
title = {Biden Mistral Adapter},
year = {2023},
publisher = {Hugging Face},
howpublished = {\url{https://huggingface.co/nnat03/biden-mistral-adapter}}
}
```
## 🔍 Ethical Considerations
This model is created for educational and research purposes. It attempts to mimic the speaking style of a public figure but does not represent their actual views or statements. Use responsibly.
---
Framework version: PEFT 0.15.0
Made with ❤️ for NLP research and education