English
Eluwa-2.7b / README.md
Yudhanjaya's picture
Update README.md
f50f250
|
raw
history blame
2.12 kB
metadata
license: cc-by-nc-4.0
datasets:
  - tatsu-lab/alpaca
language:
  - en

Eluwa: A Conversational LoRA for Facebook's OPT 2.7b Architecture

logo

Eluwa is a fine-tuned Low-Rank Adapter (LoRA) model for Facebook's OPT 2.7b. It is trained on the Stanford Alpaca dataset. The idea was that OPT 2.7 was too curt (and frankly, a bit of an asshole) for a model of its size, and that we could finetune it like Alpaca did to Llama.

This repository contains the Eluwa 2.7b 2 epoch model, which represents a significant improvements in question-answering ability compared to the default OPT 2.7b model. Below are the results of Vicuna-style testing: 80 questions in various categories, with the responses rated by GPT-4.

Model OPT 2.7b base Eluwa 2.7b 1000 iter Eluwa 2.7b 2 epoch
Generic 22 44 57
Knowledge 35 60 72
Roleplay 29 38 58
Common sense 20 48 50
Fermi 4 28 23
Counterfactual 5 24 23
Coding 2 7 7
Math 0 3 3
Writing 8 19 19
Total 125 271 312

(A sheet of questions, answers and GPT's reviews are also included in this repo).

Because of its small size, Eluwa can be used as research into conversational models with older and slower hardware. To load it in a UI like oobabooga, download the model's .bin and .json files, put them in a folder inside the /loras folder, and load it with the OPT 2.7b model.