qywu commited on
Commit
f449dd4
·
verified ·
1 Parent(s): 089358b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +30 -20
README.md CHANGED
@@ -27,9 +27,6 @@ model-index:
27
  - type: acc_norm
28
  value: 51.88
29
  name: normalized accuracy
30
- source:
31
- url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=wandb/gemma-2b-zephyr-sft
32
- name: Open LLM Leaderboard
33
  - task:
34
  type: text-generation
35
  name: Text Generation
@@ -43,9 +40,6 @@ model-index:
43
  - type: acc_norm
44
  value: 72.63
45
  name: normalized accuracy
46
- source:
47
- url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=wandb/gemma-2b-zephyr-sft
48
- name: Open LLM Leaderboard
49
  - task:
50
  type: text-generation
51
  name: Text Generation
@@ -60,9 +54,6 @@ model-index:
60
  - type: acc
61
  value: 42.20
62
  name: accuracy
63
- source:
64
- url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=wandb/gemma-2b-zephyr-sft
65
- name: Open LLM Leaderboard
66
  - task:
67
  type: text-generation
68
  name: Text Generation
@@ -76,9 +67,6 @@ model-index:
76
  metrics:
77
  - type: mc2
78
  value: 41.96
79
- source:
80
- url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=wandb/gemma-2b-zephyr-sft
81
- name: Open LLM Leaderboard
82
  - task:
83
  type: text-generation
84
  name: Text Generation
@@ -93,9 +81,6 @@ model-index:
93
  - type: acc
94
  value: 63.85
95
  name: accuracy
96
- source:
97
- url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=wandb/gemma-2b-zephyr-sft
98
- name: Open LLM Leaderboard
99
  - task:
100
  type: text-generation
101
  name: Text Generation
@@ -110,12 +95,37 @@ model-index:
110
  - type: acc
111
  value: 20.09
112
  name: accuracy
113
- source:
114
- url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=wandb/gemma-2b-zephyr-sft
115
- name: Open LLM Leaderboard
116
  ---
117
 
118
- # gemma-2b-zephyr
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
119
 
120
- <!-- Provide a quick summary of what the model is/does. -->
121
 
 
27
  - type: acc_norm
28
  value: 51.88
29
  name: normalized accuracy
 
 
 
30
  - task:
31
  type: text-generation
32
  name: Text Generation
 
40
  - type: acc_norm
41
  value: 72.63
42
  name: normalized accuracy
 
 
 
43
  - task:
44
  type: text-generation
45
  name: Text Generation
 
54
  - type: acc
55
  value: 42.20
56
  name: accuracy
 
 
 
57
  - task:
58
  type: text-generation
59
  name: Text Generation
 
67
  metrics:
68
  - type: mc2
69
  value: 41.96
 
 
 
70
  - task:
71
  type: text-generation
72
  name: Text Generation
 
81
  - type: acc
82
  value: 63.85
83
  name: accuracy
 
 
 
84
  - task:
85
  type: text-generation
86
  name: Text Generation
 
95
  - type: acc
96
  value: 20.09
97
  name: accuracy
 
 
 
98
  ---
99
 
100
+ # Model Card for Gemma 2B Zephyr SFT
101
+
102
+ We trained the [google/gemma-2b](https://huggingface.co/google/gemma-2b) with [deita-10k-v0-sft](https://huggingface.co/datasets/HuggingFaceH4/deita-10k-v0-sft).
103
+ We carefully selected the hyper-parameters and masked the user tokens during training to achieve the best supervised fine-tuning performance.
104
+
105
+ ## Model description
106
+
107
+ - **Model type:** A 2.5B parameter GPT-like model fine-tuned on a mix of publicly available, synthetic datasets.
108
+ - **Language(s) (NLP):** Primarily English
109
+ - **License:** Gemma Terms of Use
110
+ - **Finetuned from model:** [google/gemma-2b](https://huggingface.co/google/gemma-2b)
111
+
112
+
113
+ ## License
114
+ This model has the same license as the [original Gemma model collection](https://ai.google.dev/gemma/terms)
115
+
116
+ ## OpenLLM Leaderboard Performance
117
+
118
+ | Models | Avg. | ARC-C | HellaSwag | MMLU | TruthfulQA | Winogrande | GSM8k |
119
+ |--------------------------------------|-------|-------|-----------|-------|------------|------------|-------|
120
+ | google/gemma-2b | 46.37 | 48.38 | 71.77 | 41.77 | 33.08 | 34.42 | 16.91 |
121
+ | wandb/gemma-2b-zephyr-sft | 47.18 | 49.74 | 72.38 | 41.37 | 34.42 | 66.93 | 18.27 |
122
+ | wandb/gemma-2b-zephyr-dpo | 46.92 | 49.66 | 72.23 | 41.13 | 34.47 | 66.54 | 17.51 |
123
+ | **Columbia-NLP/gemma-2b-zephyr-sft** | 48.75 | 51.8 | 72.63 | 42.20 | 41.96 | 63.85 | 20.09 |
124
+ | Columbia-NLP/gemma-2b-zephyr-dpo | 49.14 | 52.22 | 73.11 | 42.55 | 42.64 | 64.40 | 19.94 |
125
+
126
+ ## MT-Bench
127
+
128
+
129
+
130
 
 
131