Update status of vicgalle/CarbonBeagle-11B_eval_request_False_float16_Original to FINISHED
Browse files
vicgalle/CarbonBeagle-11B_eval_request_False_float16_Original.json
CHANGED
@@ -8,10 +8,24 @@
|
|
8 |
"architectures": "MistralForCausalLM",
|
9 |
"weight_type": "Original",
|
10 |
"main_language": "English",
|
11 |
-
"status": "
|
12 |
"submitted_time": "2024-06-17T17:18:43Z",
|
13 |
"model_type": "🤝 : base merges and moerges",
|
14 |
"source": "leaderboard",
|
15 |
"job_id": 842,
|
16 |
-
"job_start_time": "2024-06-21T01-31-39.472689"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
17 |
}
|
|
|
8 |
"architectures": "MistralForCausalLM",
|
9 |
"weight_type": "Original",
|
10 |
"main_language": "English",
|
11 |
+
"status": "FINISHED",
|
12 |
"submitted_time": "2024-06-17T17:18:43Z",
|
13 |
"model_type": "🤝 : base merges and moerges",
|
14 |
"source": "leaderboard",
|
15 |
"job_id": 842,
|
16 |
+
"job_start_time": "2024-06-21T01-31-39.472689",
|
17 |
+
"eval_version": "1.1.0",
|
18 |
+
"result_metrics": {
|
19 |
+
"enem_challenge": 0.6906927921623512,
|
20 |
+
"bluex": 0.5577190542420027,
|
21 |
+
"oab_exams": 0.47425968109339406,
|
22 |
+
"assin2_rte": 0.9210937556104944,
|
23 |
+
"assin2_sts": 0.8146595670108915,
|
24 |
+
"faquad_nli": 0.751597669118611,
|
25 |
+
"hatebr_offensive": 0.8482580645161291,
|
26 |
+
"portuguese_hate_speech": 0.6870704747394097,
|
27 |
+
"tweetsentbr": 0.5220275792806917
|
28 |
+
},
|
29 |
+
"result_metrics_average": 0.6963754041971083,
|
30 |
+
"result_metrics_npm": 0.5473926792450741
|
31 |
}
|