Model save
Browse files- README.md +3 -5
- all_results.json +7 -1
- train_results.json +4 -4
- trainer_state.json +46 -10
README.md
CHANGED
@@ -1,11 +1,9 @@
|
|
1 |
---
|
2 |
base_model: Qwen/Qwen2.5-Coder-7B
|
3 |
-
datasets: Genies/Reasoning_SQL_BIRD
|
4 |
library_name: transformers
|
5 |
model_name: text2sql-reasoning-sql-no-ex
|
6 |
tags:
|
7 |
- generated_from_trainer
|
8 |
-
- open-r1
|
9 |
- trl
|
10 |
- grpo
|
11 |
licence: license
|
@@ -13,7 +11,7 @@ licence: license
|
|
13 |
|
14 |
# Model Card for text2sql-reasoning-sql-no-ex
|
15 |
|
16 |
-
This model is a fine-tuned version of [Qwen/Qwen2.5-Coder-7B](https://huggingface.co/Qwen/Qwen2.5-Coder-7B)
|
17 |
It has been trained using [TRL](https://github.com/huggingface/trl).
|
18 |
|
19 |
## Quick start
|
@@ -29,7 +27,7 @@ print(output["generated_text"])
|
|
29 |
|
30 |
## Training procedure
|
31 |
|
32 |
-
[<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/genies-rnd/text2sql-rl/runs/
|
33 |
|
34 |
|
35 |
This model was trained with GRPO, a method introduced in [DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models](https://huggingface.co/papers/2402.03300).
|
@@ -39,7 +37,7 @@ This model was trained with GRPO, a method introduced in [DeepSeekMath: Pushing
|
|
39 |
- TRL: 0.18.0
|
40 |
- Transformers: 4.52.3
|
41 |
- Pytorch: 2.6.0
|
42 |
-
- Datasets:
|
43 |
- Tokenizers: 0.21.1
|
44 |
|
45 |
## Citations
|
|
|
1 |
---
|
2 |
base_model: Qwen/Qwen2.5-Coder-7B
|
|
|
3 |
library_name: transformers
|
4 |
model_name: text2sql-reasoning-sql-no-ex
|
5 |
tags:
|
6 |
- generated_from_trainer
|
|
|
7 |
- trl
|
8 |
- grpo
|
9 |
licence: license
|
|
|
11 |
|
12 |
# Model Card for text2sql-reasoning-sql-no-ex
|
13 |
|
14 |
+
This model is a fine-tuned version of [Qwen/Qwen2.5-Coder-7B](https://huggingface.co/Qwen/Qwen2.5-Coder-7B).
|
15 |
It has been trained using [TRL](https://github.com/huggingface/trl).
|
16 |
|
17 |
## Quick start
|
|
|
27 |
|
28 |
## Training procedure
|
29 |
|
30 |
+
[<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/genies-rnd/text2sql-rl/runs/v2nly9i2)
|
31 |
|
32 |
|
33 |
This model was trained with GRPO, a method introduced in [DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models](https://huggingface.co/papers/2402.03300).
|
|
|
37 |
- TRL: 0.18.0
|
38 |
- Transformers: 4.52.3
|
39 |
- Pytorch: 2.6.0
|
40 |
+
- Datasets: 4.0.0
|
41 |
- Tokenizers: 0.21.1
|
42 |
|
43 |
## Citations
|
all_results.json
CHANGED
@@ -4,5 +4,11 @@
|
|
4 |
"eval_runtime": 1327.7604,
|
5 |
"eval_samples": 1534,
|
6 |
"eval_samples_per_second": 1.155,
|
7 |
-
"eval_steps_per_second": 0.024
|
|
|
|
|
|
|
|
|
|
|
|
|
8 |
}
|
|
|
4 |
"eval_runtime": 1327.7604,
|
5 |
"eval_samples": 1534,
|
6 |
"eval_samples_per_second": 1.155,
|
7 |
+
"eval_steps_per_second": 0.024,
|
8 |
+
"total_flos": 0.0,
|
9 |
+
"train_loss": -3.7846087062426205e-06,
|
10 |
+
"train_runtime": 110.1977,
|
11 |
+
"train_samples": 9428,
|
12 |
+
"train_samples_per_second": 256.666,
|
13 |
+
"train_steps_per_second": 8.031
|
14 |
}
|
train_results.json
CHANGED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"total_flos": 0.0,
|
3 |
-
"train_loss":
|
4 |
-
"train_runtime":
|
5 |
"train_samples": 9428,
|
6 |
-
"train_samples_per_second":
|
7 |
-
"train_steps_per_second":
|
8 |
}
|
|
|
1 |
{
|
2 |
"total_flos": 0.0,
|
3 |
+
"train_loss": -3.7846087062426205e-06,
|
4 |
+
"train_runtime": 110.1977,
|
5 |
"train_samples": 9428,
|
6 |
+
"train_samples_per_second": 256.666,
|
7 |
+
"train_steps_per_second": 8.031
|
8 |
}
|
trainer_state.json
CHANGED
@@ -2,9 +2,9 @@
|
|
2 |
"best_global_step": null,
|
3 |
"best_metric": null,
|
4 |
"best_model_checkpoint": null,
|
5 |
-
"epoch": 2.
|
6 |
"eval_steps": 100,
|
7 |
-
"global_step":
|
8 |
"is_hyper_param_search": false,
|
9 |
"is_local_process_zero": true,
|
10 |
"is_world_process_zero": true,
|
@@ -31834,18 +31834,54 @@
|
|
31834 |
"step": 884
|
31835 |
},
|
31836 |
{
|
31837 |
-
"
|
31838 |
-
"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
31839 |
"total_flos": 0.0,
|
31840 |
-
"train_loss":
|
31841 |
-
"train_runtime":
|
31842 |
-
"train_samples_per_second":
|
31843 |
-
"train_steps_per_second":
|
31844 |
}
|
31845 |
],
|
31846 |
"logging_steps": 1,
|
31847 |
"max_steps": 885,
|
31848 |
-
"num_input_tokens_seen":
|
31849 |
"num_train_epochs": 3,
|
31850 |
"save_steps": 52,
|
31851 |
"stateful_callbacks": {
|
@@ -31855,7 +31891,7 @@
|
|
31855 |
"should_evaluate": false,
|
31856 |
"should_log": false,
|
31857 |
"should_save": true,
|
31858 |
-
"should_training_stop":
|
31859 |
},
|
31860 |
"attributes": {}
|
31861 |
}
|
|
|
2 |
"best_global_step": null,
|
3 |
"best_metric": null,
|
4 |
"best_model_checkpoint": null,
|
5 |
+
"epoch": 2.003392705682782,
|
6 |
"eval_steps": 100,
|
7 |
+
"global_step": 885,
|
8 |
"is_hyper_param_search": false,
|
9 |
"is_local_process_zero": true,
|
10 |
"is_world_process_zero": true,
|
|
|
31834 |
"step": 884
|
31835 |
},
|
31836 |
{
|
31837 |
+
"clip_ratio/high_max": 0.0,
|
31838 |
+
"clip_ratio/high_mean": 0.0,
|
31839 |
+
"clip_ratio/low_mean": 0.0,
|
31840 |
+
"clip_ratio/low_min": 0.0,
|
31841 |
+
"clip_ratio/region_mean": 0.0,
|
31842 |
+
"completions/clipped_ratio": -1.1875,
|
31843 |
+
"completions/max_length": 256.0,
|
31844 |
+
"completions/max_terminated_length": 254.0,
|
31845 |
+
"completions/mean_length": 222.8072967529297,
|
31846 |
+
"completions/mean_terminated_length": 164.95713806152344,
|
31847 |
+
"completions/min_length": 50.0,
|
31848 |
+
"completions/min_terminated_length": 50.0,
|
31849 |
+
"epoch": 2.003392705682782,
|
31850 |
+
"frac_reward_zero_std": 0.03125,
|
31851 |
+
"grad_norm": 0.7709642648696899,
|
31852 |
+
"kl": 0.05810546875,
|
31853 |
+
"learning_rate": 3.8941533949121694e-12,
|
31854 |
+
"loss": -0.0033,
|
31855 |
+
"num_tokens": 226847464.0,
|
31856 |
+
"reward": 8.306476593017578,
|
31857 |
+
"reward_std": 1.4022259712219238,
|
31858 |
+
"rewards/accuracy_reward/mean": 1.171875,
|
31859 |
+
"rewards/accuracy_reward/std": 1.4674979448318481,
|
31860 |
+
"rewards/format_reward/mean": 0.9479166865348816,
|
31861 |
+
"rewards/format_reward/std": 0.22277598083019257,
|
31862 |
+
"rewards/judge_reward/mean": 1.0458332300186157,
|
31863 |
+
"rewards/judge_reward/std": 0.9318019151687622,
|
31864 |
+
"rewards/ngrams_iou_reward/mean": 0.1845172643661499,
|
31865 |
+
"rewards/ngrams_iou_reward/std": 0.1875908374786377,
|
31866 |
+
"rewards/schema_keywords_iou_reward/mean": 0.7230003476142883,
|
31867 |
+
"rewards/schema_keywords_iou_reward/std": 0.18600532412528992,
|
31868 |
+
"rewards/syntax_reward/mean": 0.84375,
|
31869 |
+
"rewards/syntax_reward/std": 0.36404144763946533,
|
31870 |
+
"step": 885
|
31871 |
+
},
|
31872 |
+
{
|
31873 |
+
"epoch": 2.003392705682782,
|
31874 |
+
"step": 885,
|
31875 |
"total_flos": 0.0,
|
31876 |
+
"train_loss": -3.7846087062426205e-06,
|
31877 |
+
"train_runtime": 110.1977,
|
31878 |
+
"train_samples_per_second": 256.666,
|
31879 |
+
"train_steps_per_second": 8.031
|
31880 |
}
|
31881 |
],
|
31882 |
"logging_steps": 1,
|
31883 |
"max_steps": 885,
|
31884 |
+
"num_input_tokens_seen": 226847464,
|
31885 |
"num_train_epochs": 3,
|
31886 |
"save_steps": 52,
|
31887 |
"stateful_callbacks": {
|
|
|
31891 |
"should_evaluate": false,
|
31892 |
"should_log": false,
|
31893 |
"should_save": true,
|
31894 |
+
"should_training_stop": true
|
31895 |
},
|
31896 |
"attributes": {}
|
31897 |
}
|