Rustamshry commited on
Commit
494db16
·
verified ·
1 Parent(s): fff50e1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -8
README.md CHANGED
@@ -14,10 +14,6 @@ tags:
14
 
15
  # Model Card for Model ID
16
 
17
-
18
- ## Model Details
19
-
20
-
21
  ### Model Description
22
 
23
  This model is a fine-tuned version of Qwen2.5-1.5B-Instruct on an Azerbaijani medical reasoning dataset.
@@ -61,7 +57,6 @@ Performance in zero-shot generalisation (e.g., English → Azerbaijani medical t
61
  ## How to Get Started with the Model
62
 
63
  ```python
64
-
65
  login(token="")
66
 
67
  tokenizer = AutoTokenizer.from_pretrained("unsloth/Qwen2.5-1.5B-Instruct",)
@@ -87,7 +82,6 @@ outputs = model.generate(
87
  )
88
 
89
  print(tokenizer.decode(outputs[0]))
90
-
91
  ```
92
 
93
  ## Training Details
@@ -97,7 +91,8 @@ print(tokenizer.decode(outputs[0]))
97
  The model was fine-tuned on a translated and cleaned version of FreedomIntelligence/medical-o1-reasoning-SFT, which was manually converted into Azerbaijani.
98
  All examples were filtered for translation quality and medical relevance.
99
 
100
- Dataset(Translated): Rustamshry/medical_o1_reasoning_SFT_az
 
101
  Link of Original Dataset: huggingface.co/datasets/FreedomIntelligence/medical-o1-reasoning-SFT
102
 
103
 
@@ -118,7 +113,7 @@ The model was trained using supervised fine-tuning (SFT) with parameter-efficien
118
  - **Target Modules:** q_proj, k_proj, v_proj, o_proj, gate_proj, up_proj, down_proj
119
 
120
 
121
- #### Speeds, Sizes, Times [optional]
122
 
123
  - **Training speed:** 0.12 steps/sec
124
  - **Total training time:** 11 hours, 26 minutes
 
14
 
15
  # Model Card for Model ID
16
 
 
 
 
 
17
  ### Model Description
18
 
19
  This model is a fine-tuned version of Qwen2.5-1.5B-Instruct on an Azerbaijani medical reasoning dataset.
 
57
  ## How to Get Started with the Model
58
 
59
  ```python
 
60
  login(token="")
61
 
62
  tokenizer = AutoTokenizer.from_pretrained("unsloth/Qwen2.5-1.5B-Instruct",)
 
82
  )
83
 
84
  print(tokenizer.decode(outputs[0]))
 
85
  ```
86
 
87
  ## Training Details
 
91
  The model was fine-tuned on a translated and cleaned version of FreedomIntelligence/medical-o1-reasoning-SFT, which was manually converted into Azerbaijani.
92
  All examples were filtered for translation quality and medical relevance.
93
 
94
+ Dataset (Translated): Rustamshry/medical_o1_reasoning_SFT_az
95
+
96
  Link of Original Dataset: huggingface.co/datasets/FreedomIntelligence/medical-o1-reasoning-SFT
97
 
98
 
 
113
  - **Target Modules:** q_proj, k_proj, v_proj, o_proj, gate_proj, up_proj, down_proj
114
 
115
 
116
+ #### Speeds, Sizes, Times
117
 
118
  - **Training speed:** 0.12 steps/sec
119
  - **Total training time:** 11 hours, 26 minutes