LG-AI-EXAONE commited on
Commit
1694662
·
1 Parent(s): beab9cf

Update HuggingFace transformers support

Browse files
Files changed (3) hide show
  1. README.md +2 -8
  2. config.json +67 -1
  3. generation_config.json +1 -1
README.md CHANGED
@@ -20,6 +20,7 @@ library_name: transformers
20
  <img src="assets/EXAONE_Symbol+BI_3d.png", width="300", style="margin: 40 auto;">
21
  🎉 License Updated! We are pleased to announce our more flexible licensing terms 🤗
22
  <br>✈️ Try on <a href="https://friendli.ai/suite/~/serverless-endpoints/LGAI-EXAONE/EXAONE-4.0-32B/overview">FriendliAI</a>
 
23
  <br>
24
 
25
  # EXAONE-4.0-32B-FP8
@@ -50,14 +51,7 @@ For more details, please refer to our [technical report](https://arxiv.org/abs/2
50
 
51
  ## Quickstart
52
 
53
- You should install the transformers library forked from the original, available in our [PR](https://github.com/huggingface/transformers/pull/39129).
54
- Once this PR is merged and released, we will update this section.
55
-
56
- You can install the latest version of transformers with support for EXAONE 4.0 by following the command:
57
-
58
- ```bash
59
- pip install git+https://github.com/lgai-exaone/transformers@add-exaone4
60
- ```
61
 
62
  ### Non-reasoning mode
63
 
 
20
  <img src="assets/EXAONE_Symbol+BI_3d.png", width="300", style="margin: 40 auto;">
21
  🎉 License Updated! We are pleased to announce our more flexible licensing terms 🤗
22
  <br>✈️ Try on <a href="https://friendli.ai/suite/~/serverless-endpoints/LGAI-EXAONE/EXAONE-4.0-32B/overview">FriendliAI</a>
23
+ <br><br><i>📢 EXAONE 4.0 is officially supported by HuggingFace transformers! Please check out the guide <a href="#quickstart">below</a></i>
24
  <br>
25
 
26
  # EXAONE-4.0-32B-FP8
 
51
 
52
  ## Quickstart
53
 
54
+ You should install the transformers library with version >= `4.54.0`.
 
 
 
 
 
 
 
55
 
56
  ### Non-reasoning mode
57
 
config.json CHANGED
@@ -10,6 +10,72 @@
10
  "hidden_size": 5120,
11
  "initializer_range": 0.02,
12
  "intermediate_size": 27392,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
13
  "max_position_embeddings": 131072,
14
  "model_type": "exaone4",
15
  "num_attention_heads": 40,
@@ -38,7 +104,7 @@
38
  "sliding_window_pattern": "LLLG",
39
  "tie_word_embeddings": false,
40
  "torch_dtype": "bfloat16",
41
- "transformers_version": "4.54.0.dev0",
42
  "use_cache": true,
43
  "vocab_size": 102400
44
  }
 
10
  "hidden_size": 5120,
11
  "initializer_range": 0.02,
12
  "intermediate_size": 27392,
13
+ "layer_types": [
14
+ "sliding_attention",
15
+ "sliding_attention",
16
+ "sliding_attention",
17
+ "full_attention",
18
+ "sliding_attention",
19
+ "sliding_attention",
20
+ "sliding_attention",
21
+ "full_attention",
22
+ "sliding_attention",
23
+ "sliding_attention",
24
+ "sliding_attention",
25
+ "full_attention",
26
+ "sliding_attention",
27
+ "sliding_attention",
28
+ "sliding_attention",
29
+ "full_attention",
30
+ "sliding_attention",
31
+ "sliding_attention",
32
+ "sliding_attention",
33
+ "full_attention",
34
+ "sliding_attention",
35
+ "sliding_attention",
36
+ "sliding_attention",
37
+ "full_attention",
38
+ "sliding_attention",
39
+ "sliding_attention",
40
+ "sliding_attention",
41
+ "full_attention",
42
+ "sliding_attention",
43
+ "sliding_attention",
44
+ "sliding_attention",
45
+ "full_attention",
46
+ "sliding_attention",
47
+ "sliding_attention",
48
+ "sliding_attention",
49
+ "full_attention",
50
+ "sliding_attention",
51
+ "sliding_attention",
52
+ "sliding_attention",
53
+ "full_attention",
54
+ "sliding_attention",
55
+ "sliding_attention",
56
+ "sliding_attention",
57
+ "full_attention",
58
+ "sliding_attention",
59
+ "sliding_attention",
60
+ "sliding_attention",
61
+ "full_attention",
62
+ "sliding_attention",
63
+ "sliding_attention",
64
+ "sliding_attention",
65
+ "full_attention",
66
+ "sliding_attention",
67
+ "sliding_attention",
68
+ "sliding_attention",
69
+ "full_attention",
70
+ "sliding_attention",
71
+ "sliding_attention",
72
+ "sliding_attention",
73
+ "full_attention",
74
+ "sliding_attention",
75
+ "sliding_attention",
76
+ "sliding_attention",
77
+ "full_attention"
78
+ ],
79
  "max_position_embeddings": 131072,
80
  "model_type": "exaone4",
81
  "num_attention_heads": 40,
 
104
  "sliding_window_pattern": "LLLG",
105
  "tie_word_embeddings": false,
106
  "torch_dtype": "bfloat16",
107
+ "transformers_version": "4.54.0",
108
  "use_cache": true,
109
  "vocab_size": 102400
110
  }
generation_config.json CHANGED
@@ -4,5 +4,5 @@
4
  "cache_implementation": "hybrid",
5
  "eos_token_id": 361,
6
  "pad_token_id": 0,
7
- "transformers_version": "4.54.0.dev0"
8
  }
 
4
  "cache_implementation": "hybrid",
5
  "eos_token_id": 361,
6
  "pad_token_id": 0,
7
+ "transformers_version": "4.54.0"
8
  }