jsaizant commited on
Commit
ba74d96
·
2 Parent(s): c6a5019 6416025

Merging version 1.1

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "/home/jsainz/PycharmProjects/ibm-quantized-models/salamandra-2b-instruct-gptq-grp64",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
@@ -18,13 +18,14 @@
18
  "num_attention_heads": 16,
19
  "num_hidden_layers": 24,
20
  "num_key_value_heads": 16,
21
- "num_layers": 24,
22
  "pretraining_tp": 1,
23
  "quantization_config": {
 
24
  "batch_size": 1,
25
  "bits": 4,
26
  "block_name_to_quantize": null,
27
  "cache_block_outputs": true,
 
28
  "damp_percent": 0.1,
29
  "dataset": null,
30
  "desc_act": false,
@@ -33,6 +34,7 @@
33
  },
34
  "group_size": 64,
35
  "max_input_length": null,
 
36
  "model_seqlen": null,
37
  "module_name_preceding_first_block": null,
38
  "modules_in_block_to_quantize": null,
@@ -49,7 +51,7 @@
49
  "rope_theta": 10000.0,
50
  "tie_word_embeddings": false,
51
  "torch_dtype": "float16",
52
- "transformers_version": "4.46.2",
53
  "use_cache": true,
54
  "vocab_size": 256000
55
  }
 
1
  {
2
+ "_name_or_path": "/gpfs/projects/bsc88/text/models/salamandra_quantized_speculative/V1.1/salamandra-2b-instruct-gptq",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
 
18
  "num_attention_heads": 16,
19
  "num_hidden_layers": 24,
20
  "num_key_value_heads": 16,
 
21
  "pretraining_tp": 1,
22
  "quantization_config": {
23
+ "backend": null,
24
  "batch_size": 1,
25
  "bits": 4,
26
  "block_name_to_quantize": null,
27
  "cache_block_outputs": true,
28
+ "checkpoint_format": "gptq",
29
  "damp_percent": 0.1,
30
  "dataset": null,
31
  "desc_act": false,
 
34
  },
35
  "group_size": 64,
36
  "max_input_length": null,
37
+ "meta": null,
38
  "model_seqlen": null,
39
  "module_name_preceding_first_block": null,
40
  "modules_in_block_to_quantize": null,
 
51
  "rope_theta": 10000.0,
52
  "tie_word_embeddings": false,
53
  "torch_dtype": "float16",
54
+ "transformers_version": "4.49.0",
55
  "use_cache": true,
56
  "vocab_size": 256000
57
  }
generation_config.json CHANGED
@@ -2,5 +2,5 @@
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
- "transformers_version": "4.46.2"
6
  }
 
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
+ "transformers_version": "4.49.0"
6
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:99c95a632c806d16daee0d5fceaed550c1f12befda2ca812a7927c041ad07852
3
  size 2748604656
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b8c2e21d5c72bbd141a08b5e4813d6febcb4ff71e29762b1f7b2651d01f4d56
3
  size 2748604656
tokenizer_config.json CHANGED
@@ -1092,6 +1092,7 @@
1092
  "chat_template": "{%- if not date_string is defined %}{%- set date_string = \"2024-09-30\" %}{%- endif %}{%- set system_message = messages[0].content if messages[0].role == \"system\" else \"I am Salamandra, an AI language model developed at the Barcelona Supercomputing Centre (BSC) by the Language Technologies Unit. My knowledge base was last updated on August 2023. Today Date: \"+ date_string +\"\nSoy Salamandra, un modelo lingüístico de IA desarrollado en el Barcelona Supercomputing Centre (BSC) por la Language Technologies Unit. Mi base de conocimientos se actualizó por última vez en agosto de 2023.\nSoc Salamandra, un model de llenguatge d'IA desenvolupat al Barcelona Supercomputing Centre (BSC) per la Language Technologies Unit. La meva base de coneixement es va actualitzar per última vegada l'agost de 2023.\" -%}{%- if messages[0].role == \"system\" -%}{%- set messages = messages[1:] -%}{%- endif -%}{{ \"<|im_start|>system\n\" + system_message + \"<|im_end|>\n\" }}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
1093
  "clean_up_tokenization_spaces": false,
1094
  "eos_token": "</s>",
 
1095
  "legacy": true,
1096
  "model_max_length": 8192,
1097
  "pad_token": "<unk>",
 
1092
  "chat_template": "{%- if not date_string is defined %}{%- set date_string = \"2024-09-30\" %}{%- endif %}{%- set system_message = messages[0].content if messages[0].role == \"system\" else \"I am Salamandra, an AI language model developed at the Barcelona Supercomputing Centre (BSC) by the Language Technologies Unit. My knowledge base was last updated on August 2023. Today Date: \"+ date_string +\"\nSoy Salamandra, un modelo lingüístico de IA desarrollado en el Barcelona Supercomputing Centre (BSC) por la Language Technologies Unit. Mi base de conocimientos se actualizó por última vez en agosto de 2023.\nSoc Salamandra, un model de llenguatge d'IA desenvolupat al Barcelona Supercomputing Centre (BSC) per la Language Technologies Unit. La meva base de coneixement es va actualitzar per última vegada l'agost de 2023.\" -%}{%- if messages[0].role == \"system\" -%}{%- set messages = messages[1:] -%}{%- endif -%}{{ \"<|im_start|>system\n\" + system_message + \"<|im_end|>\n\" }}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
1093
  "clean_up_tokenization_spaces": false,
1094
  "eos_token": "</s>",
1095
+ "extra_special_tokens": {},
1096
  "legacy": true,
1097
  "model_max_length": 8192,
1098
  "pad_token": "<unk>",