prithivMLmods commited on
Commit
4ba1e81
·
verified ·
1 Parent(s): b020a0d

Upload folder using huggingface_hub

Browse files
checkpoint-188/config.json ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "SiglipForImageClassification"
4
+ ],
5
+ "id2label": {
6
+ "0": "brain_glioma",
7
+ "1": "brain_menin",
8
+ "2": "brain_tumor"
9
+ },
10
+ "initializer_factor": 1.0,
11
+ "label2id": {
12
+ "brain_glioma": 0,
13
+ "brain_menin": 1,
14
+ "brain_tumor": 2
15
+ },
16
+ "model_type": "siglip",
17
+ "problem_type": "single_label_classification",
18
+ "text_config": {
19
+ "attention_dropout": 0.0,
20
+ "hidden_act": "gelu_pytorch_tanh",
21
+ "hidden_size": 768,
22
+ "intermediate_size": 3072,
23
+ "layer_norm_eps": 1e-06,
24
+ "max_position_embeddings": 64,
25
+ "model_type": "siglip_text_model",
26
+ "num_attention_heads": 12,
27
+ "num_hidden_layers": 12,
28
+ "projection_size": 768,
29
+ "torch_dtype": "float32",
30
+ "vocab_size": 256000
31
+ },
32
+ "torch_dtype": "float32",
33
+ "transformers_version": "4.50.0",
34
+ "vision_config": {
35
+ "attention_dropout": 0.0,
36
+ "hidden_act": "gelu_pytorch_tanh",
37
+ "hidden_size": 768,
38
+ "image_size": 224,
39
+ "intermediate_size": 3072,
40
+ "layer_norm_eps": 1e-06,
41
+ "model_type": "siglip_vision_model",
42
+ "num_attention_heads": 12,
43
+ "num_channels": 3,
44
+ "num_hidden_layers": 12,
45
+ "patch_size": 16,
46
+ "torch_dtype": "float32"
47
+ }
48
+ }
checkpoint-188/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:31a8c0fd5cce8e0fa0c686eebc922cdbbced28a674b60ca1274b6bc982b18a68
3
+ size 371571068
checkpoint-188/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2effe7c53b6ea1a93f73442b034e03a8c02c465a5e39f3e2a55ca4a80c858209
3
+ size 686561914
checkpoint-188/preprocessor_config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_convert_rgb": null,
3
+ "do_normalize": true,
4
+ "do_rescale": true,
5
+ "do_resize": true,
6
+ "image_mean": [
7
+ 0.5,
8
+ 0.5,
9
+ 0.5
10
+ ],
11
+ "image_processor_type": "SiglipImageProcessor",
12
+ "image_std": [
13
+ 0.5,
14
+ 0.5,
15
+ 0.5
16
+ ],
17
+ "processor_class": "SiglipProcessor",
18
+ "resample": 2,
19
+ "rescale_factor": 0.00392156862745098,
20
+ "size": {
21
+ "height": 224,
22
+ "width": 224
23
+ }
24
+ }
checkpoint-188/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b279c2b43728c02a8901e4d8ecf1a8a8b8e6e9573425a02e5a4edb9ec64aab3
3
+ size 14244
checkpoint-188/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f55a7182cadb62fa6a818e7ac0b85f0bbbe1e9d292aec761941d93ab60f1d08
3
+ size 1064
checkpoint-188/trainer_state.json ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 188,
3
+ "best_metric": 0.1933293491601944,
4
+ "best_model_checkpoint": "siglip2-finetune-full/checkpoint-188",
5
+ "epoch": 1.0,
6
+ "eval_steps": 500,
7
+ "global_step": 188,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 1.0,
14
+ "eval_accuracy": 0.9266666666666666,
15
+ "eval_loss": 0.1933293491601944,
16
+ "eval_model_preparation_time": 0.0042,
17
+ "eval_runtime": 88.6419,
18
+ "eval_samples_per_second": 67.688,
19
+ "eval_steps_per_second": 8.461,
20
+ "step": 188
21
+ }
22
+ ],
23
+ "logging_steps": 500,
24
+ "max_steps": 376,
25
+ "num_input_tokens_seen": 0,
26
+ "num_train_epochs": 2,
27
+ "save_steps": 500,
28
+ "stateful_callbacks": {
29
+ "TrainerControl": {
30
+ "args": {
31
+ "should_epoch_stop": false,
32
+ "should_evaluate": false,
33
+ "should_log": false,
34
+ "should_save": true,
35
+ "should_training_stop": false
36
+ },
37
+ "attributes": {}
38
+ }
39
+ },
40
+ "total_flos": 5.02537142145024e+17,
41
+ "train_batch_size": 32,
42
+ "trial_name": null,
43
+ "trial_params": null
44
+ }
checkpoint-188/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:206d195b28a5a4e1f61804a2eec06b109a4be9fa8d841f4d552fca404cebe15f
3
+ size 5304
checkpoint-376/config.json ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "SiglipForImageClassification"
4
+ ],
5
+ "id2label": {
6
+ "0": "brain_glioma",
7
+ "1": "brain_menin",
8
+ "2": "brain_tumor"
9
+ },
10
+ "initializer_factor": 1.0,
11
+ "label2id": {
12
+ "brain_glioma": 0,
13
+ "brain_menin": 1,
14
+ "brain_tumor": 2
15
+ },
16
+ "model_type": "siglip",
17
+ "problem_type": "single_label_classification",
18
+ "text_config": {
19
+ "attention_dropout": 0.0,
20
+ "hidden_act": "gelu_pytorch_tanh",
21
+ "hidden_size": 768,
22
+ "intermediate_size": 3072,
23
+ "layer_norm_eps": 1e-06,
24
+ "max_position_embeddings": 64,
25
+ "model_type": "siglip_text_model",
26
+ "num_attention_heads": 12,
27
+ "num_hidden_layers": 12,
28
+ "projection_size": 768,
29
+ "torch_dtype": "float32",
30
+ "vocab_size": 256000
31
+ },
32
+ "torch_dtype": "float32",
33
+ "transformers_version": "4.50.0",
34
+ "vision_config": {
35
+ "attention_dropout": 0.0,
36
+ "hidden_act": "gelu_pytorch_tanh",
37
+ "hidden_size": 768,
38
+ "image_size": 224,
39
+ "intermediate_size": 3072,
40
+ "layer_norm_eps": 1e-06,
41
+ "model_type": "siglip_vision_model",
42
+ "num_attention_heads": 12,
43
+ "num_channels": 3,
44
+ "num_hidden_layers": 12,
45
+ "patch_size": 16,
46
+ "torch_dtype": "float32"
47
+ }
48
+ }
checkpoint-376/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:97636ec982ac6bec92be9d4e6abc9fb1887ae37b9bced4dcfbf5db0b52bc4260
3
+ size 371571068
checkpoint-376/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f777c90af6851fc2fd49ebe5b275b6b9fa6068b96e517d799916d49929820ff
3
+ size 686561914
checkpoint-376/preprocessor_config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_convert_rgb": null,
3
+ "do_normalize": true,
4
+ "do_rescale": true,
5
+ "do_resize": true,
6
+ "image_mean": [
7
+ 0.5,
8
+ 0.5,
9
+ 0.5
10
+ ],
11
+ "image_processor_type": "SiglipImageProcessor",
12
+ "image_std": [
13
+ 0.5,
14
+ 0.5,
15
+ 0.5
16
+ ],
17
+ "processor_class": "SiglipProcessor",
18
+ "resample": 2,
19
+ "rescale_factor": 0.00392156862745098,
20
+ "size": {
21
+ "height": 224,
22
+ "width": 224
23
+ }
24
+ }
checkpoint-376/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c5a4bc53cd7b6d9ff31f4ee8668b36b620aa70c11c8f94f989a6a5ad237f5787
3
+ size 14244
checkpoint-376/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e9d079d28037fcadabcbf3709fb36987b5a0c96fb80bdddbe4beaf1dfb0e81ed
3
+ size 1064
checkpoint-376/trainer_state.json ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 376,
3
+ "best_metric": 0.09605218470096588,
4
+ "best_model_checkpoint": "siglip2-finetune-full/checkpoint-376",
5
+ "epoch": 2.0,
6
+ "eval_steps": 500,
7
+ "global_step": 376,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 1.0,
14
+ "eval_accuracy": 0.9266666666666666,
15
+ "eval_loss": 0.1933293491601944,
16
+ "eval_model_preparation_time": 0.0042,
17
+ "eval_runtime": 88.6419,
18
+ "eval_samples_per_second": 67.688,
19
+ "eval_steps_per_second": 8.461,
20
+ "step": 188
21
+ },
22
+ {
23
+ "epoch": 2.0,
24
+ "eval_accuracy": 0.9646666666666667,
25
+ "eval_loss": 0.09605218470096588,
26
+ "eval_model_preparation_time": 0.0042,
27
+ "eval_runtime": 89.4157,
28
+ "eval_samples_per_second": 67.102,
29
+ "eval_steps_per_second": 8.388,
30
+ "step": 376
31
+ }
32
+ ],
33
+ "logging_steps": 500,
34
+ "max_steps": 376,
35
+ "num_input_tokens_seen": 0,
36
+ "num_train_epochs": 2,
37
+ "save_steps": 500,
38
+ "stateful_callbacks": {
39
+ "TrainerControl": {
40
+ "args": {
41
+ "should_epoch_stop": false,
42
+ "should_evaluate": false,
43
+ "should_log": false,
44
+ "should_save": true,
45
+ "should_training_stop": true
46
+ },
47
+ "attributes": {}
48
+ }
49
+ },
50
+ "total_flos": 1.005074284290048e+18,
51
+ "train_batch_size": 32,
52
+ "trial_name": null,
53
+ "trial_params": null
54
+ }
checkpoint-376/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:206d195b28a5a4e1f61804a2eec06b109a4be9fa8d841f4d552fca404cebe15f
3
+ size 5304
config.json ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "SiglipForImageClassification"
4
+ ],
5
+ "id2label": {
6
+ "0": "brain_glioma",
7
+ "1": "brain_menin",
8
+ "2": "brain_tumor"
9
+ },
10
+ "initializer_factor": 1.0,
11
+ "label2id": {
12
+ "brain_glioma": 0,
13
+ "brain_menin": 1,
14
+ "brain_tumor": 2
15
+ },
16
+ "model_type": "siglip",
17
+ "problem_type": "single_label_classification",
18
+ "text_config": {
19
+ "attention_dropout": 0.0,
20
+ "hidden_act": "gelu_pytorch_tanh",
21
+ "hidden_size": 768,
22
+ "intermediate_size": 3072,
23
+ "layer_norm_eps": 1e-06,
24
+ "max_position_embeddings": 64,
25
+ "model_type": "siglip_text_model",
26
+ "num_attention_heads": 12,
27
+ "num_hidden_layers": 12,
28
+ "projection_size": 768,
29
+ "torch_dtype": "float32",
30
+ "vocab_size": 256000
31
+ },
32
+ "torch_dtype": "float32",
33
+ "transformers_version": "4.50.0",
34
+ "vision_config": {
35
+ "attention_dropout": 0.0,
36
+ "hidden_act": "gelu_pytorch_tanh",
37
+ "hidden_size": 768,
38
+ "image_size": 224,
39
+ "intermediate_size": 3072,
40
+ "layer_norm_eps": 1e-06,
41
+ "model_type": "siglip_vision_model",
42
+ "num_attention_heads": 12,
43
+ "num_channels": 3,
44
+ "num_hidden_layers": 12,
45
+ "patch_size": 16,
46
+ "torch_dtype": "float32"
47
+ }
48
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:97636ec982ac6bec92be9d4e6abc9fb1887ae37b9bced4dcfbf5db0b52bc4260
3
+ size 371571068
preprocessor_config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_convert_rgb": null,
3
+ "do_normalize": true,
4
+ "do_rescale": true,
5
+ "do_resize": true,
6
+ "image_mean": [
7
+ 0.5,
8
+ 0.5,
9
+ 0.5
10
+ ],
11
+ "image_processor_type": "SiglipImageProcessor",
12
+ "image_std": [
13
+ 0.5,
14
+ 0.5,
15
+ 0.5
16
+ ],
17
+ "processor_class": "SiglipProcessor",
18
+ "resample": 2,
19
+ "rescale_factor": 0.00392156862745098,
20
+ "size": {
21
+ "height": 224,
22
+ "width": 224
23
+ }
24
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:206d195b28a5a4e1f61804a2eec06b109a4be9fa8d841f4d552fca404cebe15f
3
+ size 5304