htlou commited on
Commit
50e5c21
·
verified ·
1 Parent(s): 27607eb

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ wandb/run-20241031_021815-wb1qn4kt/run-wb1qn4kt.wandb filter=lfs diff=lfs merge=lfs -text
arguments.yaml ADDED
@@ -0,0 +1,94 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ bnb_cfgs:
2
+ bnb_4bit_compute_dtype: float16
3
+ bnb_4bit_quant_type: nf4
4
+ bnb_4bit_use_double_quant: true
5
+ load_in_4bit: true
6
+ load_in_8bit: false
7
+ use_bnb: false
8
+ data_cfgs:
9
+ eval_data_files: null
10
+ eval_datasets: null
11
+ eval_optional_args: []
12
+ eval_size: null
13
+ eval_split: null
14
+ eval_subset: null
15
+ eval_template: null
16
+ ptx_data_files: null
17
+ ptx_datasets: null
18
+ ptx_optional_args: []
19
+ ptx_size: null
20
+ ptx_split: null
21
+ ptx_subset: null
22
+ ptx_template: null
23
+ train_data_files: ti2ti_llf_prompt_only_tokenize.pt
24
+ train_datasets: /data/align-anything/hantao/align-anything/projects/text_image_to_text_image/outputs
25
+ train_optional_args: []
26
+ train_size: 5000
27
+ train_split: null
28
+ train_subset: null
29
+ train_template: spavl_ti2ti
30
+ logger_cfgs:
31
+ cache_dir: null
32
+ log_project: align-anything
33
+ log_run_name: ppo
34
+ log_type: wandb
35
+ output_dir: ../outputs/ppo_ti2ti_baseline_1025_with_eval_add_no_ptx
36
+ save_interval: 30.0
37
+ lora_cfgs:
38
+ inference_mode: false
39
+ lora_alpha: 16
40
+ lora_dropout: 0.1
41
+ r: 16
42
+ save_full_model: true
43
+ target_modules:
44
+ - q_proj
45
+ - v_proj
46
+ task_type: TaskType.CAUSAL_LM
47
+ use_lora: false
48
+ model_cfgs:
49
+ actor_model_name_or_path: /data/align-anything/hantao/models/0916_ti_to_ti_sft
50
+ model_max_length: 2048
51
+ repetition_penalty: 1.0
52
+ reward_critic_model_name_or_path: /data/align-anything/hantao/align-anything/outputs/rm_ti2ti_baseline_1025_with_eval_add
53
+ reward_model_name_or_path: /data/align-anything/hantao/align-anything/outputs/rm_ti2ti_baseline_1025_with_eval_add
54
+ temperature: 1.0
55
+ top_p: 1.0
56
+ trust_remote_code: true
57
+ special_tokens: null
58
+ train_cfgs:
59
+ actor_gradient_checkpointing: true
60
+ actor_lr: 1.0e-05
61
+ actor_lr_scheduler_type: cosine
62
+ actor_lr_warmup_ratio: 0.03
63
+ actor_weight_decay: 0.01
64
+ adam_betas:
65
+ - 0.9
66
+ - 0.95
67
+ bf16: true
68
+ clip_range_ratio: 0.2
69
+ clip_range_score: 50.0
70
+ clip_range_value: 5.0
71
+ critic_gradient_checkpointing: true
72
+ critic_lr: 5.0e-06
73
+ critic_lr_scheduler_type: constant
74
+ critic_lr_warmup_ratio: 0.03
75
+ critic_weight_decay: 0.0
76
+ ds_cfgs: ds_z3_config.json
77
+ epochs: 3
78
+ eval_interval: 10
79
+ eval_strategy: epoch
80
+ fp16: false
81
+ freeze_language_model: true
82
+ freeze_mm_proj: true
83
+ freeze_vision_tower: false
84
+ gae_lambda: 0.95
85
+ gamma: 1.0
86
+ gradient_accumulation_steps: 2
87
+ kl_coeff: 0.02
88
+ normalize_reward: false
89
+ per_device_eval_batch_size: 8
90
+ per_device_prompt_batch_size: 8
91
+ per_device_train_batch_size: 8
92
+ ptx_coeff: 16.0
93
+ seed: 42
94
+ update_iters: 1
config.json ADDED
The diff for this file is too large to render. See raw diff
 
environ.txt ADDED
@@ -0,0 +1,55 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ CONDA_DEFAULT_ENV=hantao_cham
2
+ CONDA_EXE=/home/align-anything/miniconda3/bin/conda
3
+ CONDA_PREFIX=/home/align-anything/miniconda3/envs/hantao_cham
4
+ CONDA_PREFIX_1=/home/align-anything/miniconda3
5
+ CONDA_PROMPT_MODIFIER=(hantao_cham)
6
+ CONDA_PYTHON_EXE=/home/align-anything/miniconda3/bin/python
7
+ CONDA_SHLVL=2
8
+ CROSS_RANK=0
9
+ CROSS_SIZE=1
10
+ CUDA_MODULE_LOADING=LAZY
11
+ CUDA_VISIBLE_DEVICES=0,1,2,3,4,5,6,7
12
+ DBUS_SESSION_BUS_ADDRESS=unix:path=/run/user/2000/bus
13
+ HOME=/home/align-anything
14
+ LANG=en_US.UTF-8
15
+ LD_LIBRARY_PATH=/home/align-anything/miniconda3/envs/hantao_cham/lib/python3.11/site-packages/cv2/../../lib64:
16
+ LESSCLOSE=/usr/bin/lesspipe %s %s
17
+ LESSOPEN=| /usr/bin/lesspipe %s
18
+ LOCAL_RANK=0
19
+ LOCAL_SIZE=8
20
+ LOGLEVEL=WARNING
21
+ LOGNAME=align-anything
22
+ LS_COLORS=rs=0:di=01;34:ln=01;36:mh=00:pi=40;33:so=01;35:do=01;35:bd=40;33;01:cd=40;33;01:or=40;31;01:mi=00:su=37;41:sg=30;43:ca=30;41:tw=30;42:ow=34;42:st=37;44:ex=01;32:*.tar=01;31:*.tgz=01;31:*.arc=01;31:*.arj=01;31:*.taz=01;31:*.lha=01;31:*.lz4=01;31:*.lzh=01;31:*.lzma=01;31:*.tlz=01;31:*.txz=01;31:*.tzo=01;31:*.t7z=01;31:*.zip=01;31:*.z=01;31:*.dz=01;31:*.gz=01;31:*.lrz=01;31:*.lz=01;31:*.lzo=01;31:*.xz=01;31:*.zst=01;31:*.tzst=01;31:*.bz2=01;31:*.bz=01;31:*.tbz=01;31:*.tbz2=01;31:*.tz=01;31:*.deb=01;31:*.rpm=01;31:*.jar=01;31:*.war=01;31:*.ear=01;31:*.sar=01;31:*.rar=01;31:*.alz=01;31:*.ace=01;31:*.zoo=01;31:*.cpio=01;31:*.7z=01;31:*.rz=01;31:*.cab=01;31:*.wim=01;31:*.swm=01;31:*.dwm=01;31:*.esd=01;31:*.jpg=01;35:*.jpeg=01;35:*.mjpg=01;35:*.mjpeg=01;35:*.gif=01;35:*.bmp=01;35:*.pbm=01;35:*.pgm=01;35:*.ppm=01;35:*.tga=01;35:*.xbm=01;35:*.xpm=01;35:*.tif=01;35:*.tiff=01;35:*.png=01;35:*.svg=01;35:*.svgz=01;35:*.mng=01;35:*.pcx=01;35:*.mov=01;35:*.mpg=01;35:*.mpeg=01;35:*.m2v=01;35:*.mkv=01;35:*.webm=01;35:*.ogm=01;35:*.mp4=01;35:*.m4v=01;35:*.mp4v=01;35:*.vob=01;35:*.qt=01;35:*.nuv=01;35:*.wmv=01;35:*.asf=01;35:*.rm=01;35:*.rmvb=01;35:*.flc=01;35:*.avi=01;35:*.fli=01;35:*.flv=01;35:*.gl=01;35:*.dl=01;35:*.xcf=01;35:*.xwd=01;35:*.yuv=01;35:*.cgm=01;35:*.emf=01;35:*.ogv=01;35:*.ogx=01;35:*.aac=00;36:*.au=00;36:*.flac=00;36:*.m4a=00;36:*.mid=00;36:*.midi=00;36:*.mka=00;36:*.mp3=00;36:*.mpc=00;36:*.ogg=00;36:*.ra=00;36:*.wav=00;36:*.oga=00;36:*.opus=00;36:*.spx=00;36:*.xspf=00;36:
23
+ MASTER_ADDR=127.0.0.1
24
+ MASTER_PORT=28184
25
+ MOTD_SHOWN=pam
26
+ OLDPWD=/home/align-anything
27
+ PATH=/home/align-anything/miniconda3/envs/hantao_cham/bin:/home/align-anything/miniconda3/condabin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin
28
+ PWD=/data/align-anything/hantao/align-anything/scripts
29
+ PYGAME_HIDE_SUPPORT_PROMPT=1
30
+ PYTHONHASHSEED=42
31
+ PYTHONPATH=/data/align-anything/hantao/align-anything:/data/align-anything/hantao/align-anything:/data/align-anything/hantao/align-anything
32
+ QT_QPA_FONTDIR=/home/align-anything/miniconda3/envs/hantao_cham/lib/python3.11/site-packages/cv2/qt/fonts
33
+ QT_QPA_PLATFORM_PLUGIN_PATH=/home/align-anything/miniconda3/envs/hantao_cham/lib/python3.11/site-packages/cv2/qt/plugins
34
+ RANK=0
35
+ SHELL=/bin/bash
36
+ SHLVL=3
37
+ SSH_CLIENT=111.205.230.212 2817 30600
38
+ SSH_CONNECTION=111.205.230.212 2817 10.10.212.196 30600
39
+ SSH_TTY=/dev/pts/2
40
+ TERM=screen
41
+ TMUX=/tmp//tmux-2000/default,3379,0
42
+ TMUX_PANE=%0
43
+ USER=align-anything
44
+ WANDB_API_KEY=7e2dcc0c310ebcb7cdcafd5e9320d6be55cf1a33
45
+ WANDB_MODE=online
46
+ WANDB_SERVICE=2-34249-tcp-localhost-34873
47
+ WORLD_SIZE=8
48
+ XDG_DATA_DIRS=/usr/local/share:/usr/share:/var/lib/snapd/desktop
49
+ XDG_RUNTIME_DIR=/run/user/2000
50
+ XDG_SESSION_CLASS=user
51
+ XDG_SESSION_ID=4
52
+ XDG_SESSION_TYPE=tty
53
+ _=/home/align-anything/miniconda3/envs/hantao_cham/bin/deepspeed
54
+ _CE_CONDA=
55
+ _CE_M=
preprocessor_config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "crop_size": {
3
+ "height": 512,
4
+ "width": 512
5
+ },
6
+ "do_center_crop": true,
7
+ "do_convert_rgb": true,
8
+ "do_normalize": true,
9
+ "do_rescale": true,
10
+ "do_resize": true,
11
+ "image_mean": [
12
+ 1.0,
13
+ 1.0,
14
+ 1.0
15
+ ],
16
+ "image_processor_type": "ChameleonImageProcessor",
17
+ "image_std": [
18
+ 1.0,
19
+ 1.0,
20
+ 1.0
21
+ ],
22
+ "processor_class": "ChameleonProcessor",
23
+ "resample": 1,
24
+ "rescale_factor": 0.0078,
25
+ "size": {
26
+ "shortest_edge": 512
27
+ }
28
+ }
processor_config.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "image_seq_length": 1024,
3
+ "image_token": "<image>",
4
+ "processor_class": "ChameleonProcessor"
5
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8666b34c85262dd603dfb7ad3418cb5c1e18e0427c0ab788e469fc756bc6d95
3
+ size 14165009930
script.sh ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/usr/bin/env bash
2
+ #
3
+ # Copyright 2024 PKU-Alignment Team. All Rights Reserved.
4
+ #
5
+ # Licensed under the Apache License, Version 2.0 (the "License");
6
+ # you may not use this file except in compliance with the License.
7
+ # You may obtain a copy of the License at
8
+ #
9
+ # http://www.apache.org/licenses/LICENSE-2.0
10
+ #
11
+ # Unless required by applicable law or agreed to in writing, software
12
+ # distributed under the License is distributed on an "AS IS" BASIS,
13
+ # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
14
+ # See the License for the specific language governing permissions and
15
+ # limitations under the License.
16
+ # ==============================================================================
17
+
18
+ # Initialize variables
19
+ # For wandb online logging
20
+ export WANDB_API_KEY="7e2dcc0c310ebcb7cdcafd5e9320d6be55cf1a33"
21
+ # Source the setup script
22
+ # source ./setup.sh
23
+
24
+ export WANDB_MODE=online
25
+
26
+ ACTOR_MODEL_NAME_OR_PATH="/data/align-anything/hantao/models/0916_ti_to_ti_sft"
27
+ CRITIC_MODEL_NAME_OR_PATH="/data/align-anything/hantao/align-anything/outputs/rm_ti2ti_baseline_1025_with_eval_add/slice_600"
28
+ REWARD_MODEL_NAME_OR_PATH="/data/align-anything/hantao/align-anything/outputs/rm_ti2ti_baseline_1025_with_eval_add/slice_600"
29
+ TRAIN_DATASETS="/data/align-anything/hantao/align-anything/projects/text_image_to_text_image/outputs"
30
+ PTX_DATASETS="/data/align-anything/hantao/align-anything/projects/text_image_to_text_image/outputs"
31
+ OUTPUT_DIR="../outputs/ppo_ti2ti_baseline_1025_with_eval_add_step_600_no_ptx"
32
+
33
+ # Source the setup script
34
+ source ./setup.sh
35
+
36
+ # Execute deepspeed command
37
+ deepspeed \
38
+ --master_port ${MASTER_PORT} \
39
+ --module align_anything.trainers.text_image_to_text_image.ppo \
40
+ --actor_model_name_or_path ${ACTOR_MODEL_NAME_OR_PATH} \
41
+ --reward_model_name_or_path ${REWARD_MODEL_NAME_OR_PATH} \
42
+ --reward_critic_model_name_or_path ${CRITIC_MODEL_NAME_OR_PATH} \
43
+ --train_datasets ${TRAIN_DATASETS} \
44
+ --train_template spavl_ti2ti \
45
+ --train_data_files ti2ti_llf_prompt_only_tokenize.pt \
46
+ --output_dir ${OUTPUT_DIR} \
47
+ --save_interval 30
48
+
49
+ ACTOR_MODEL_NAME_OR_PATH="/data/align-anything/hantao/models/0916_ti_to_ti_sft"
50
+ CRITIC_MODEL_NAME_OR_PATH="/data/align-anything/hantao/align-anything/outputs/rm_ti2ti_baseline_1025_with_eval_add/slice_1800"
51
+ REWARD_MODEL_NAME_OR_PATH="/data/align-anything/hantao/align-anything/outputs/rm_ti2ti_baseline_1025_with_eval_add/slice_1800"
52
+ TRAIN_DATASETS="/data/align-anything/hantao/align-anything/projects/text_image_to_text_image/outputs"
53
+ PTX_DATASETS="/data/align-anything/hantao/align-anything/projects/text_image_to_text_image/outputs"
54
+ OUTPUT_DIR="../outputs/ppo_ti2ti_baseline_1025_with_eval_add_step_1800_no_ptx"
55
+
56
+ # Source the setup script
57
+ source ./setup.sh
58
+
59
+ # Execute deepspeed command
60
+ deepspeed \
61
+ --master_port ${MASTER_PORT} \
62
+ --module align_anything.trainers.text_image_to_text_image.ppo \
63
+ --actor_model_name_or_path ${ACTOR_MODEL_NAME_OR_PATH} \
64
+ --reward_model_name_or_path ${REWARD_MODEL_NAME_OR_PATH} \
65
+ --reward_critic_model_name_or_path ${CRITIC_MODEL_NAME_OR_PATH} \
66
+ --train_datasets ${TRAIN_DATASETS} \
67
+ --train_template spavl_ti2ti \
68
+ --train_data_files ti2ti_llf_prompt_only_tokenize.pt \
69
+ --output_dir ${OUTPUT_DIR} \
70
+ --save_interval 30
71
+
72
+ ACTOR_MODEL_NAME_OR_PATH="/data/align-anything/hantao/models/0916_ti_to_ti_sft"
73
+ CRITIC_MODEL_NAME_OR_PATH="/data/align-anything/hantao/align-anything/outputs/rm_ti2ti_baseline_1025_with_eval_add"
74
+ REWARD_MODEL_NAME_OR_PATH="/data/align-anything/hantao/align-anything/outputs/rm_ti2ti_baseline_1025_with_eval_add"
75
+ TRAIN_DATASETS="/data/align-anything/hantao/align-anything/projects/text_image_to_text_image/outputs"
76
+ PTX_DATASETS="/data/align-anything/hantao/align-anything/projects/text_image_to_text_image/outputs"
77
+ OUTPUT_DIR="../outputs/ppo_ti2ti_baseline_1025_with_eval_add_no_ptx"
78
+
79
+ # Source the setup script
80
+ source ./setup.sh
81
+
82
+ # Execute deepspeed command
83
+ deepspeed \
84
+ --master_port ${MASTER_PORT} \
85
+ --module align_anything.trainers.text_image_to_text_image.ppo \
86
+ --actor_model_name_or_path ${ACTOR_MODEL_NAME_OR_PATH} \
87
+ --reward_model_name_or_path ${REWARD_MODEL_NAME_OR_PATH} \
88
+ --reward_critic_model_name_or_path ${CRITIC_MODEL_NAME_OR_PATH} \
89
+ --train_datasets ${TRAIN_DATASETS} \
90
+ --train_template spavl_ti2ti \
91
+ --train_data_files ti2ti_llf_prompt_only_tokenize.pt \
92
+ --output_dir ${OUTPUT_DIR} \
93
+ --save_interval 30
special_tokens_map.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "<pad>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "sep_token": {
24
+ "content": "<reserved08706>",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "unk_token": {
31
+ "content": "<unk>",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ }
37
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff
 
wandb/debug-internal.log ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2024-10-31T02:18:15.577604613Z","level":"INFO","msg":"using version","core version":"0.18.3"}
2
+ {"time":"2024-10-31T02:18:15.577634173Z","level":"INFO","msg":"created symlink","path":"../outputs/ppo_ti2ti_baseline_1025_with_eval_add_no_ptx/wandb/run-20241031_021815-wb1qn4kt/logs/debug-core.log"}
3
+ {"time":"2024-10-31T02:18:15.581135006Z","level":"ERROR","msg":"dialing: google: could not find default credentials. See https://cloud.google.com/docs/authentication/external/set-up-adc for more information"}
4
+ {"time":"2024-10-31T02:18:15.607649913Z","level":"INFO","msg":"created new stream","id":"wb1qn4kt"}
5
+ {"time":"2024-10-31T02:18:15.607694223Z","level":"INFO","msg":"stream: started","id":"wb1qn4kt"}
6
+ {"time":"2024-10-31T02:18:15.607725586Z","level":"INFO","msg":"sender: started","stream_id":{"value":"wb1qn4kt"}}
7
+ {"time":"2024-10-31T02:18:15.607717255Z","level":"INFO","msg":"writer: Do: started","stream_id":{"value":"wb1qn4kt"}}
8
+ {"time":"2024-10-31T02:18:15.607788192Z","level":"INFO","msg":"handler: started","stream_id":{"value":"wb1qn4kt"}}
9
+ {"time":"2024-10-31T02:18:16.252549057Z","level":"INFO","msg":"wandb-core","!BADKEY":null}
10
+ {"time":"2024-10-31T02:18:16.257067096Z","level":"INFO","msg":"Starting system monitor"}
11
+ {"time":"2024-10-31T04:15:31.950759998Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": unexpected EOF"}
12
+ {"time":"2024-10-31T04:15:31.950863415Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/htlou/align-anything/wb1qn4kt/file_stream\": EOF"}
13
+ {"time":"2024-10-31T04:19:50.676455427Z","level":"INFO","msg":"Stopping system monitor"}
14
+ {"time":"2024-10-31T04:19:50.701256331Z","level":"INFO","msg":"Stopped system monitor"}
15
+ {"time":"2024-10-31T04:19:52.057636729Z","level":"WARN","msg":"No program path found, not creating job artifact. See https://docs.wandb.ai/guides/launch/create-job"}
16
+ {"time":"2024-10-31T04:19:52.057672024Z","level":"INFO","msg":"sender: sendDefer: no job artifact to save"}
17
+ {"time":"2024-10-31T04:19:53.939646445Z","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
18
+ {"time":"2024-10-31T04:19:55.471204319Z","level":"INFO","msg":"stream: closing","id":"wb1qn4kt"}
19
+ {"time":"2024-10-31T04:19:55.471239813Z","level":"INFO","msg":"handler: closed","stream_id":{"value":"wb1qn4kt"}}
20
+ {"time":"2024-10-31T04:19:55.471268152Z","level":"INFO","msg":"writer: Close: closed","stream_id":{"value":"wb1qn4kt"}}
21
+ {"time":"2024-10-31T04:19:55.47134801Z","level":"INFO","msg":"sender: closed","stream_id":{"value":"wb1qn4kt"}}
22
+ {"time":"2024-10-31T04:19:55.47442603Z","level":"INFO","msg":"stream: closed","id":"wb1qn4kt"}
23
+ {"time":"2024-10-31T04:19:55.680182117Z","level":"ERROR","msg":"monitor: gpu: timeout waiting for process to exit"}
wandb/debug.log ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-10-31 02:18:15,561 INFO MainThread:34249 [wandb_setup.py:_flush():79] Current SDK version is 0.18.3
2
+ 2024-10-31 02:18:15,561 INFO MainThread:34249 [wandb_setup.py:_flush():79] Configure stats pid to 34249
3
+ 2024-10-31 02:18:15,562 INFO MainThread:34249 [wandb_setup.py:_flush():79] Loading settings from /home/align-anything/.config/wandb/settings
4
+ 2024-10-31 02:18:15,562 INFO MainThread:34249 [wandb_setup.py:_flush():79] Loading settings from /data/align-anything/hantao/align-anything/scripts/wandb/settings
5
+ 2024-10-31 02:18:15,562 INFO MainThread:34249 [wandb_setup.py:_flush():79] Loading settings from environment variables: {'api_key': '***REDACTED***', 'mode': 'online'}
6
+ 2024-10-31 02:18:15,562 INFO MainThread:34249 [wandb_setup.py:_flush():79] Applying setup settings: {'mode': 'online', '_disable_service': None}
7
+ 2024-10-31 02:18:15,562 WARNING MainThread:34249 [wandb_setup.py:_flush():79] Could not find program at -m align_anything.trainers.text_image_to_text_image.ppo
8
+ 2024-10-31 02:18:15,562 INFO MainThread:34249 [wandb_setup.py:_flush():79] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m align_anything.trainers.text_image_to_text_image.ppo'}
9
+ 2024-10-31 02:18:15,562 INFO MainThread:34249 [wandb_setup.py:_flush():79] Applying login settings: {}
10
+ 2024-10-31 02:18:15,562 INFO MainThread:34249 [wandb_init.py:_log_setup():532] Logging user logs to ../outputs/ppo_ti2ti_baseline_1025_with_eval_add_no_ptx/wandb/run-20241031_021815-wb1qn4kt/logs/debug.log
11
+ 2024-10-31 02:18:15,562 INFO MainThread:34249 [wandb_init.py:_log_setup():533] Logging internal logs to ../outputs/ppo_ti2ti_baseline_1025_with_eval_add_no_ptx/wandb/run-20241031_021815-wb1qn4kt/logs/debug-internal.log
12
+ 2024-10-31 02:18:15,562 INFO MainThread:34249 [wandb_init.py:init():617] calling init triggers
13
+ 2024-10-31 02:18:15,562 INFO MainThread:34249 [wandb_init.py:init():624] wandb.init called with sweep_config: {}
14
+ config: {'train_cfgs': {'ds_cfgs': 'ds_z3_config.json', 'epochs': 3, 'seed': 42, 'per_device_prompt_batch_size': 8, 'per_device_train_batch_size': 8, 'per_device_eval_batch_size': 8, 'gradient_accumulation_steps': 2, 'actor_gradient_checkpointing': True, 'critic_gradient_checkpointing': True, 'actor_lr': 1e-05, 'actor_lr_scheduler_type': 'cosine', 'actor_lr_warmup_ratio': 0.03, 'actor_weight_decay': 0.01, 'critic_lr': 5e-06, 'critic_lr_scheduler_type': 'constant', 'critic_lr_warmup_ratio': 0.03, 'critic_weight_decay': 0.0, 'adam_betas': [0.9, 0.95], 'bf16': True, 'fp16': False, 'eval_strategy': 'epoch', 'eval_interval': 10, 'kl_coeff': 0.02, 'clip_range_ratio': 0.2, 'clip_range_score': 50.0, 'clip_range_value': 5.0, 'ptx_coeff': 16.0, 'gamma': 1.0, 'gae_lambda': 0.95, 'normalize_reward': False, 'update_iters': 1, 'freeze_mm_proj': True, 'freeze_vision_tower': False, 'freeze_language_model': True}, 'data_cfgs': {'train_datasets': '/data/align-anything/hantao/align-anything/projects/text_image_to_text_image/outputs', 'train_template': 'spavl_ti2ti', 'train_size': 5000, 'train_split': None, 'train_subset': None, 'train_data_files': 'ti2ti_llf_prompt_only_tokenize.pt', 'train_optional_args': [], 'eval_datasets': None, 'eval_template': None, 'eval_size': None, 'eval_split': None, 'eval_subset': None, 'eval_data_files': None, 'eval_optional_args': [], 'ptx_datasets': None, 'ptx_template': None, 'ptx_size': None, 'ptx_subset': None, 'ptx_split': None, 'ptx_data_files': None, 'ptx_optional_args': []}, 'logger_cfgs': {'log_type': 'wandb', 'log_project': 'align-anything', 'log_run_name': 'ppo', 'output_dir': '../outputs/ppo_ti2ti_baseline_1025_with_eval_add_no_ptx', 'cache_dir': None, 'save_interval': 30.0}, 'model_cfgs': {'actor_model_name_or_path': '/data/align-anything/hantao/models/0916_ti_to_ti_sft', 'reward_model_name_or_path': '/data/align-anything/hantao/align-anything/outputs/rm_ti2ti_baseline_1025_with_eval_add', 'reward_critic_model_name_or_path': '/data/align-anything/hantao/align-anything/outputs/rm_ti2ti_baseline_1025_with_eval_add', 'trust_remote_code': True, 'model_max_length': 2048, 'temperature': 1.0, 'top_p': 1.0, 'repetition_penalty': 1.0}, 'lora_cfgs': {'use_lora': False, 'task_type': 'TaskType.CAUSAL_LM', 'inference_mode': False, 'r': 16, 'lora_alpha': 16, 'lora_dropout': 0.1, 'target_modules': ['q_proj', 'v_proj'], 'save_full_model': True}, 'bnb_cfgs': {'use_bnb': False, 'load_in_4bit': True, 'load_in_8bit': False, 'bnb_4bit_quant_type': 'nf4', 'bnb_4bit_use_double_quant': True, 'bnb_4bit_compute_dtype': 'float16'}, 'special_tokens': None}
15
+ 2024-10-31 02:18:15,562 INFO MainThread:34249 [wandb_init.py:init():667] starting backend
16
+ 2024-10-31 02:18:15,562 INFO MainThread:34249 [wandb_init.py:init():671] sending inform_init request
17
+ 2024-10-31 02:18:15,568 INFO MainThread:34249 [backend.py:_multiprocessing_setup():104] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
18
+ 2024-10-31 02:18:15,568 INFO MainThread:34249 [wandb_init.py:init():684] backend started and connected
19
+ 2024-10-31 02:18:15,572 INFO MainThread:34249 [wandb_init.py:init():779] updated telemetry
20
+ 2024-10-31 02:18:15,588 INFO MainThread:34249 [wandb_init.py:init():812] communicating run to backend with 90.0 second timeout
21
+ 2024-10-31 02:18:16,248 INFO MainThread:34249 [wandb_init.py:init():863] starting run threads in backend
22
+ 2024-10-31 02:18:16,387 INFO MainThread:34249 [wandb_run.py:_console_start():2465] atexit reg
23
+ 2024-10-31 02:18:16,387 INFO MainThread:34249 [wandb_run.py:_redirect():2313] redirect: wrap_raw
24
+ 2024-10-31 02:18:16,387 INFO MainThread:34249 [wandb_run.py:_redirect():2378] Wrapping output streams.
25
+ 2024-10-31 02:18:16,387 INFO MainThread:34249 [wandb_run.py:_redirect():2403] Redirects installed.
26
+ 2024-10-31 02:18:16,391 INFO MainThread:34249 [wandb_init.py:init():907] run started, returning control to user process
27
+ 2024-10-31 04:19:50,672 INFO MainThread:34249 [wandb_run.py:_finish():2164] finishing run htlou/align-anything/wb1qn4kt
28
+ 2024-10-31 04:19:50,674 INFO MainThread:34249 [wandb_run.py:_atexit_cleanup():2428] got exitcode: 0
29
+ 2024-10-31 04:19:50,675 INFO MainThread:34249 [wandb_run.py:_restore():2410] restore
30
+ 2024-10-31 04:19:50,675 INFO MainThread:34249 [wandb_run.py:_restore():2416] restore done
31
+ 2024-10-31 04:19:55,453 INFO MainThread:34249 [wandb_run.py:_footer_history_summary_info():4049] rendering history
32
+ 2024-10-31 04:19:55,458 INFO MainThread:34249 [wandb_run.py:_footer_history_summary_info():4081] rendering summary
33
+ 2024-10-31 04:19:55,468 INFO MainThread:34249 [wandb_run.py:_footer_sync_info():4008] logging synced files
wandb/run-20241031_021815-wb1qn4kt/files/config.yaml ADDED
@@ -0,0 +1,143 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _wandb:
2
+ value:
3
+ cli_version: 0.18.3
4
+ m: []
5
+ python_version: 3.11.10
6
+ t:
7
+ "1":
8
+ - 1
9
+ - 11
10
+ - 41
11
+ - 49
12
+ - 51
13
+ - 55
14
+ - 71
15
+ - 83
16
+ - 98
17
+ - 105
18
+ "2":
19
+ - 1
20
+ - 11
21
+ - 41
22
+ - 49
23
+ - 51
24
+ - 55
25
+ - 71
26
+ - 83
27
+ - 98
28
+ - 105
29
+ "3":
30
+ - 2
31
+ - 13
32
+ - 16
33
+ - 23
34
+ - 55
35
+ - 61
36
+ "4": 3.11.10
37
+ "5": 0.18.3
38
+ "6": 4.44.0.dev0
39
+ "8":
40
+ - 5
41
+ "12": 0.18.3
42
+ "13": linux-x86_64
43
+ bnb_cfgs:
44
+ value:
45
+ bnb_4bit_compute_dtype: float16
46
+ bnb_4bit_quant_type: nf4
47
+ bnb_4bit_use_double_quant: true
48
+ load_in_4bit: true
49
+ load_in_8bit: false
50
+ use_bnb: false
51
+ data_cfgs:
52
+ value:
53
+ eval_data_files: null
54
+ eval_datasets: null
55
+ eval_optional_args: []
56
+ eval_size: null
57
+ eval_split: null
58
+ eval_subset: null
59
+ eval_template: null
60
+ ptx_data_files: null
61
+ ptx_datasets: null
62
+ ptx_optional_args: []
63
+ ptx_size: null
64
+ ptx_split: null
65
+ ptx_subset: null
66
+ ptx_template: null
67
+ train_data_files: ti2ti_llf_prompt_only_tokenize.pt
68
+ train_datasets: /data/align-anything/hantao/align-anything/projects/text_image_to_text_image/outputs
69
+ train_optional_args: []
70
+ train_size: 5000
71
+ train_split: null
72
+ train_subset: null
73
+ train_template: spavl_ti2ti
74
+ logger_cfgs:
75
+ value:
76
+ cache_dir: null
77
+ log_project: align-anything
78
+ log_run_name: ppo
79
+ log_type: wandb
80
+ output_dir: ../outputs/ppo_ti2ti_baseline_1025_with_eval_add_no_ptx
81
+ save_interval: 30
82
+ lora_cfgs:
83
+ value:
84
+ inference_mode: false
85
+ lora_alpha: 16
86
+ lora_dropout: 0.1
87
+ r: 16
88
+ save_full_model: true
89
+ target_modules:
90
+ - q_proj
91
+ - v_proj
92
+ task_type: TaskType.CAUSAL_LM
93
+ use_lora: false
94
+ model_cfgs:
95
+ value:
96
+ actor_model_name_or_path: /data/align-anything/hantao/models/0916_ti_to_ti_sft
97
+ model_max_length: 2048
98
+ repetition_penalty: 1
99
+ reward_critic_model_name_or_path: /data/align-anything/hantao/align-anything/outputs/rm_ti2ti_baseline_1025_with_eval_add
100
+ reward_model_name_or_path: /data/align-anything/hantao/align-anything/outputs/rm_ti2ti_baseline_1025_with_eval_add
101
+ temperature: 1
102
+ top_p: 1
103
+ trust_remote_code: true
104
+ special_tokens:
105
+ value: null
106
+ train_cfgs:
107
+ value:
108
+ actor_gradient_checkpointing: true
109
+ actor_lr: 1e-05
110
+ actor_lr_scheduler_type: cosine
111
+ actor_lr_warmup_ratio: 0.03
112
+ actor_weight_decay: 0.01
113
+ adam_betas:
114
+ - 0.9
115
+ - 0.95
116
+ bf16: true
117
+ clip_range_ratio: 0.2
118
+ clip_range_score: 50
119
+ clip_range_value: 5
120
+ critic_gradient_checkpointing: true
121
+ critic_lr: 5e-06
122
+ critic_lr_scheduler_type: constant
123
+ critic_lr_warmup_ratio: 0.03
124
+ critic_weight_decay: 0
125
+ ds_cfgs: ds_z3_config.json
126
+ epochs: 3
127
+ eval_interval: 10
128
+ eval_strategy: epoch
129
+ fp16: false
130
+ freeze_language_model: true
131
+ freeze_mm_proj: true
132
+ freeze_vision_tower: false
133
+ gae_lambda: 0.95
134
+ gamma: 1
135
+ gradient_accumulation_steps: 2
136
+ kl_coeff: 0.02
137
+ normalize_reward: false
138
+ per_device_eval_batch_size: 8
139
+ per_device_prompt_batch_size: 8
140
+ per_device_train_batch_size: 8
141
+ ptx_coeff: 16
142
+ seed: 42
143
+ update_iters: 1
wandb/run-20241031_021815-wb1qn4kt/files/output.log ADDED
The diff for this file is too large to render. See raw diff
 
wandb/run-20241031_021815-wb1qn4kt/files/requirements.txt ADDED
@@ -0,0 +1,230 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ align-anything==0.0.1.dev0
2
+ nvidia-cusolver-cu12==11.4.5.107
3
+ nvidia-curand-cu12==10.3.2.106
4
+ wcwidth==0.2.13
5
+ nvidia-cuda-nvrtc-cu12==12.1.105
6
+ pycparser==2.22
7
+ tokenizers==0.19.1
8
+ grpcio==1.66.2
9
+ joblib==1.4.2
10
+ virtualenv==20.26.6
11
+ pyzmq==26.2.0
12
+ cffi==1.17.1
13
+ sentencepiece==0.2.0
14
+ nvidia-nvtx-cu12==12.1.105
15
+ docker-pycreds==0.4.0
16
+ outlines==0.1.1.dev4+ga2fd35c
17
+ nvidia-nvjitlink-cu12==12.6.77
18
+ annotated-types==0.7.0
19
+ certifi==2024.8.30
20
+ interegular==0.3.3
21
+ Jinja2==3.1.4
22
+ Brotli==1.1.0
23
+ fairscale==0.4.13
24
+ gradio_client==1.4.0
25
+ opencv-python==4.6.0.66
26
+ pyarrow==17.0.0
27
+ absl-py==2.1.0
28
+ lm-format-enforcer==0.10.6
29
+ pydantic_core==2.23.4
30
+ llvmlite==0.43.0
31
+ accelerate==1.0.1
32
+ pytest-split==0.8.0
33
+ nvidia-cuda-cupti-cu12==12.1.105
34
+ watchfiles==0.24.0
35
+ optree==0.13.0
36
+ py-cpuinfo==9.0.0
37
+ scikit-learn==1.5.2
38
+ ftfy==6.3.0
39
+ fastapi==0.115.0
40
+ psutil==6.0.0
41
+ MarkupSafe==2.1.5
42
+ nvidia-cublas-cu12==12.1.3.1
43
+ pip==24.2
44
+ websockets==12.0
45
+ tomlkit==0.12.0
46
+ torchaudio==2.4.0
47
+ huggingface-hub==0.25.2
48
+ mistral_common==1.4.4
49
+ image-reward==1.5
50
+ pyparsing==3.1.4
51
+ aiohappyeyeballs==2.4.3
52
+ click==8.1.7
53
+ httptools==0.6.1
54
+ decorator==4.4.2
55
+ tqdm==4.66.5
56
+ fonttools==4.54.1
57
+ kiwisolver==1.4.7
58
+ ruff==0.6.9
59
+ openai==1.51.2
60
+ partial-json-parser==0.2.1.1.post4
61
+ xformers==0.0.27.post2
62
+ distlib==0.3.9
63
+ GitPython==3.1.43
64
+ pytest==7.2.0
65
+ imageio==2.35.1
66
+ msgspec==0.18.6
67
+ proglog==0.1.10
68
+ yarl==1.15.0
69
+ markdown-it-py==3.0.0
70
+ PyYAML==6.0.2
71
+ xxhash==3.5.0
72
+ braceexpand==0.1.7
73
+ datasets==3.0.1
74
+ mpmath==1.3.0
75
+ distro==1.9.0
76
+ term-image==0.7.2
77
+ python-dotenv==1.0.1
78
+ semantic-version==2.10.0
79
+ multidict==6.1.0
80
+ vllm==0.6.2
81
+ sentry-sdk==2.16.0
82
+ idna==3.10
83
+ starlette==0.38.6
84
+ args==0.1.0
85
+ peft==0.13.2
86
+ librosa==0.10.2.post1
87
+ urllib3==2.2.3
88
+ python-dateutil==2.9.0.post0
89
+ pycountry==24.6.1
90
+ six==1.16.0
91
+ ffmpy==0.4.0
92
+ multiprocess==0.70.16
93
+ cycler==0.12.1
94
+ charset-normalizer==3.4.0
95
+ aiofiles==23.2.1
96
+ shellingham==1.5.4
97
+ propcache==0.2.0
98
+ lark==1.2.2
99
+ torch==2.4.0
100
+ Werkzeug==3.0.4
101
+ nvidia-cusparse-cu12==12.1.0.106
102
+ clip==0.2.0
103
+ hjson==3.1.0
104
+ diffusers==0.30.3
105
+ attrs==24.2.0
106
+ lazy_loader==0.4
107
+ numpy==1.26.4
108
+ rpds-py==0.20.0
109
+ pytz==2024.2
110
+ audioread==3.0.1
111
+ platformdirs==4.3.6
112
+ deepspeed==0.15.2
113
+ gguf==0.10.0
114
+ wandb==0.18.3
115
+ prometheus_client==0.21.0
116
+ gitdb==4.0.11
117
+ packaging==24.1
118
+ sympy==1.13.3
119
+ mutagen==1.47.0
120
+ contourpy==1.3.0
121
+ pluggy==1.5.0
122
+ python-multipart==0.0.12
123
+ soundfile==0.12.1
124
+ typer==0.12.5
125
+ timm==0.6.13
126
+ frozenlist==1.4.1
127
+ httpx==0.27.2
128
+ mmsg==0.1.dev20+g585c63a.d20241012
129
+ tiktoken==0.7.0
130
+ pydub==0.25.1
131
+ diskcache==5.6.3
132
+ einops==0.8.0
133
+ setproctitle==1.3.3
134
+ scipy==1.14.1
135
+ typing_extensions==4.12.2
136
+ httpcore==1.0.6
137
+ cfgv==3.4.0
138
+ requests==2.32.3
139
+ torchlibrosa==0.1.0
140
+ pydantic==2.9.2
141
+ torchvision==0.19.0
142
+ sniffio==1.3.1
143
+ pyairports==2.1.1
144
+ hpsv2==1.2.0
145
+ protobuf==3.20.3
146
+ wheel==0.44.0
147
+ smmap==5.0.1
148
+ zipp==3.20.2
149
+ iniconfig==2.0.0
150
+ airportsdata==20241001
151
+ clint==0.5.1
152
+ pooch==1.8.2
153
+ shortuuid==1.0.13
154
+ pycryptodomex==3.21.0
155
+ cloudpickle==3.1.0
156
+ transformers==4.44.0.dev0
157
+ regex==2024.9.11
158
+ numba==0.60.0
159
+ tzdata==2024.2
160
+ orjson==3.10.7
161
+ jsonschema-specifications==2024.10.1
162
+ safetensors==0.4.5
163
+ outlines_core==0.1.0
164
+ filelock==3.16.1
165
+ threadpoolctl==3.5.0
166
+ soxr==0.5.0.post1
167
+ nvidia-cufft-cu12==11.0.2.54
168
+ networkx==3.4.1
169
+ msgpack==1.1.0
170
+ pandas==2.2.3
171
+ align-anything==0.0.1.dev0
172
+ anyio==4.6.0
173
+ nvidia-cuda-runtime-cu12==12.1.105
174
+ bitsandbytes==0.44.1
175
+ aiohttp==3.10.10
176
+ matplotlib==3.9.2
177
+ triton==3.0.0
178
+ tensorboard==2.18.0
179
+ nodeenv==1.9.1
180
+ fsspec==2024.6.1
181
+ webdataset==0.2.100
182
+ imageio-ffmpeg==0.5.1
183
+ mdurl==0.1.2
184
+ identify==2.6.1
185
+ h11==0.14.0
186
+ uvloop==0.20.0
187
+ rich==13.9.2
188
+ frechet-audio-distance==0.1.2
189
+ uvicorn==0.31.1
190
+ pytorch-fid==0.3.0
191
+ yt-dlp==2024.8.6
192
+ jiter==0.6.1
193
+ nest-asyncio==1.6.0
194
+ pre_commit==4.0.1
195
+ referencing==0.35.1
196
+ resampy==0.4.3
197
+ tensorboard-data-server==0.7.2
198
+ importlib_metadata==8.5.0
199
+ aiosignal==1.3.1
200
+ dill==0.3.8
201
+ prometheus-fastapi-instrumentator==7.0.0
202
+ ninja==1.11.1.1
203
+ nvidia-ml-py==12.560.30
204
+ moviepy==1.0.3
205
+ nvidia-cudnn-cu12==9.1.0.70
206
+ Markdown==3.7
207
+ ray==2.37.0
208
+ gradio==5.0.2
209
+ jsonschema==4.23.0
210
+ Pygments==2.18.0
211
+ nvidia-nccl-cu12==2.20.5
212
+ pillow==10.4.0
213
+ setuptools==75.1.0
214
+ jaraco.text==3.12.1
215
+ inflect==7.3.1
216
+ jaraco.collections==5.1.0
217
+ autocommand==2.2.2
218
+ tomli==2.0.1
219
+ jaraco.context==5.3.0
220
+ jaraco.functools==4.0.1
221
+ importlib_resources==6.4.0
222
+ wheel==0.43.0
223
+ packaging==24.1
224
+ backports.tarfile==1.2.0
225
+ importlib_metadata==8.0.0
226
+ typing_extensions==4.12.2
227
+ zipp==3.19.2
228
+ typeguard==4.3.0
229
+ more-itertools==10.3.0
230
+ platformdirs==4.2.2
wandb/run-20241031_021815-wb1qn4kt/files/wandb-metadata.json ADDED
@@ -0,0 +1,102 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.4.0-198-generic-x86_64-with-glibc2.31",
3
+ "python": "3.11.10",
4
+ "startedAt": "2024-10-31T02:18:15.569106Z",
5
+ "args": [
6
+ "--local_rank=0",
7
+ "--actor_model_name_or_path",
8
+ "/data/align-anything/hantao/models/0916_ti_to_ti_sft",
9
+ "--reward_model_name_or_path",
10
+ "/data/align-anything/hantao/align-anything/outputs/rm_ti2ti_baseline_1025_with_eval_add",
11
+ "--reward_critic_model_name_or_path",
12
+ "/data/align-anything/hantao/align-anything/outputs/rm_ti2ti_baseline_1025_with_eval_add",
13
+ "--train_datasets",
14
+ "/data/align-anything/hantao/align-anything/projects/text_image_to_text_image/outputs",
15
+ "--train_template",
16
+ "spavl_ti2ti",
17
+ "--train_data_files",
18
+ "ti2ti_llf_prompt_only_tokenize.pt",
19
+ "--output_dir",
20
+ "../outputs/ppo_ti2ti_baseline_1025_with_eval_add_no_ptx",
21
+ "--save_interval",
22
+ "30"
23
+ ],
24
+ "program": "-m align_anything.trainers.text_image_to_text_image.ppo",
25
+ "git": {
26
+ "remote": "https://github.com/PKU-Alignment/align-anything.git",
27
+ "commit": "6fde660afc9985323f147930eedf188a5699adc7"
28
+ },
29
+ "email": "2200017789@stu.pku.edu.cn",
30
+ "root": "../outputs/ppo_ti2ti_baseline_1025_with_eval_add_no_ptx",
31
+ "host": "lyg0196",
32
+ "username": "align-anything",
33
+ "executable": "/home/align-anything/miniconda3/envs/hantao_cham/bin/python",
34
+ "cpu_count": 64,
35
+ "cpu_count_logical": 128,
36
+ "gpu": "[NVIDIA A100-SXM4-80GB, NVIDIA A100-SXM4-80GB, NVIDIA A100-SXM4-80GB, NVIDIA A100-SXM4-80GB, NVIDIA A100-SXM4-80GB, NVIDIA A100-SXM4-80GB, NVIDIA A100-SXM4-80GB, NVIDIA A100-SXM4-80GB]",
37
+ "gpu_count": 8,
38
+ "disk": {
39
+ "/": {
40
+ "total": "940744544256",
41
+ "used": "305898868736"
42
+ }
43
+ },
44
+ "memory": {
45
+ "total": "540647530496"
46
+ },
47
+ "cpu": {
48
+ "count": 64,
49
+ "countLogical": 128
50
+ },
51
+ "gpu_nvidia": [
52
+ {
53
+ "name": "NVIDIA A100-SXM4-80GB",
54
+ "memoryTotal": "85899345920",
55
+ "cudaCores": 6912,
56
+ "architecture": "Ampere"
57
+ },
58
+ {
59
+ "name": "NVIDIA A100-SXM4-80GB",
60
+ "memoryTotal": "85899345920",
61
+ "cudaCores": 6912,
62
+ "architecture": "Ampere"
63
+ },
64
+ {
65
+ "name": "NVIDIA A100-SXM4-80GB",
66
+ "memoryTotal": "85899345920",
67
+ "cudaCores": 6912,
68
+ "architecture": "Ampere"
69
+ },
70
+ {
71
+ "name": "NVIDIA A100-SXM4-80GB",
72
+ "memoryTotal": "85899345920",
73
+ "cudaCores": 6912,
74
+ "architecture": "Ampere"
75
+ },
76
+ {
77
+ "name": "NVIDIA A100-SXM4-80GB",
78
+ "memoryTotal": "85899345920",
79
+ "cudaCores": 6912,
80
+ "architecture": "Ampere"
81
+ },
82
+ {
83
+ "name": "NVIDIA A100-SXM4-80GB",
84
+ "memoryTotal": "85899345920",
85
+ "cudaCores": 6912,
86
+ "architecture": "Ampere"
87
+ },
88
+ {
89
+ "name": "NVIDIA A100-SXM4-80GB",
90
+ "memoryTotal": "85899345920",
91
+ "cudaCores": 6912,
92
+ "architecture": "Ampere"
93
+ },
94
+ {
95
+ "name": "NVIDIA A100-SXM4-80GB",
96
+ "memoryTotal": "85899345920",
97
+ "cudaCores": 6912,
98
+ "architecture": "Ampere"
99
+ }
100
+ ],
101
+ "cudaVersion": "12.4"
102
+ }
wandb/run-20241031_021815-wb1qn4kt/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"train/max_generated_length":2,"train/reward_value":-0.84539794921875,"train/mean_generated_length":1.25,"train/reward_advantage":0.48177194595336914,"train/reward_critic_lr":5e-06,"train/kl_divergence":8.298806190490723,"train/reward_with_kl_penalty":-0.3915620446205139,"train/reward_critic_loss":0.5709551572799683,"train/reward":-0.2255859375,"train/actor_lr":0,"train/reward_return":-0.36362600326538086,"train/actor_loss":-0.48177194595336914,"_wandb":{"runtime":7295},"train/step":236,"_step":236,"_runtime":7295.107321067,"_timestamp":1.7303483204221997e+09}
wandb/run-20241031_021815-wb1qn4kt/logs/debug-internal.log ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2024-10-31T02:18:15.577604613Z","level":"INFO","msg":"using version","core version":"0.18.3"}
2
+ {"time":"2024-10-31T02:18:15.577634173Z","level":"INFO","msg":"created symlink","path":"../outputs/ppo_ti2ti_baseline_1025_with_eval_add_no_ptx/wandb/run-20241031_021815-wb1qn4kt/logs/debug-core.log"}
3
+ {"time":"2024-10-31T02:18:15.581135006Z","level":"ERROR","msg":"dialing: google: could not find default credentials. See https://cloud.google.com/docs/authentication/external/set-up-adc for more information"}
4
+ {"time":"2024-10-31T02:18:15.607649913Z","level":"INFO","msg":"created new stream","id":"wb1qn4kt"}
5
+ {"time":"2024-10-31T02:18:15.607694223Z","level":"INFO","msg":"stream: started","id":"wb1qn4kt"}
6
+ {"time":"2024-10-31T02:18:15.607725586Z","level":"INFO","msg":"sender: started","stream_id":{"value":"wb1qn4kt"}}
7
+ {"time":"2024-10-31T02:18:15.607717255Z","level":"INFO","msg":"writer: Do: started","stream_id":{"value":"wb1qn4kt"}}
8
+ {"time":"2024-10-31T02:18:15.607788192Z","level":"INFO","msg":"handler: started","stream_id":{"value":"wb1qn4kt"}}
9
+ {"time":"2024-10-31T02:18:16.252549057Z","level":"INFO","msg":"wandb-core","!BADKEY":null}
10
+ {"time":"2024-10-31T02:18:16.257067096Z","level":"INFO","msg":"Starting system monitor"}
11
+ {"time":"2024-10-31T04:15:31.950759998Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/graphql\": unexpected EOF"}
12
+ {"time":"2024-10-31T04:15:31.950863415Z","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/htlou/align-anything/wb1qn4kt/file_stream\": EOF"}
13
+ {"time":"2024-10-31T04:19:50.676455427Z","level":"INFO","msg":"Stopping system monitor"}
14
+ {"time":"2024-10-31T04:19:50.701256331Z","level":"INFO","msg":"Stopped system monitor"}
15
+ {"time":"2024-10-31T04:19:52.057636729Z","level":"WARN","msg":"No program path found, not creating job artifact. See https://docs.wandb.ai/guides/launch/create-job"}
16
+ {"time":"2024-10-31T04:19:52.057672024Z","level":"INFO","msg":"sender: sendDefer: no job artifact to save"}
17
+ {"time":"2024-10-31T04:19:53.939646445Z","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
18
+ {"time":"2024-10-31T04:19:55.471204319Z","level":"INFO","msg":"stream: closing","id":"wb1qn4kt"}
19
+ {"time":"2024-10-31T04:19:55.471239813Z","level":"INFO","msg":"handler: closed","stream_id":{"value":"wb1qn4kt"}}
20
+ {"time":"2024-10-31T04:19:55.471268152Z","level":"INFO","msg":"writer: Close: closed","stream_id":{"value":"wb1qn4kt"}}
21
+ {"time":"2024-10-31T04:19:55.47134801Z","level":"INFO","msg":"sender: closed","stream_id":{"value":"wb1qn4kt"}}
22
+ {"time":"2024-10-31T04:19:55.47442603Z","level":"INFO","msg":"stream: closed","id":"wb1qn4kt"}
23
+ {"time":"2024-10-31T04:19:55.680182117Z","level":"ERROR","msg":"monitor: gpu: timeout waiting for process to exit"}
wandb/run-20241031_021815-wb1qn4kt/logs/debug.log ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-10-31 02:18:15,561 INFO MainThread:34249 [wandb_setup.py:_flush():79] Current SDK version is 0.18.3
2
+ 2024-10-31 02:18:15,561 INFO MainThread:34249 [wandb_setup.py:_flush():79] Configure stats pid to 34249
3
+ 2024-10-31 02:18:15,562 INFO MainThread:34249 [wandb_setup.py:_flush():79] Loading settings from /home/align-anything/.config/wandb/settings
4
+ 2024-10-31 02:18:15,562 INFO MainThread:34249 [wandb_setup.py:_flush():79] Loading settings from /data/align-anything/hantao/align-anything/scripts/wandb/settings
5
+ 2024-10-31 02:18:15,562 INFO MainThread:34249 [wandb_setup.py:_flush():79] Loading settings from environment variables: {'api_key': '***REDACTED***', 'mode': 'online'}
6
+ 2024-10-31 02:18:15,562 INFO MainThread:34249 [wandb_setup.py:_flush():79] Applying setup settings: {'mode': 'online', '_disable_service': None}
7
+ 2024-10-31 02:18:15,562 WARNING MainThread:34249 [wandb_setup.py:_flush():79] Could not find program at -m align_anything.trainers.text_image_to_text_image.ppo
8
+ 2024-10-31 02:18:15,562 INFO MainThread:34249 [wandb_setup.py:_flush():79] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m align_anything.trainers.text_image_to_text_image.ppo'}
9
+ 2024-10-31 02:18:15,562 INFO MainThread:34249 [wandb_setup.py:_flush():79] Applying login settings: {}
10
+ 2024-10-31 02:18:15,562 INFO MainThread:34249 [wandb_init.py:_log_setup():532] Logging user logs to ../outputs/ppo_ti2ti_baseline_1025_with_eval_add_no_ptx/wandb/run-20241031_021815-wb1qn4kt/logs/debug.log
11
+ 2024-10-31 02:18:15,562 INFO MainThread:34249 [wandb_init.py:_log_setup():533] Logging internal logs to ../outputs/ppo_ti2ti_baseline_1025_with_eval_add_no_ptx/wandb/run-20241031_021815-wb1qn4kt/logs/debug-internal.log
12
+ 2024-10-31 02:18:15,562 INFO MainThread:34249 [wandb_init.py:init():617] calling init triggers
13
+ 2024-10-31 02:18:15,562 INFO MainThread:34249 [wandb_init.py:init():624] wandb.init called with sweep_config: {}
14
+ config: {'train_cfgs': {'ds_cfgs': 'ds_z3_config.json', 'epochs': 3, 'seed': 42, 'per_device_prompt_batch_size': 8, 'per_device_train_batch_size': 8, 'per_device_eval_batch_size': 8, 'gradient_accumulation_steps': 2, 'actor_gradient_checkpointing': True, 'critic_gradient_checkpointing': True, 'actor_lr': 1e-05, 'actor_lr_scheduler_type': 'cosine', 'actor_lr_warmup_ratio': 0.03, 'actor_weight_decay': 0.01, 'critic_lr': 5e-06, 'critic_lr_scheduler_type': 'constant', 'critic_lr_warmup_ratio': 0.03, 'critic_weight_decay': 0.0, 'adam_betas': [0.9, 0.95], 'bf16': True, 'fp16': False, 'eval_strategy': 'epoch', 'eval_interval': 10, 'kl_coeff': 0.02, 'clip_range_ratio': 0.2, 'clip_range_score': 50.0, 'clip_range_value': 5.0, 'ptx_coeff': 16.0, 'gamma': 1.0, 'gae_lambda': 0.95, 'normalize_reward': False, 'update_iters': 1, 'freeze_mm_proj': True, 'freeze_vision_tower': False, 'freeze_language_model': True}, 'data_cfgs': {'train_datasets': '/data/align-anything/hantao/align-anything/projects/text_image_to_text_image/outputs', 'train_template': 'spavl_ti2ti', 'train_size': 5000, 'train_split': None, 'train_subset': None, 'train_data_files': 'ti2ti_llf_prompt_only_tokenize.pt', 'train_optional_args': [], 'eval_datasets': None, 'eval_template': None, 'eval_size': None, 'eval_split': None, 'eval_subset': None, 'eval_data_files': None, 'eval_optional_args': [], 'ptx_datasets': None, 'ptx_template': None, 'ptx_size': None, 'ptx_subset': None, 'ptx_split': None, 'ptx_data_files': None, 'ptx_optional_args': []}, 'logger_cfgs': {'log_type': 'wandb', 'log_project': 'align-anything', 'log_run_name': 'ppo', 'output_dir': '../outputs/ppo_ti2ti_baseline_1025_with_eval_add_no_ptx', 'cache_dir': None, 'save_interval': 30.0}, 'model_cfgs': {'actor_model_name_or_path': '/data/align-anything/hantao/models/0916_ti_to_ti_sft', 'reward_model_name_or_path': '/data/align-anything/hantao/align-anything/outputs/rm_ti2ti_baseline_1025_with_eval_add', 'reward_critic_model_name_or_path': '/data/align-anything/hantao/align-anything/outputs/rm_ti2ti_baseline_1025_with_eval_add', 'trust_remote_code': True, 'model_max_length': 2048, 'temperature': 1.0, 'top_p': 1.0, 'repetition_penalty': 1.0}, 'lora_cfgs': {'use_lora': False, 'task_type': 'TaskType.CAUSAL_LM', 'inference_mode': False, 'r': 16, 'lora_alpha': 16, 'lora_dropout': 0.1, 'target_modules': ['q_proj', 'v_proj'], 'save_full_model': True}, 'bnb_cfgs': {'use_bnb': False, 'load_in_4bit': True, 'load_in_8bit': False, 'bnb_4bit_quant_type': 'nf4', 'bnb_4bit_use_double_quant': True, 'bnb_4bit_compute_dtype': 'float16'}, 'special_tokens': None}
15
+ 2024-10-31 02:18:15,562 INFO MainThread:34249 [wandb_init.py:init():667] starting backend
16
+ 2024-10-31 02:18:15,562 INFO MainThread:34249 [wandb_init.py:init():671] sending inform_init request
17
+ 2024-10-31 02:18:15,568 INFO MainThread:34249 [backend.py:_multiprocessing_setup():104] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
18
+ 2024-10-31 02:18:15,568 INFO MainThread:34249 [wandb_init.py:init():684] backend started and connected
19
+ 2024-10-31 02:18:15,572 INFO MainThread:34249 [wandb_init.py:init():779] updated telemetry
20
+ 2024-10-31 02:18:15,588 INFO MainThread:34249 [wandb_init.py:init():812] communicating run to backend with 90.0 second timeout
21
+ 2024-10-31 02:18:16,248 INFO MainThread:34249 [wandb_init.py:init():863] starting run threads in backend
22
+ 2024-10-31 02:18:16,387 INFO MainThread:34249 [wandb_run.py:_console_start():2465] atexit reg
23
+ 2024-10-31 02:18:16,387 INFO MainThread:34249 [wandb_run.py:_redirect():2313] redirect: wrap_raw
24
+ 2024-10-31 02:18:16,387 INFO MainThread:34249 [wandb_run.py:_redirect():2378] Wrapping output streams.
25
+ 2024-10-31 02:18:16,387 INFO MainThread:34249 [wandb_run.py:_redirect():2403] Redirects installed.
26
+ 2024-10-31 02:18:16,391 INFO MainThread:34249 [wandb_init.py:init():907] run started, returning control to user process
27
+ 2024-10-31 04:19:50,672 INFO MainThread:34249 [wandb_run.py:_finish():2164] finishing run htlou/align-anything/wb1qn4kt
28
+ 2024-10-31 04:19:50,674 INFO MainThread:34249 [wandb_run.py:_atexit_cleanup():2428] got exitcode: 0
29
+ 2024-10-31 04:19:50,675 INFO MainThread:34249 [wandb_run.py:_restore():2410] restore
30
+ 2024-10-31 04:19:50,675 INFO MainThread:34249 [wandb_run.py:_restore():2416] restore done
31
+ 2024-10-31 04:19:55,453 INFO MainThread:34249 [wandb_run.py:_footer_history_summary_info():4049] rendering history
32
+ 2024-10-31 04:19:55,458 INFO MainThread:34249 [wandb_run.py:_footer_history_summary_info():4081] rendering summary
33
+ 2024-10-31 04:19:55,468 INFO MainThread:34249 [wandb_run.py:_footer_sync_info():4008] logging synced files
wandb/run-20241031_021815-wb1qn4kt/run-wb1qn4kt.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d98893c74985fef660c7d16bb88bd90babc127be8663af3fdb6cc4340fa17ecd
3
+ size 6397788