arch: | |
H_cycles: 2 | |
H_layers: 4 | |
L_cycles: 2 | |
L_layers: 4 | |
expansion: 4 | |
halt_exploration_prob: 0.1 | |
halt_max_steps: 16 | |
hidden_size: 512 | |
loss: | |
loss_type: stablemax_cross_entropy | |
name: losses@ACTLossHead | |
name: hrm.hrm_act_v1@HierarchicalReasoningModel_ACTV1 | |
num_heads: 8 | |
pos_encodings: rope | |
puzzle_emb_ndim: 512 | |
beta1: 0.9 | |
beta2: 0.95 | |
checkpoint_every_eval: true | |
checkpoint_path: checkpoints/Arc-2-aug-1000 ACT-torch/HierarchicalReasoningModel_ACTV1 | |
bright-mustang | |
data_path: data/arc-2-aug-1000 | |
epochs: 100000 | |
eval_interval: 10000 | |
eval_save_outputs: [] | |
global_batch_size: 768 | |
lr: 0.0001 | |
lr_min_ratio: 1.0 | |
lr_warmup_steps: 2000 | |
project_name: Arc-2-aug-1000 ACT-torch | |
puzzle_emb_lr: 0.01 | |
puzzle_emb_weight_decay: 0.1 | |
run_name: HierarchicalReasoningModel_ACTV1 bright-mustang | |
seed: 0 | |
weight_decay: 0.1 | |