| run_name: glue_20251002_155411 |
| seed: 6198 |
| epoch: null |
| dry_run: false |
| model: |
| d_model: 3584 |
| n_heads: 28 |
| n_kv_heads: 4 |
| qkv_bias: true |
| clip_qkv: null |
| n_layers: 28 |
| mlp_ratio: 4 |
| mlp_hidden_size: 37888 |
| activation_type: swiglu |
| block_type: sequential |
| block_group_size: 1 |
| rope: true |
| rope_full_precision: true |
| rope_theta: 1000000.0 |
| vision_backbone: |
| image_model_type: openai |
| image_default_input_size: |
| - 336 |
| - 336 |
| image_patch_size: 14 |
| image_pos_patch_size: 14 |
| image_emb_dim: 1024 |
| image_num_heads: 16 |
| image_num_key_value_heads: 16 |
| image_num_layers: 23 |
| image_head_dim: 64 |
| image_mlp_dim: 4096 |
| image_mlp_activations: quick_gelu |
| image_dropout_rate: 0.0 |
| image_num_pos: 577 |
| image_norm_eps: 1.0e-05 |
| attention_dropout: 0.0 |
| residual_dropout: 0.0 |
| initializer_range: 0.02 |
| fsdp_wrap: false |
| resize_mode: default |
| vit_load_path: /weka/oe-training-default/mm-olmo/pretrained_image_encoders/vit-l-14-336.pt |
| llm_load_path: /weka/oe-training-default/mm-olmo/pretrained_llms/qwen2-7b.pt |
| low_cpu_fsdp: true |
| attention_type: sdpa |
| float32_attention: true |
| attention_dropout: 0.0 |
| attention_layer_norm: false |
| residual_dropout: 0.1 |
| response_residual_dropout: 0.0 |
| embedding_dropout: 0.0 |
| layer_norm_type: rms |
| layer_norm_with_affine: true |
| layer_norm_eps: 1.0e-06 |
| attention_layer_norm_with_affine: true |
| max_sequence_length: 4096 |
| max_position_embeddings: null |
| include_bias: false |
| bias_for_layer_norm: null |
| scale_logits: false |
| vocab_size: 152064 |
| embedding_size: 152064 |
| ff_out_size: null |
| additional_vocab_size: 128 |
| new_embedding_init_range: 0.02 |
| weight_tying: false |
| init_device: null |
| init_fn: normal |
| init_std: 0.02 |
| init_cutoff_factor: null |
| norm_after: false |
| precision: amp_bf16 |
| max_crops: 12 |
| crop_mode: overlap-and-resize-c2 |
| use_col_tokens: true |
| prompt_type: uber_model |
| system_prompt_kind: demo_or_style |
| message_formatting: role |
| always_start_with_space: true |
| multi_annotation_weighting: root_subsegments |
| default_inference_len: 65 |
| overlap_margins: |
| - 4 |
| - 4 |
| pad_value: 0.0 |
| image_padding_embed: pad_and_partial_pad |
| fix_image_padding: true |
| vit_layers: |
| - -2 |
| - -9 |
| image_pooling_h: 2 |
| image_pooling_w: 2 |
| image_pooling_2d: attention_meanq |
| image_projector: mlp |
| image_feature_dropout: 0.0 |
| initializer_range: 0.02 |
| normalize_input_embeds: false |
| use_position_ids: true |
| head_dim: null |
| action_tokenizer: |
| identifier: physical-intelligence/fast |
| tokenizer_dir: null |
| action_dim: 7 |
| horizon: 8 |
| tokenizer: |
| identifier: Qwen/Qwen2-7B |
| tokenizer_dir: null |
| pad_tokenizer: true |
| moe_num_experts: 8 |
| moe_top_k: 2 |
| moe_mlp_impl: sparse |
| moe_log_expert_assignment: false |
| moe_shared_expert: false |
| moe_lbl_in_fp32: false |
| moe_interleave: false |
| moe_loss_weight: 0.1 |
| moe_zloss_weight: null |
| moe_dropless: true |
| moe_capacity_factor: 1.25 |
| action_head: l1_regression |
| num_diffusion_steps: 1000 |
| num_diffusion_inference_steps: 30 |
| use_proprio: true |
| action_head_dit_hidden_size: 1152 |
| action_head_dit_depth: 28 |
| action_head_dit_num_heads: 16 |
| llm_causal_attention: false |
| action_use_left_eef: true |
| action_use_mobile_base: false |
| allow_resume: false |
| ft_llm: true |
| ft_vit: false |
| ft_connector: false |
| ft_embedding: lm_head |
| lora: false |
| use_lora: true |
| lora_rank: 8 |
| lora_llm: false |
| lora_vit: false |
| lora_connector: false |
| early_exit: false |
| train_exit_random_layer: false |
| optimizer: |
| name: adamw |
| learning_rate: 0.0001 |
| weight_decay: 0.01 |
| betas: |
| - 0.9 |
| - 0.95 |
| eps: 1.0e-05 |
| connector_learning_rate: 0.0002 |
| vit_learning_rate: 6.0e-06 |
| llm_learning_rate: 5.0e-05 |
| connector_weight_decay: 0.0 |
| vit_weight_decay: 0.0 |
| llm_weight_decay: 0.0 |
| connector_betas: |
| - 0.9 |
| - 0.95 |
| vit_betas: |
| - 0.9 |
| - 0.95 |
| llm_betas: |
| - 0.9 |
| - 0.95 |
| connector_eps: 1.0e-06 |
| vit_eps: 1.0e-06 |
| llm_eps: 1.0e-06 |
| metrics_log_interval: 20 |
| scheduler: |
| name: multimodal |
| units: steps |
| t_warmup: 100 |
| t_max: null |
| alpha_f: 0.1 |
| connector_t_warmup: 200 |
| vit_t_warmup: 2000 |
| llm_t_warmup: 2000 |
| grad_clip_warmup_steps: null |
| grad_clip_warmup_factor: null |
| warmup_min_lr: 0.0 |
| data: |
| dataset: vla_dataset_realworld |
| mixture: null |
| root_size_mixture: null |
| split: train |
| seed: 95818 |
| shuffle_messages: false |
| pad: to_max |
| sequence_length: 1600 |
| shuffle: true |
| for_inference: false |
| multi_modal: torch |
| num_workers: 0 |
| drop_last: true |
| pin_memory: true |
| prefetch_factor: null |
| persistent_workers: false |
| timeout: 0 |
| rlds_dataset_name: libero_4_task_suites_no_noops |
| rlds_data_root_dir: /mnt/data/zhangjian/dataset/Simulation/datasets--openvla--modified_libero_rlds |
| use_wrist_image: true |
| use_proprio: true |
| rlds_shuffle_buffer_size: 100000 |
| rlds_traj_threads: 8 |
| rlds_read_threads: 8 |
| lerobot_episode_index_start: null |
| lerobot_episode_index_end: null |
| restore_dataloader: true |
| fast_forward_batches: null |
| evaluators: |
| - label: val |
| data: |
| dataset: vla_dataset_realworld |
| mixture: null |
| root_size_mixture: null |
| split: validation |
| seed: null |
| shuffle_messages: false |
| pad: to_max |
| sequence_length: 1600 |
| shuffle: false |
| for_inference: false |
| multi_modal: torch |
| num_workers: 0 |
| drop_last: true |
| pin_memory: true |
| prefetch_factor: null |
| persistent_workers: true |
| timeout: 0 |
| rlds_dataset_name: libero_4_task_suites_no_noops |
| rlds_data_root_dir: /mnt/data/zhangjian/dataset/Simulation/datasets--openvla--modified_libero_rlds |
| use_wrist_image: true |
| use_proprio: true |
| rlds_shuffle_buffer_size: 256000 |
| rlds_traj_threads: 8 |
| rlds_read_threads: 8 |
| lerobot_episode_index_start: 353 |
| lerobot_episode_index_end: 765 |
| device_eval_batch_size: null |
| subset_num_batches: 64 |
| max_examples: null |
| max_new_tokens: 448 |
| mm_evaluator: null |
| save_dir: null |
| save_to_checkpoint_dir: false |
| eval_name: null |
| skip_if_metrics_cached: true |
| eval_interval: 0 |
| inf_eval_interval: -1 |
| inf_evaluators: [] |
| save_folder: /vast/users/xiaodan/workspace/minghao.guo/warehouse_a1/ckpt/ |
| remote_save_folder: null |
| canceled_check_interval: 50 |
| save_interval: 500 |
| save_interval_unsharded: 500 |
| save_interval_ephemeral: null |
| save_interval_action_head: 500 |
| save_num_checkpoints_to_keep: 1 |
| save_num_unsharded_checkpoints_to_keep: 1 |
| save_num_action_head_checkpoints_to_keep: 2 |
| save_overwrite: true |
| force_save_unsharded: false |
| no_pre_train_checkpoint: true |
| initial_model_checkpoint: /vast/users/xiaodan/zhangjian/molmo_data/Molmo-7B-D-0924 |
| load_model_config: null |
| checkpoint_dir: /vast/users/xiaodan/zhangjian/molmo_data/Molmo-7B-D-0924 |
| load_path: null |
| load_path_sharded_checkpointer: null |
| reset_optimizer_state: false |
| reset_trainer_state: false |
| save_dataloader_state: false |
| reset_dataloader_state: false |
| keep_lr_on_load: true |
| sharded_checkpointer: torch_legacy |
| max_duration: 500000 |
| global_train_batch_size: 126 |
| device_train_batch_size: 15 |
| device_train_microbatch_size: 16 |
| device_eval_batch_size: 4 |
| eval_subset_num_batches: -1 |
| eval_on_load: false |
| device_inf_eval_batch_size: 16 |
| inf_eval_subset_num_batches: -1 |
| device_train_grad_accum: 0 |
| max_grad_norm: 1.0 |
| multi_component_grad_norm: true |
| batch_divisor: global_batch |
| max_grad_norm_ratio: null |
| precision: amp_bf16 |
| wandb: |
| project: a1-realworld |
| entity: henryeap |
| group: null |
| name: glue_20251002_155411 |
| tags: |
| - watching |
| log_artifacts: false |
| rank_zero_only: true |
| log_interval: 1 |
| speed_monitor: |
| window_size: 20 |
| gpu_flops_available: null |
| console_log_interval: 1 |
| gen1_gc_interval: 1 |
| compile: null |
| fsdp: |
| use_orig_params: true |
| sharding_strategy: FULL_SHARD |
| wrapping_strategy: by_block_and_size |
| precision: float |
| hybrid_sharding_num_model_replicas: null |
| softmax_auxiliary_loss: true |
| softmax_auxiliary_loss_scale: 0.0001 |
| time_limit: null |
| extra_steps_after_cancel: 10 |
| python_profiling: false |
| torch_profiling: false |
| stop_at: 500000 |
| stop_after: null |
| activation_checkpointing: whole_layer |
| fused_loss: null |
|
|