zhaohui commited on
Commit
23a3cfe
·
1 Parent(s): cf302c2

add r2r models

Browse files
r2r/fine_tuned/cma/config.json ADDED
@@ -0,0 +1,134 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "CMANet"
4
+ ],
5
+ "model_cfg": {
6
+ "checkpoint_folder": "data/checkpoints/20250619_cma_train/ckpts",
7
+ "eval": {
8
+ "action": null,
9
+ "ckpt_to_load": "",
10
+ "len_traj_act": null,
11
+ "max_steps": 195,
12
+ "num_sample": null,
13
+ "pm_threshold": null,
14
+ "rotation_threshold": null,
15
+ "sample": false,
16
+ "save_results": true,
17
+ "split": [
18
+ "val_seen"
19
+ ],
20
+ "start_eval_epoch": -1,
21
+ "step_interval": 50,
22
+ "stop_mode": null,
23
+ "success_distance": 3.0,
24
+ "use_ckpt_config": false
25
+ },
26
+ "il": {
27
+ "batch_size": 2,
28
+ "camera_name": "pano_camera_0",
29
+ "ckpt_to_load": "",
30
+ "dataset_3dgs_root_dir": "data/datasets/3dgs",
31
+ "dataset_grutopia10_root_dir": "data/datasets/grutopia10",
32
+ "dataset_r2r_root_dir": "data/datasets/R2R_VLNCE_v1-3_preprocessed",
33
+ "epochs": 80,
34
+ "filter_failure": {
35
+ "min_rgb_nums": 15,
36
+ "use": true
37
+ },
38
+ "inflection_weight_coef": 3.2,
39
+ "lmdb_features_dir": "data/sample_episodes/20250211_sample_origin/sample_data.lmdb",
40
+ "lmdb_map_size": 1000000000000.0,
41
+ "load_from_ckpt": false,
42
+ "load_from_pretrain": false,
43
+ "loss": {
44
+ "alpha": 0.0001,
45
+ "dist_scale": 1
46
+ },
47
+ "lr": 0.0001,
48
+ "num_workers": 8,
49
+ "report_to": "wandb",
50
+ "save_filter_frozen_weights": false,
51
+ "save_interval_epochs": 5,
52
+ "save_interval_steps": null,
53
+ "use_descrete_dataset": null,
54
+ "use_iw": true,
55
+ "warmup_ratio": 0.05,
56
+ "weight_decay": 1e-05
57
+ },
58
+ "local_rank": 0,
59
+ "log_dir": "data/checkpoints/20250619_cma_train/logs",
60
+ "model": {
61
+ "ablate_depth": null,
62
+ "ablate_instruction": null,
63
+ "ablate_rgb": null,
64
+ "cross_modal_encoder": null,
65
+ "depth_encoder": {
66
+ "backbone": "resnet50",
67
+ "cnn_type": "VlnResnetDepthEncoder",
68
+ "ddppo_checkpoint": "data/ddppo-models/gibson-4plus-mp3d-train-val-test-resnet50.pth",
69
+ "output_size": 128,
70
+ "trainable": false
71
+ },
72
+ "diffusion_policy": null,
73
+ "distance_predictor": null,
74
+ "image_encoder": null,
75
+ "imu_encoder": null,
76
+ "instruction_encoder": {
77
+ "bidirectional": true,
78
+ "dataset_vocab": "data/datasets/R2R_VLNCE_v1-3_preprocessed/train/train.json.gz",
79
+ "embedding_file": "data/datasets/R2R_VLNCE_v1-3_preprocessed/embeddings.json.gz",
80
+ "embedding_size": 50,
81
+ "final_state_only": true,
82
+ "fine_tune_embeddings": false,
83
+ "hidden_size": 128,
84
+ "load_model": null,
85
+ "max_length": null,
86
+ "rnn_type": "LSTM",
87
+ "sensor_uuid": "instruction",
88
+ "use_pretrained_embeddings": true,
89
+ "vocab_size": 2504
90
+ },
91
+ "learn_angle": null,
92
+ "len_traj_act": 4,
93
+ "max_step": 200,
94
+ "normalize_rgb": null,
95
+ "policy_name": "CMA_Policy",
96
+ "prev_action_encoder": null,
97
+ "progress_monitor": {
98
+ "alpha": 1.0,
99
+ "concat_state_txt": null,
100
+ "use": true
101
+ },
102
+ "rgb_encoder": {
103
+ "cnn_type": "TorchVisionResNet50",
104
+ "output_size": 256,
105
+ "trainable": false
106
+ },
107
+ "seq2seq": null,
108
+ "state_encoder": {
109
+ "dropout_rate": null,
110
+ "hidden_size": 512,
111
+ "num_recurrent_layers": 2,
112
+ "rgb_depth_embed_method": null,
113
+ "rnn_type": "GRU",
114
+ "use_dropout": null
115
+ },
116
+ "stop_progress_predictor": null,
117
+ "text_encoder": null
118
+ },
119
+ "model_name": "cma",
120
+ "name": "20250619_cma_train",
121
+ "num_gpus": 1,
122
+ "output_dir": "data/checkpoints/20250619_cma_train/ckpts",
123
+ "seed": 0,
124
+ "tensorboard_dir": "data/checkpoints/20250619_cma_train/tensorboard",
125
+ "torch_gpu_id": 0,
126
+ "torch_gpu_ids": [
127
+ 0
128
+ ],
129
+ "world_size": 1
130
+ },
131
+ "model_type": "cma",
132
+ "torch_dtype": "float32",
133
+ "transformers_version": "4.26.1"
134
+ }
r2r/fine_tuned/cma/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:572c625623155f301a114485987128d1317541940fe94432fa92e62e532d66d3
3
+ size 49260090
r2r/fine_tuned/cma/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:686f32f754bd0382831574ba3a29090b5e1d54edc0cbd48d89f5a816a67a4b69
3
+ size 147809914
r2r/fine_tuned/cma/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:91a9d8b0c5f7466cc65f08b3acfeb448ddb0262da4161bb9d8e89a0ac4f17caa
3
+ size 14244
r2r/fine_tuned/cma/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8f081d2d07b19d3825f13778a499dadd1508d8cabea562ed222a9c7c64094e28
3
+ size 1064
r2r/fine_tuned/cma/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
r2r/fine_tuned/cma_plus/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:91737ae7fa4a190a53efa6c843c9cbe0319db1d024c9da55eaea84fbef4e96b3
3
+ size 147803318
r2r/fine_tuned/rdp/config.json ADDED
@@ -0,0 +1,253 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "RDPNet"
4
+ ],
5
+ "model_cfg": {
6
+ "checkpoint_folder": "data/checkpoints/20250619_rdp_train_new/ckpts",
7
+ "eval": {
8
+ "action": "descrete",
9
+ "ckpt_to_load": null,
10
+ "len_traj_act": 2,
11
+ "max_steps": 195,
12
+ "num_sample": 1,
13
+ "pm_threshold": 0.9,
14
+ "rotation_threshold": 0.01,
15
+ "sample": true,
16
+ "save_results": true,
17
+ "split": [
18
+ "val_unseen"
19
+ ],
20
+ "start_eval_epoch": -1,
21
+ "step_interval": 80,
22
+ "stop_mode": "stop_progress",
23
+ "success_distance": 3.0,
24
+ "use_ckpt_config": false
25
+ },
26
+ "il": {
27
+ "batch_size": 4,
28
+ "camera_name": "pano_camera_0",
29
+ "ckpt_to_load": "",
30
+ "dataset_3dgs_root_dir": "data/datasets/3dgs",
31
+ "dataset_grutopia10_root_dir": "data/datasets/grutopia10",
32
+ "dataset_r2r_root_dir": "data/datasets/R2R_VLNCE_v1-3_corrected",
33
+ "epochs": 50,
34
+ "filter_failure": {
35
+ "min_rgb_nums": 15,
36
+ "use": true
37
+ },
38
+ "inflection_weight_coef": null,
39
+ "lmdb_features_dir": "data/sample_episodes/20250211_sample_origin/sample_data.lmdb",
40
+ "load_from_ckpt": false,
41
+ "load_from_pretrain": true,
42
+ "loss": {
43
+ "alpha": 0.0001,
44
+ "dist_scale": 1
45
+ },
46
+ "lr": 0.0001,
47
+ "num_workers": 8,
48
+ "report_to": "wandb",
49
+ "save_filter_frozen_weights": true,
50
+ "save_interval_epochs": 5,
51
+ "save_interval_steps": null,
52
+ "use_descrete_dataset": true,
53
+ "use_iw": null,
54
+ "warmup_ratio": 0.1,
55
+ "weight_decay": 0.0001
56
+ },
57
+ "local_rank": 0,
58
+ "log_dir": "data/checkpoints/20250619_rdp_train_new/logs",
59
+ "model": {
60
+ "ablate_depth": null,
61
+ "ablate_instruction": null,
62
+ "ablate_rgb": null,
63
+ "cross_modal_encoder": {
64
+ "hidden_size": 512,
65
+ "input_type": 3,
66
+ "load_model": false,
67
+ "num_attention_heads": 8,
68
+ "num_x_layers": 2,
69
+ "txt_to_img": true,
70
+ "txt_to_img_layer": 2
71
+ },
72
+ "depth_encoder": null,
73
+ "diffusion_policy": {
74
+ "action_stats": {
75
+ "max": [
76
+ 0.25,
77
+ 0.25,
78
+ 0.27
79
+ ],
80
+ "min": [
81
+ -0.25,
82
+ -0.25,
83
+ -0.27
84
+ ]
85
+ },
86
+ "clip_sample": true,
87
+ "cls_free_guidance_scale": 1.5,
88
+ "cls_mask_method": "mask_token",
89
+ "cls_mask_ratio": 0.25,
90
+ "len_traj_pred": 8,
91
+ "metric_waypoint_spacing": 1,
92
+ "num_diffusion_iters": 20,
93
+ "pred_type": "epsilon",
94
+ "random_mask_instr": true,
95
+ "random_mask_rgb": true,
96
+ "scheduler": "DDPM",
97
+ "stop_weight": 1,
98
+ "transformer_encoding_size": 512,
99
+ "transformer_n_cond_layers": 1,
100
+ "transformer_n_layers": 3,
101
+ "transformer_p_drop_emb": 0.2,
102
+ "txt_len": 80,
103
+ "type": "transformer",
104
+ "use": true,
105
+ "use_cls_free_guidance": true,
106
+ "waypoint_spacing": 1
107
+ },
108
+ "distance_predictor": {
109
+ "normalize": false,
110
+ "use": false
111
+ },
112
+ "eval": {
113
+ "action": "descrete",
114
+ "auto_remove": false,
115
+ "ckpt_to_load": null,
116
+ "distance_threshold": 1.5,
117
+ "episode_count": -1,
118
+ "len_traj_act": 4,
119
+ "load_eval_subset": true,
120
+ "max_len_traj_act": 8,
121
+ "max_steps": 195,
122
+ "min_displacement": 0.15,
123
+ "min_len_traj_act": 3,
124
+ "num_sample": 1,
125
+ "pm_threshold": 0.9,
126
+ "re_eval": false,
127
+ "rotation_threshold": 0.01,
128
+ "sample": null,
129
+ "save_results": true,
130
+ "sim_cfg_file": "vln/configs/sim_cfg_policy_h1_eval.yaml",
131
+ "split": [
132
+ "val_unseen"
133
+ ],
134
+ "start_eval_epoch": -1,
135
+ "step_interval": 80,
136
+ "stop_mode": "stop_progress",
137
+ "stop_progress_threshold": 0.85,
138
+ "stop_x_threshold": 0.015,
139
+ "stop_y_threshold": 0.015,
140
+ "stop_yaw_threshold": 0.05,
141
+ "success_distance": 3.0,
142
+ "train_eval_interval": 100,
143
+ "use_ckpt_config": false,
144
+ "use_dynamic_len_traj_act": false,
145
+ "vln_cfg_file": "vln/configs/vln_cfg_policy_eval.yaml"
146
+ },
147
+ "image_encoder": {
148
+ "depth": {
149
+ "backbone": "resnet50",
150
+ "bottleneck": "resnet",
151
+ "cnn_type": "VlnResnetDepthEncoder",
152
+ "ddppo_checkpoint": "data/ddppo-models/gibson-4plus-mp3d-train-val-test-resnet50.pth",
153
+ "feature_dim": 768,
154
+ "load_model": true,
155
+ "output_size": 128,
156
+ "projection_dim": 512,
157
+ "update_depth_encoder": false
158
+ },
159
+ "dropout": 0.1,
160
+ "env_drop": 0.3,
161
+ "img_stack_nums": 4,
162
+ "rgb": {
163
+ "feature_dim": 768,
164
+ "img_mod": "multi_patches_avg_pooling",
165
+ "load_model": true,
166
+ "model_name": "clip-long",
167
+ "model_path": "data/pretrained/clip-long/longclip-B.pt",
168
+ "multi_patches_num": 5,
169
+ "projection_dim": 512,
170
+ "rgb_proj": false,
171
+ "update_rgb_encoder": false
172
+ },
173
+ "use_env_drop": true,
174
+ "use_stack": false
175
+ },
176
+ "imu_encoder": {
177
+ "encoding_size": 64,
178
+ "input_size": 3,
179
+ "to_local_coords": true,
180
+ "use": true
181
+ },
182
+ "instruction_encoder": null,
183
+ "learn_angle": true,
184
+ "len_traj_act": 4,
185
+ "max_step": 200,
186
+ "normalize_rgb": null,
187
+ "policy_name": "RDP_Policy",
188
+ "prev_action_encoder": {
189
+ "encoding_size": 64,
190
+ "input_size": null,
191
+ "to_local_coords": null,
192
+ "type": "continuous",
193
+ "use": null
194
+ },
195
+ "progress_monitor": {
196
+ "alpha": null,
197
+ "concat_state_txt": true,
198
+ "use": true
199
+ },
200
+ "rgb_encoder": null,
201
+ "seq2seq": null,
202
+ "state_encoder": {
203
+ "dropout_rate": 0.2,
204
+ "hidden_size": 512,
205
+ "num_recurrent_layers": 1,
206
+ "rgb_depth_embed_method": "flat",
207
+ "rnn_type": "GRU",
208
+ "use_dropout": false
209
+ },
210
+ "stop_progress_predictor": {
211
+ "concat_state_txt": true,
212
+ "loss_alpha": 10,
213
+ "type": "continuous",
214
+ "use": true
215
+ },
216
+ "text_encoder": {
217
+ "ablate": null,
218
+ "embedding_size": 512,
219
+ "eot_token": 49407,
220
+ "final_state_only": null,
221
+ "hidden_size": 512,
222
+ "load_model": true,
223
+ "max_length": 248,
224
+ "model_name": "clip-long",
225
+ "model_path": "data/pretrained/clip-long/longclip-B.pt",
226
+ "num_l_layers": 6,
227
+ "pad_token": 0,
228
+ "sot_token": 49406,
229
+ "type": "clip-long",
230
+ "update_text_encoder": false,
231
+ "vocab_size": 50265
232
+ },
233
+ "use_iw": false
234
+ },
235
+ "model_name": "rdp",
236
+ "name": "20250619_rdp_train_new",
237
+ "num_gpus": 4,
238
+ "output_dir": "data/checkpoints/20250619_rdp_train_new/ckpts",
239
+ "seed": 0,
240
+ "tensorboard_dir": "data/checkpoints/20250619_rdp_train_new/tensorboard",
241
+ "torch_gpu_id": 0,
242
+ "torch_gpu_ids": [
243
+ 0,
244
+ 1,
245
+ 2,
246
+ 3
247
+ ],
248
+ "world_size": 4
249
+ },
250
+ "model_type": "rdp",
251
+ "torch_dtype": "float32",
252
+ "transformers_version": "4.26.1"
253
+ }
r2r/fine_tuned/rdp/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37d7fab707312afd3ea12a65c84edbeb1fa50c668d0036ff93e21a3919ff06b1
3
+ size 368891962
r2r/fine_tuned/rdp/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0536d140be240a0fd5fb5894b179d453a06fa1c7f96a8c987b35543b01d00599
3
+ size 565168086
r2r/fine_tuned/rdp/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dbbdc54615092c7ae7b6703cfcbf9188e8913651c406ec7bc482f02df67dcb9d
3
+ size 15006
r2r/fine_tuned/rdp/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4640b4a7e0c9ff2f9b7bead4a5038207678e9cbe45e41030074a0f2ed9b42ee5
3
+ size 1064
r2r/fine_tuned/rdp/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
r2r/fine_tuned/seq2seq/config.json ADDED
@@ -0,0 +1,136 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Seq2SeqNet"
4
+ ],
5
+ "model_cfg": {
6
+ "checkpoint_folder": "data/checkpoints/20250619_seq2seq_train/ckpts",
7
+ "eval": {
8
+ "action": null,
9
+ "ckpt_to_load": null,
10
+ "len_traj_act": null,
11
+ "max_steps": 195,
12
+ "num_sample": null,
13
+ "pm_threshold": null,
14
+ "rotation_threshold": 0.01,
15
+ "sample": false,
16
+ "save_results": true,
17
+ "split": [
18
+ "val_seen"
19
+ ],
20
+ "start_eval_epoch": -1,
21
+ "step_interval": 50,
22
+ "stop_mode": null,
23
+ "success_distance": 3.0,
24
+ "use_ckpt_config": false
25
+ },
26
+ "il": {
27
+ "batch_size": 2,
28
+ "camera_name": "pano_camera_0",
29
+ "ckpt_to_load": "",
30
+ "dataset_3dgs_root_dir": "data/datasets/3dgs",
31
+ "dataset_grutopia10_root_dir": "data/datasets/grutopia10",
32
+ "dataset_r2r_root_dir": "data/datasets/R2R_VLNCE_v1-3_preprocessed",
33
+ "epochs": 100,
34
+ "filter_failure": {
35
+ "min_rgb_nums": 15,
36
+ "use": true
37
+ },
38
+ "inflection_weight_coef": 3.2,
39
+ "lmdb_features_dir": "data/sample_episodes/20250211_sample_origin/sample_data.lmdb",
40
+ "lmdb_map_size": 1000000000000.0,
41
+ "load_from_ckpt": false,
42
+ "load_from_pretrain": true,
43
+ "loss": {
44
+ "alpha": 0.0001,
45
+ "dist_scale": 1
46
+ },
47
+ "lr": 0.0001,
48
+ "num_workers": 8,
49
+ "report_to": "wandb",
50
+ "save_filter_frozen_weights": false,
51
+ "save_interval_epochs": 5,
52
+ "save_interval_steps": null,
53
+ "use_descrete_dataset": null,
54
+ "use_iw": true,
55
+ "warmup_ratio": 0.1,
56
+ "weight_decay": 0.0001
57
+ },
58
+ "local_rank": 0,
59
+ "log_dir": "data/checkpoints/20250619_seq2seq_train/logs",
60
+ "model": {
61
+ "ablate_depth": null,
62
+ "ablate_instruction": null,
63
+ "ablate_rgb": null,
64
+ "cross_modal_encoder": null,
65
+ "depth_encoder": {
66
+ "backbone": "resnet50",
67
+ "cnn_type": "VlnResnetDepthEncoder",
68
+ "ddppo_checkpoint": "data/ddppo-models/gibson-4plus-mp3d-train-val-test-resnet50.pth",
69
+ "output_size": 128,
70
+ "trainable": false
71
+ },
72
+ "diffusion_policy": null,
73
+ "distance_predictor": null,
74
+ "image_encoder": null,
75
+ "imu_encoder": null,
76
+ "instruction_encoder": {
77
+ "bidirectional": false,
78
+ "dataset_vocab": "data/datasets/R2R_VLNCE_v1-3_preprocessed/train/train.json.gz",
79
+ "embedding_file": "data/datasets/R2R_VLNCE_v1-3_preprocessed/embeddings.json.gz",
80
+ "embedding_size": 50,
81
+ "final_state_only": true,
82
+ "fine_tune_embeddings": false,
83
+ "hidden_size": 128,
84
+ "load_model": null,
85
+ "max_length": null,
86
+ "rnn_type": "LSTM",
87
+ "sensor_uuid": "instruction",
88
+ "use_pretrained_embeddings": true,
89
+ "vocab_size": 2504
90
+ },
91
+ "learn_angle": null,
92
+ "len_traj_act": 4,
93
+ "max_step": 200,
94
+ "normalize_rgb": null,
95
+ "policy_name": "Seq2Seq_Policy",
96
+ "prev_action_encoder": null,
97
+ "progress_monitor": {
98
+ "alpha": 1.0,
99
+ "concat_state_txt": null,
100
+ "use": true
101
+ },
102
+ "rgb_encoder": {
103
+ "cnn_type": "TorchVisionResNet50",
104
+ "output_size": 256,
105
+ "trainable": false
106
+ },
107
+ "seq2seq": {
108
+ "use_prev_action": false
109
+ },
110
+ "state_encoder": {
111
+ "dropout_rate": null,
112
+ "hidden_size": 512,
113
+ "num_recurrent_layers": 1,
114
+ "rgb_depth_embed_method": null,
115
+ "rnn_type": "GRU",
116
+ "use_dropout": null
117
+ },
118
+ "stop_progress_predictor": null,
119
+ "text_encoder": null
120
+ },
121
+ "model_name": "seq2seq",
122
+ "name": "20250619_seq2seq_train",
123
+ "num_gpus": 1,
124
+ "output_dir": "data/checkpoints/20250619_seq2seq_train/ckpts",
125
+ "seed": 0,
126
+ "tensorboard_dir": "data/checkpoints/20250619_seq2seq_train/tensorboard",
127
+ "torch_gpu_id": 0,
128
+ "torch_gpu_ids": [
129
+ 0
130
+ ],
131
+ "world_size": 1
132
+ },
133
+ "model_type": "seq2seq",
134
+ "torch_dtype": "float32",
135
+ "transformers_version": "4.26.1"
136
+ }
r2r/fine_tuned/seq2seq/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b1a9c011abecf2958a513a5269d1f650ff54dd3ce72ba9c813005b629604c771
3
+ size 19673710
r2r/fine_tuned/seq2seq/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c5570070c8697664f46668bdc96727381124c4c4d1f0e20e75c8940d4274d1c4
3
+ size 133019958
r2r/fine_tuned/seq2seq/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b1c1483b2f971b5b4fb5c7742b5ae188a8eafdb358c870fb90a320a187afabd1
3
+ size 14244
r2r/fine_tuned/seq2seq/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b0508238d3474230955f9d0198dbf096ccc53817711c80525217e2d61170a4af
3
+ size 1064
r2r/fine_tuned/seq2seq/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
r2r/fine_tuned/seq2seq_plus/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1b510fc3e18a1f57cc714af9fc975dcb60eddfcf76aa51daae582bd864dbec4a
3
+ size 133013046
r2r/zero_shot/cma/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:693259a1e8d426c64573bbc838e430dba0e57cbe2e2b8d9a14b3b80e8c4fd448
3
+ size 147809576
r2r/zero_shot/seq2seq/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:81284a6a74998db7c91f4bf9cac6865715a7ced3300ff12386675d861aa66faf
3
+ size 133012722