Upload policy weights, train config and readme
Browse files- README.md +1 -1
- config.json +2 -2
- model.safetensors +2 -2
- train_config.json +5 -5
README.md
CHANGED
@@ -6,8 +6,8 @@ model_name: reward_classifier
|
|
6 |
pipeline_tag: robotics
|
7 |
tags:
|
8 |
- robotics
|
9 |
-
- lerobot
|
10 |
- reward_classifier
|
|
|
11 |
---
|
12 |
|
13 |
# Model Card for reward_classifier
|
|
|
6 |
pipeline_tag: robotics
|
7 |
tags:
|
8 |
- robotics
|
|
|
9 |
- reward_classifier
|
10 |
+
- lerobot
|
11 |
---
|
12 |
|
13 |
# Model Card for reward_classifier
|
config.json
CHANGED
@@ -44,8 +44,8 @@
|
|
44 |
"latent_dim": 256,
|
45 |
"image_embedding_pooling_dim": 8,
|
46 |
"dropout_rate": 0.1,
|
47 |
-
"model_name": "
|
48 |
-
"model_type": "
|
49 |
"num_cameras": 1,
|
50 |
"learning_rate": 0.0003,
|
51 |
"weight_decay": 0.01,
|
|
|
44 |
"latent_dim": 256,
|
45 |
"image_embedding_pooling_dim": 8,
|
46 |
"dropout_rate": 0.1,
|
47 |
+
"model_name": "openai/clip-vit-base-patch32",
|
48 |
+
"model_type": "transformer",
|
49 |
"num_cameras": 1,
|
50 |
"learning_rate": 0.0003,
|
51 |
"weight_decay": 0.01,
|
model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:174034fa15024461897747db635b4d17c04efbede9aa24f84226186e34763cd4
|
3 |
+
size 350638652
|
train_config.json
CHANGED
@@ -71,14 +71,14 @@
|
|
71 |
"latent_dim": 256,
|
72 |
"image_embedding_pooling_dim": 8,
|
73 |
"dropout_rate": 0.1,
|
74 |
-
"model_name": "
|
75 |
-
"model_type": "
|
76 |
"num_cameras": 1,
|
77 |
"learning_rate": 0.0003,
|
78 |
"weight_decay": 0.01,
|
79 |
"grad_clip_norm": 1.0
|
80 |
},
|
81 |
-
"output_dir": "outputs/train/2025-09-05/16-
|
82 |
"job_name": "reward-classifier",
|
83 |
"resume": false,
|
84 |
"seed": 2,
|
@@ -91,7 +91,7 @@
|
|
91 |
"save_freq": 1000,
|
92 |
"use_validation": true,
|
93 |
"val_split": 0.05,
|
94 |
-
"val_freq":
|
95 |
"val_batch_size": 16,
|
96 |
"use_policy_training_preset": true,
|
97 |
"optimizer": {
|
@@ -117,7 +117,7 @@
|
|
117 |
"project": "reward-classifier",
|
118 |
"entity": null,
|
119 |
"notes": null,
|
120 |
-
"run_id": "
|
121 |
"mode": null
|
122 |
},
|
123 |
"use_peft": false,
|
|
|
71 |
"latent_dim": 256,
|
72 |
"image_embedding_pooling_dim": 8,
|
73 |
"dropout_rate": 0.1,
|
74 |
+
"model_name": "openai/clip-vit-base-patch32",
|
75 |
+
"model_type": "transformer",
|
76 |
"num_cameras": 1,
|
77 |
"learning_rate": 0.0003,
|
78 |
"weight_decay": 0.01,
|
79 |
"grad_clip_norm": 1.0
|
80 |
},
|
81 |
+
"output_dir": "outputs/train/2025-09-05/16-56-32_reward-classifier",
|
82 |
"job_name": "reward-classifier",
|
83 |
"resume": false,
|
84 |
"seed": 2,
|
|
|
91 |
"save_freq": 1000,
|
92 |
"use_validation": true,
|
93 |
"val_split": 0.05,
|
94 |
+
"val_freq": 500,
|
95 |
"val_batch_size": 16,
|
96 |
"use_policy_training_preset": true,
|
97 |
"optimizer": {
|
|
|
117 |
"project": "reward-classifier",
|
118 |
"entity": null,
|
119 |
"notes": null,
|
120 |
+
"run_id": "9scb7mxz",
|
121 |
"mode": null
|
122 |
},
|
123 |
"use_peft": false,
|