Hugging Face
Models
Datasets
Spaces
Community
Docs
Enterprise
Pricing
Log In
Sign Up
9
16
37
PommesPeter
PommesPeter
Follow
21world's profile picture
HAUZachery's profile picture
ltim's profile picture
5 followers
·
19 following
PommesPeter
AI & ML interests
MM-LLM
Recent Activity
updated
a model
2 days ago
PommesPeter/dp_ckpts
reacted
to
IliaLarchenko
's
post
with 👍
4 days ago
I am presenting Decoder-Only Transformer (DOT) Policy a simple Behavioral Control policy that outperforms SOTA models on two simple benchmark tasks: ✅ PushT (pushing an object to a goal) – 84% success on keypoints, 74% on images (previous best: 75% / 69%) ✅ ALOHA Insert (precise bimanual insertion) – 30% success (previous best: ~21%) The best part? DOT is much smaller (sometimes 100 times less parameters) than previous SOTA models, trains faster, and avoids complexity: 🚫 No generative models (Diffusion, VAE, GANs) 🚫 No discretization/tokenization of actions 🚫 No reinforcement learning or multi-stage training ✅ Just learns from human demos, plain and simple This is still early — more complex real-life tasks need testing, and no guarantees it will actually work well there, but I think it's interesting to share. Sometimes, simpler approaches can be just as effective (or even better) than complex ones. 🔗 Open-source code and detailed description: https://github.com/IliaLarchenko/dot_policy Trained models on Hugging Face: https://huggingface.co/IliaLarchenko/dot_pusht_keypoints https://huggingface.co/IliaLarchenko/dot_pusht_images https://huggingface.co/IliaLarchenko/dot_bimanual_insert
reacted
to
IliaLarchenko
's
post
with 🔥
4 days ago
I am presenting Decoder-Only Transformer (DOT) Policy a simple Behavioral Control policy that outperforms SOTA models on two simple benchmark tasks: ✅ PushT (pushing an object to a goal) – 84% success on keypoints, 74% on images (previous best: 75% / 69%) ✅ ALOHA Insert (precise bimanual insertion) – 30% success (previous best: ~21%) The best part? DOT is much smaller (sometimes 100 times less parameters) than previous SOTA models, trains faster, and avoids complexity: 🚫 No generative models (Diffusion, VAE, GANs) 🚫 No discretization/tokenization of actions 🚫 No reinforcement learning or multi-stage training ✅ Just learns from human demos, plain and simple This is still early — more complex real-life tasks need testing, and no guarantees it will actually work well there, but I think it's interesting to share. Sometimes, simpler approaches can be just as effective (or even better) than complex ones. 🔗 Open-source code and detailed description: https://github.com/IliaLarchenko/dot_policy Trained models on Hugging Face: https://huggingface.co/IliaLarchenko/dot_pusht_keypoints https://huggingface.co/IliaLarchenko/dot_pusht_images https://huggingface.co/IliaLarchenko/dot_bimanual_insert
View all activity
Organizations
PommesPeter
's datasets
3
Sort: Recently updated
PommesPeter/libero_pick_up_bowl_split
Viewer
•
Updated
18 days ago
•
22.9k
•
138
PommesPeter/imelodist-increment
Updated
Mar 13, 2024
•
4
•
1
PommesPeter/imelodist-sft
Updated
Mar 13, 2024
•
9
•
1