Running openvla on multiple gpus/distributed?

#1
by mmajek - opened

Hi! Thank you so much for the research and open sourcing your code.
I am currently running openvla at 6 actions per second on a single rtx 4090 on a 2x4090 machine.
I've been tinkering with the code trying to get it working with device_map = auto, with no luck.
Have I missed something?

OpenVLA Collaboration org

Hmm… not sure about exactly what should happen under the hood when parallelizing. When you load with device_map=auto are you seeing the model split across GPUs/is it just not any faster?

Because our model always needs to encode the image features before generating actions there’s always an upfront cost; we also don’t support generation with a batch size > 1 right now, so this could further impact things.

skaramcheti changed discussion status to closed

Sign up or log in to comment