ut-austin-rpl / viola Goto Github PK
View Code? Open in Web Editor NEWOfficial implementation for VIOLA
License: MIT License
Official implementation for VIOLA
License: MIT License
Hello! Thanks for releasing this great works!
I am trying to reproduce this model in the real-world, so i use the viola dataset. I was wondering what is the scale of the dataset ? Such as, the action is absolute or delta? The translation is measaured as "m" or "cm" ? and the rotation is recorded as "rad" or "degree". I find the measures are confusing.
Thanks for answering my questions ~
How to visualize when running the program viola_bc/final_eval_script.py? Where does the offscreen_visualization function in img = offscreen_visualization(env, use_eye_in_hand=cfg.algo.use_eye_in_hand) come from? I didn't find this function
action_token_out = transformer_out[:, :, 0, :].
Hello, i don't know why directly take the first dimension of the output as the action_token_out. After your grouping, the grouped input should follow this order: spatial_context_feature + region_feature + action_token + other obs feature. Would the dimension be changed when they pass through the transformer_decoder?
In addition, about the image augmentation (padding + random_crop), how many crops did you take? I saw around the code, only take the default value: num_crops=1. Doesn't the global feature really get lost if there is only one? Because i saw your code, the feature map is extracted from the cropped image.
Could you help me figure out why and how? Thanks a lot
I have some questions:
For the question 1 and 2 (line 168 on here)
transformer_out = transformer_out.reshape(original_shape) action_token_out = transformer_out[:, :, 0, :] if per_step: action_token_out = action_token_out[:, -1:, :]
TensorUtils.time_distributed
is used on this line?gripper-history
?Thank you in advance!
where np.load(f"scenes/{domain_name}/normal/{eval_run_idx}_{i + rank * 50}.npz")
what is initial_mjstate = env.sim.get_state().flatten()
Thanks for releasing your great works!!
Can you share Real-Robot Dataset's experiment config file(maybe xx_viola.yaml..) and checkpoint .pth file?
I try training the model with your generated datasets, and get this error:
Cannot import name 'CropRandomizer' from 'robomimic.models.base_nets'
With suspecting incorrect installation of robomimic, I go to https://github.com/ARISE-Initiative/robomimic/blob/master/robomimic/models/base_nets.py to check. However, I can't find CropRandomizer and Randomizer either.
Could you help me figure out why and how?
Do you have the scripts to collect the demos on the real robot?
A declarative, efficient, and flexible JavaScript library for building user interfaces.
๐ Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
An Open Source Machine Learning Framework for Everyone
The Web framework for perfectionists with deadlines.
A PHP framework for web artisans
Bring data to life with SVG, Canvas and HTML. ๐๐๐
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
Some thing interesting about web. New door for the world.
A server is a program made to process requests and deliver data to clients.
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
Some thing interesting about visualization, use data art
Some thing interesting about game, make everyone happy.
We are working to build community through open source technology. NB: members must have two-factor auth.
Open source projects and samples from Microsoft.
Google โค๏ธ Open Source for everyone.
Alibaba Open Source for everyone
Data-Driven Documents codes.
China tencent open source team.