Skip to content

stanford-iprl-lab/multimodal_representation

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

11 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Multimodal Representation

Code for Making Sense of Vision and Touch. https://sites.google.com/view/visionandtouch

Code written by: Matthew Tan, Michelle Lee, Peter Zachares, Yuke Zhu

requirements

pip install -r requirements.txt

get dataset

cd multimodal/dataset
./download_data.sh

run training

python mini_main.py --config configs/training_default.yaml

ROBOT DATASET


action Dataset {50, 4}
contact Dataset {50, 50}\ depth_data Dataset {50, 128, 128, 1}
ee_forces_continuous Dataset {50, 50, 6}
ee_ori Dataset {50, 4}
ee_pos Dataset {50, 3}
ee_vel Dataset {50, 3}
ee_vel_ori Dataset {50, 3}
ee_yaw Dataset {50, 4}
ee_yaw_delta Dataset {50, 4}
image Dataset {50, 128, 128, 3}
joint_pos Dataset {50, 7}
joint_vel Dataset {50, 7}
optical_flow Dataset {50, 128, 128, 2}
proprio Dataset {50, 8}\

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published