TAX-PoseD, a method for learning relative placement prediction tasks, learns a spatially-grounded latent distribution over demonstrations without human annotations, using an architecture with geometric inductive biases.
multimodal- Stable latest branch (you are here)multimodal_dev- Latest branchmultimodal_icra2024- ICRA 2024 paper's model configurations
To install dependencies like pytorch, ndf_robot and other libraries, please follow the instructions in the TAX-Pose Github repo.
Then, install this repository with:
pip install -e .
In our paper, we use the same 1-rack NDF training dataset as TAX-Pose, as described here.
We have also experimented with environments from the RPDiff paper.
The current best model config has a long name: joint_train_pzX-dgcnn-transformer_pzY-pn2_hybridlatentz-global_gradclip1e-4_se3-action_upright-anchor_flow-fix-post-encoder-one-flow-head_joint2global-pzY-sample_anchor2action2global-opt2-pzX-sample_mod_easy_rack
python train_residual_flow_multimodal.py --config-path=../configs/rpdiff_data --config-name=joint_train_pzX-dgcnn-transformer_pzY-pn2_hybridlatentz-global_gradclip1e-4_se3-action_upright-anchor_flow-fix-post-encoder-one-flow-head_joint2global-pzY-sample_anchor2action2global-opt2-pzX-sample_mod_easy_rack dataset_root=TODO test_dataset_root=TODO log_dir=TODO rpdiff_descriptions_path=TODO
The ICRA submission model configurations can be run on the multimodal_icra2024 branch.
@article{wang2024taxposed,
title={Learning Distributional Demonstration Spaces for Task-Specific Cross-Pose Estimation},
author={Wang, Jenny and Donca, Octavian and Held, David},
journal={IEEE International Conference on Robotics and Automation (ICRA), 2024},
year={2024}
}

