• Stars
    star
    600
  • Rank 74,640 (Top 2 %)
  • Language
    C++
  • License
    Other
  • Created about 4 years ago
  • Updated about 1 year ago

Reviews

There are no reviews yet. Be the first to send feedback to the community and the maintainers!

Repository Details

[IROS 2021] BundleTrack: 6D Pose Tracking for Novel Objects without Instance or Category-Level 3D Models

This is the official implementation of our paper:

BundleTrack: 6D Pose Tracking for Novel Objects without Instance or Category-Level 3D Models

accepted in International Conference on Intelligent Robots and Systems (IROS) 2021.

Abstract

Most prior 6D object pose tracking often assume that the target object's CAD model, at least at a category-level, is available for offline training or during online template matching. This work proposes BundleTrack, a general framework for 6D pose tracking of novel objects, which does not depend upon 3D models, either at the instance or category-level. It leverages the complementary attributes of recent advances in deep learning for segmentation and robust feature extraction, as well as memory-augmented pose graph optimization for spatiotemporal consistency. This enables long-term, low-drift tracking under various challenging scenarios, including significant occlusions and object motions. Comprehensive experiments given two public benchmarks demonstrate that the proposed approach significantly outperforms state-of-art, category-level 6D tracking or dynamic SLAM methods. When compared against state-of-art methods that rely on an object instance CAD model, comparable performance is achieved, despite the proposed method's reduced information requirements. An efficient implementation in CUDA provides a real-time performance of 10Hz for the entire framework.

This repo can be readily applied to 6D pose tracking for novel unknown objects. For CAD model-based 6D pose tracking, please check out my another repository of se(3)-TrackNet

Bibtex

@inproceedings{wen2021bundletrack,
  title={BundleTrack: 6D Pose Tracking for Novel Objects without Instance or Category-Level 3D Models},
  author={Wen, B and Bekris, Kostas E},
  booktitle={IEEE/RSJ International Conference on Intelligent Robots and Systems},
  year={2021}
}

Supplementary Video

Click to watch

IROS 2021 Presentation

Click to watch

Results

Benchmark Output Results

For convenience of benchmarking and making plots, results of pose outputs can be downloaded below

Setup

For the environment setup, it's strongly recommended to use our provided docker environment (setting up from scratch is very complicated and not supported in this repo). For this, you don't have to know how docker works. Only some basic commands are needed and will be provided in the below steps.

  • Install docker (https://docs.docker.com/get-docker/).

  • Run

    docker pull wenbowen123/bundletrack:latest
    docker pull wenbowen123/lf-net-release-env:latest
    
  • Edit the docker/run_container.sh, update the paths of BUNDLETRACK_DIR, NOCS_DIR and YCBINEOAT_DIR

  • Run bash docker/run_container.sh

  • cd [PATH_TO_BUNDLETRACK]

  • rm -rf build && mkdir build && cd build && cmake .. && make

Data

Depending on what you want to run, download those data that are neccessary.

Run predictions on NOCS

  • Open a separate terminal and run

    bash lf-net-release/docker/run_container.sh
    cd [PATH_TO_BUNDLETRACK]
    cd lf-net-release && python run_server.py
    
  • Go back to the terminal where you launched the bundletrack docker in above and run below. The output will be saved to debug_dir specified in config file. By default it's /tmp/BundleTrack/. For more detailed logs, change LOG to 2 or higher in config_nocs.yml.

    python scripts/run_nocs.py --nocs_dir [PATH_TO_NOCS] --scene_id 1 --port 5555 --model_name can_arizona_tea_norm
    
  • Finally, the results will be saved in /tmp/BundleTrack/

  • For evaluating on the entire NOCS Dataset, run (NOTE that this will add noise to perturb the initial ground-truth pose for evaluation as explained in the paper. If you want to see how BundleTrack actually performs, run the above section)

    python scripts/eval_nocs.py --nocs_dir [PAHT TO NOCS]  --results_dir [PATH TO THE RUNNING OUTPUTS]
    

Run predictions on YCBInEOAT

  • Change the model_name and model_dir in config_ycbineoat.yml to the path to the .obj file (e.g. For folder bleach0, the model_name is 021_bleach_cleanser, and model_dir is [Your path to YCB Objects]/021_bleach_cleanser/textured_simple.obj)

  • Open a separate terminal and run

    bash lf-net-release/docker/run_container.sh
    cd [PATH_TO_BUNDLETRACK]
    cd lf-net-release && python run_server.py
    
  • Go back to the terminal where you launched the bundletrack docker in above, and run below. The output will be saved to debug_dir specified in config file. By default it's /tmp/BundleTrack/

    python scripts/run_ycbineoat.py --data_dir [PATH_TO_YCBInEOAT] --port 5555 --model_name [The YCB object's name, e.g. 021_bleach_cleanser]
    
  • Finally, the results will be saved in /tmp/BundleTrack/. For more detailed logs, change LOG to 2 or higher in config_ycbineoat.yml.

  • For evaluating on the entire YCBInEOAT Dataset, run

    python scripts/eval_nocs.py --ycbineoat_dir [PAHT TO YCBINEOAT] --ycb_model_dir [YCB MODELS FOLDER] --results_dir [PATH TO THE RUN OUTPUTS]
    

Run predictions on your own RGBD data

  • Download YCBInEOAT, if you haven't done so in above.

  • Open a separate terminal and run

    bash lf-net-release/docker/run_container.sh
    cd [PATH_TO_BUNDLETRACK]
    cd lf-net-release && python run_server.py
    
  • Prepare segmentation masks. In YCBInEOAT Dataset, we computed masks from robotic arm forward kinematics. If your scene is not too complicated similar to NOCS Dataset, you can run the video segmentation network to get masks as below:

    • First you need to prepare an initial mask (grayscale image, where 0 means background, else foreground).

    • python transductive-vos.pytorch/run_video.py --img_dir [THE PATH TO COLOR FILES] --init_mask_file [THE INITIAL MASK FILE YOU PREPARED ABOVE] --mask_save_dir [WHERE TO SAVE]

    • Prepare your folder structure same as any folder (e.g. "mustard_easy_00_02") in YCBInEOAT Dataset. Put it under the same directory in YCBInEOAT, i.e. next to "mustard_easy_00_02". Then edit config_ycbineoat.yml to make sure the paths at top are right.

      Structure:

      mustard_easy_00_02
      β”œβ”€β”€ rgb
      β”œβ”€β”€ masks
      β”œβ”€β”€ depth
      └── cam_K.txt
      
  • Go back to the terminal where you launched the bundletrack docker, run below. The output will be saved to debug_dir specified in config file. By default it's /tmp/BundleTrack/

    python scripts/run_ycbineoat.py --data_dir [PATH TO YOUR FOLDER ABOVE] --port 5555