Skip to content

Latest commit

 

History

History
127 lines (83 loc) · 6.32 KB

README.md

File metadata and controls

127 lines (83 loc) · 6.32 KB

TransCG: A Large-Scale Real-World Dataset for Transparent Object Depth Completion and Grasping

PWC CC BY-NC-SA 4.0

[Paper] [Project Page]

Authors: Hongjie Fang, Hao-Shu Fang, Sheng Xu, Cewu Lu.

Welcome to the official repository for the TransCG paper. This repository includes the dataset and the proposed Depth Filler Net (DFNet) models.

TransCG Dataset

TransCG dataset is now available on official page. TransCG dataset is the first large-scale real-world dataset for transparent object depth completion and grasping. In total, our dataset contains 57,715 RGB-D images of 51 transparent objects and many opaque objects captured from different perspectives of 130 scenes under various real-world settings. The 3D mesh model of the transparent objects are also provided in our dataset.

Daily Transparent Objects in Dataset Real-time Tracking SystemRobot Collection

Requirements

The code has been tested under

  • Ubuntu 18.04 + NVIDIA GeForce RTX 3090 (CUDA 11.1)
  • PyTorch 1.9.0

System dependencies can be installed by:

sudo apt-get install libhdf5-10 libhdf5-serial-dev libhdf5-dev libhdf5-cpp-11
sudo apt install libopenexr-dev zlib1g-dev openexr

Other dependencies can be installed by

pip install -r requirements.txt

Run

Quick Start

NOTE. The following checkpoint is compatible with this version. We will update the checkpoint of the latest version later.

Our pretrained checkpoint is available on Google Drive or Baidu Netdisk (Code: c01g). The checkpoint is trained with the default configuration in the configs folder. You can use our released checkpoints for inference or testing. Refer to assets/docs/DFNet.md for details about the depth completion network.

Grasping Demo

To verify the depth completion results in robotic manipulation, we select the fundamental object grasping as the downstream task of our network. Here are the grasping demos. Refer to assets/docs/grasping.md for details about the grasping method.

Configuration

You need to create a configuration file for training, testing and inference. See assets/docs/configuration.md for details.

Dataset Preparation

Inference

For inference stage, there is a Inferencer class in inference.py, you can directly call it for inference.

Example. Given an H x W x 3 RGB image rgb, and an H x W depth image depth (after scaling according to camera parameters), you can use the following code to get the refined depth according to our models.

from inferencer import Inferencer
# Initialize the inferencer. It is recommended to intiailize before starting your task for real-time performance.
inferencer = Inferencer(cfg_file = 'configs/inference.yaml') # Specify your configuration file here.
# Call inferencer for refined depth
refine_depth = inferencer.inference(rgb, depth)

For full code sample, refer to sample_inference.py.

Training (Optional)

For training from scrach, you need to create a configuration file following instruction of configuration section. Then, execute the following commands to train your own model.

python train.py --cfg [Configuration File]

If you want to fine-tune your model from some checkpoints, you may need to provide resume_lr in configuration file. See assets/docs/configuration.md for details.

Testing (Optional)

For model testing, you also need to create a configuration file following instruction of configuration section. Then, execute the following commands to test the model.

python test.py --cfg [Configuration File]

Note. For testing stage, the checkpoint specified in the configuration file should exist.

Citation

@ARTICLE{fang2022transcg,
    author  = {Fang, Hongjie and Fang, Hao-Shu and Xu, Sheng and Lu, Cewu},
    journal = {IEEE Robotics and Automation Letters}, 
    title   = {TransCG: A Large-Scale Real-World Dataset for Transparent Object Depth Completion and a Grasping Baseline}, 
    year    = {2022},
    volume  = {7},
    number  = {3},
    pages   = {7383-7390},
    doi     = {10.1109/LRA.2022.3183256}
}

License

This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License.

CC BY-NC-SA 4.0