Direkt zum Inhalt springen
Computer Vision Group
TUM School of Computation, Information and Technology
Technical University of Munich

Technical University of Munich

Menu

Links

Informatik IX
Computer Vision Group

Boltzmannstrasse 3
85748 Garching info@vision.in.tum.de

Follow us on:

News

04.03.2024

We have twelve papers accepted to CVPR 2024. Check our publication page for more details.

18.07.2023

We have four papers accepted to ICCV 2023. Check out our publication page for more details.

02.03.2023

CVPR 2023

We have six papers accepted to CVPR 2023. Check out our publication page for more details.

15.10.2022

NeurIPS 2022

We have two papers accepted to NeurIPS 2022. Check out our publication page for more details.

15.10.2022

WACV 2023

We have two papers accepted at WACV 2023. Check out our publication page for more details.

More


Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revision Previous revision
Next revision
Previous revision
research:vslam:dh3d [2020/07/30 20:04]
Rui Wang
research:vslam:dh3d [2020/11/11 20:24] (current)
Rui Wang
Line 1: Line 1:
 +~~NOCACHE~~
 ====== DH3D: Deep Hierarchical 3D Descriptors for Robust Large-Scale 6DoF Relocalization ====== ====== DH3D: Deep Hierarchical 3D Descriptors for Robust Large-Scale 6DoF Relocalization ======
 **Contact:** [[members:wangr|Rui Wang]], [[members:cremers|Prof. Daniel Cremers]] **Contact:** [[members:wangr|Rui Wang]], [[members:cremers|Prof. Daniel Cremers]]
  
-This page is still under construction.+{{ :research:vslam:dh3d:overview.png?600&nolink }}
  
 ===== Abstract ===== ===== Abstract =====
 For relocalization in large-scale point clouds, we propose the first approach that unifies global place recognition and local 6DoF pose refinement. To this end, we design a Siamese network that jointly learns 3D local feature detection and description directly from raw 3D points. It integrates FlexConv and Squeeze-and-Excitation (SE) to assure that the learned local descriptor captures multi-level geometric information and channel-wise relations. For detecting 3D keypoints we predict the discriminativeness of the local descriptors in an unsupervised manner. We generate the global descriptor by directly aggregating the learned local descriptors with an effective attention mechanism. In this way, local and global 3D descriptors are inferred in one single forward pass. Experiments on various benchmarks demonstrate that our method achieves competitive results for both global point cloud retrieval and for local point cloud registration in comparison to state-of-the-art approaches. To validate the generalizability and robustness of our 3D keypoints, we demonstrate that our method also performs favorably without fine-tuning on registration of point clouds that were generated by a visual SLAM system. For relocalization in large-scale point clouds, we propose the first approach that unifies global place recognition and local 6DoF pose refinement. To this end, we design a Siamese network that jointly learns 3D local feature detection and description directly from raw 3D points. It integrates FlexConv and Squeeze-and-Excitation (SE) to assure that the learned local descriptor captures multi-level geometric information and channel-wise relations. For detecting 3D keypoints we predict the discriminativeness of the local descriptors in an unsupervised manner. We generate the global descriptor by directly aggregating the learned local descriptors with an effective attention mechanism. In this way, local and global 3D descriptors are inferred in one single forward pass. Experiments on various benchmarks demonstrate that our method achieves competitive results for both global point cloud retrieval and for local point cloud registration in comparison to state-of-the-art approaches. To validate the generalizability and robustness of our 3D keypoints, we demonstrate that our method also performs favorably without fine-tuning on registration of point clouds that were generated by a visual SLAM system.
  
-===== Paper Summary ===== +===== Update ===== 
-The video is with audio. +  * [11.11.2020] For the ease of comparison, we upload the numbers used to draw the plots in the paper ({{:research:vslam:dh3d:DH3D_results_in_paper.txt|download}}).
-<html><center><iframe width="640" height="360" src="https://www.youtube.com/embed/1oYtqJhHOLw" frameborder="0" allowfullscreen></iframe></center></html> \\+
  
 ===== ECCV Spotlight Presentation ===== ===== ECCV Spotlight Presentation =====
Line 15: Line 15:
 <html><center><iframe width="640" height="360" src="https://www.youtube.com/embed/ZxZiwZugG14" frameborder="0" allowfullscreen></iframe></center></html> <html><center><iframe width="640" height="360" src="https://www.youtube.com/embed/ZxZiwZugG14" frameborder="0" allowfullscreen></iframe></center></html>
  
-===== Download =====+===== Citation ===== 
 +If you find our work useful in your research, please consider citing: 
 +<code> 
 +@inproceedings{du2020dh3d, 
 +    title={DH3D: Deep Hierarchical 3D Descriptors for Robust Large-Scale 6DoF Relocalization}, 
 +    author={Du, Juan and Wang, Rui and Cremers, Daniel}, 
 +    booktitle={European Conference on Computer Vision (ECCV)}, 
 +    year={2020} 
 +
 +</code> 
 + 
 +===== Code ===== 
 +Code and the pre-trained models can be accessed from the [[https://github.com/JuanDuGit/DH3D|GitHub page]]. 
 + 
 +===== Datasets ===== 
 +Our model is mainly trained and tested on the LiDAR point clouds from the [[https://robotcar-dataset.robots.ox.ac.uk/|Oxford RobotCar dataset]]. To test the generalization capability, two extra datasets are used, namely [[https://projects.asl.ethz.ch/datasets/doku.php?id=laserregistration:laserregistration|ETH]] (LiDAR point clouds from two sequences, gazebo_winter and wood_autumn) and Oxford RobotCar Stereo DSO (point clouds generated by running [[https://vision.in.tum.de/research/vslam/stereo-dso|Stereo DSO]]). As our method can be used for global place recognition (retrieval) and local pose refinement (regression), the corresponding datasets are denoted by "global" and "local" respectively. For more details on how the datasets are generated, please refer to the beginning of Section 4 in the main paper and Section 4.2 and 4.4 in the supplementary material. For examples on how to train or test our model on these datasets, please refer to the [[https://github.com/JuanDuGit/DH3D|GitHub page]]. 
 +  
 +<php> 
 +readfile("/usr/demos/cvpr/dh3d/tab.html"); 
 +</php> 
 + 
 +===== Qualitative Results ===== 
 +{{ :research:vslam:dh3d:oxford.png?700&nolink |Resuts on LiDAR points of Oxford RobotCar}} 
 + 
 +{{ :research:vslam:dh3d:oxford_stereo_dso1.png?700&nolink |}} 
 +{{ :research:vslam:dh3d:oxford_stereo_dso2.png?700&nolink |}} 
 + 
 +===== Other Materials =====
   * Paper: {{:research:vslam:dh3d:du2020dh3d.pdf|download}}, supplementary material: {{:research:vslam:dh3d:du2020dh3d-supp.pdf|download}}   * Paper: {{:research:vslam:dh3d:du2020dh3d.pdf|download}}, supplementary material: {{:research:vslam:dh3d:du2020dh3d-supp.pdf|download}}
   * arXiv: [[https://arxiv.org/abs/2007.09217|link]]   * arXiv: [[https://arxiv.org/abs/2007.09217|link]]
-  * Code: Will be online before the start of ECCV 2020. 
   * Paper summary slides: {{:research:vslam:dh3d:2754_short_video_slides.pdf|download}}   * Paper summary slides: {{:research:vslam:dh3d:2754_short_video_slides.pdf|download}}
 +  * Paper summary video: [[https://youtu.be/1oYtqJhHOLw|Youtube]]
   * ECCV spotlight presentation slides: {{:research:vslam:dh3d:ECCV2020_slides_long.pdf|download}}   * ECCV spotlight presentation slides: {{:research:vslam:dh3d:ECCV2020_slides_long.pdf|download}}
 +  * The numbers used to draw the plots (Fig.4 and Fig. 6) in the paper: {{:research:vslam:dh3d:DH3D_results_in_paper.txt|download}}.
  
 ==== Publications ==== ==== Publications ====

Rechte Seite

Informatik IX
Computer Vision Group

Boltzmannstrasse 3
85748 Garching info@vision.in.tum.de

Follow us on:

News

04.03.2024

We have twelve papers accepted to CVPR 2024. Check our publication page for more details.

18.07.2023

We have four papers accepted to ICCV 2023. Check out our publication page for more details.

02.03.2023

CVPR 2023

We have six papers accepted to CVPR 2023. Check out our publication page for more details.

15.10.2022

NeurIPS 2022

We have two papers accepted to NeurIPS 2022. Check out our publication page for more details.

15.10.2022

WACV 2023

We have two papers accepted at WACV 2023. Check out our publication page for more details.

More