Learning to Generate Realistic LiDAR Point Clouds

  • \(^1\) University of Illinois at Urbana-Champaign

ECCV 2022

overview

Abstract

We present LiDARGen, a novel, effective, and controllable generative model that produces realistic LiDAR point cloud sensory readings. Our method leverages the powerful score-matching energy-based model and formulates the point cloud generation process as a stochastic denoising process in the equirectangular view. This model allows us to sample diverse and high-quality point cloud samples with guaranteed physical feasibility and controllability. We validate the effectiveness of our method on the challenging KITTI-360 and NuScenes datasets. The quantitative and qualitative results show that our approach produces more realistic samples than other generative models. Furthermore, LiDARGen can sample point clouds conditioned on inputs without retraining. We demonstrate that our proposed generative model could be directly used to densify LiDAR point clouds.

overview

Video

Sample Generation Process from Different Angles

Unconditional Samples

Related links

LiDARGen is implemented on top of the codebase of NCSNv2.

Citation

Acknowledgements

We thank Wei-Chiu Ma and Zhijian Liu for their feedback on early drafts and all the participants in the human perceptual quality study. The project is partially funded by the Illinois Smart Transportation Initiative STII-21-07. We also thank Nvidia for the Academic Hardware Grant.
The website template was borrowed from Michaƫl Gharbi and NeurMips.