Deep Generative Modeling of LiDAR Data
| Authors |
|
|---|---|
| Publication date | 2019 |
| Book title | 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) |
| Book subtitle | Macau, China, 3-8 November 2019 |
| ISBN |
|
| ISBN (electronic) |
|
| Event | 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems |
| Pages (from-to) | 5034-5040 |
| Publisher | [Piscataway, NJ]: IEEE |
| Organisations |
|
| Abstract |
Building models capable of generating structured output is a key challenge for AI and robotics. While generative models have been explored on many types of data, little work has been done on synthesizing lidar scans, which play a key role in robot mapping and localization. In this work, we show that one can adapt deep generative models for this task by unravelling lidar scans into a 2D point map. Our approach can generate high quality samples, while simultaneously learning a meaningful latent representation of the data. We demonstrate significant improvements against state-of-the-art point cloud generation methods. Furthermore, we propose a novel data representation that augments the 2D signal with absolute positional information. We show that this helps robustness to noisy and imputed input; the learned model can recover the underlying lidar scan from seemingly uninformative data.
|
| Document type | Conference contribution |
| Note | In print edition: p. 4073-4079. |
| Language | English |
| Published at | https://doi.org/10.1109/IROS40897.2019.8968535 |
| Other links | http://www.proceedings.com/52283.html |
| Downloads |
08968535
(Final published version)
|
| Permalink to this page | |
