SE(3)-Transformers: 3D Roto-Translation Equivariant Attention Networks
| Authors |
|
|---|---|
| Publication date | 2021 |
| Host editors |
|
| Book title | 34th Concerence on Neural Information Processing Systems (NeurIPS 2020) |
| Book subtitle | online, 6-12 December 2020 |
| ISBN |
|
| Series | Advances in Neural Information Processing Systems |
| Event | Advances in Neural Information Processing Systems 2020 |
| Volume | Issue number | 3 |
| Pages (from-to) | 1970-1981 |
| Publisher | San Diego, CA: Neural Information Processing Systems Foundation |
| Organisations |
|
| Abstract |
We introduce the SE(3)-Transformer, a variant of the self-attention module for 3D point-clouds, which is equivariant under continuous 3D roto-translations. Equivariance is important to ensure stable and predictable performance in the presence of nuisance transformations of the data input. A positive corollary of equivariance is increased weight-tying within the model. The SE(3)-Transformer leverages the benefits of self-attention to operate on large point clouds with varying number of points, while guaranteeing SE(3)-equivariance for robustness. We evaluate our model on a toy N-body particle simulation dataset, showcasing the robustness of the predictions under rotations of the input. We further achieve competitive performance on two real-world datasets, ScanObjectNN and QM9. In all cases, our model outperforms a strong, non-equivariant attention baseline and an equivariant model without attention.
|
| Document type | Conference contribution |
| Note | With supplemental file |
| Language | English |
| Published at | https://papers.nips.cc/paper/2020/hash/15231a7ce4ba789d13b722cc5c955834-Abstract.html |
| Other links | https://www.proceedings.com/59066.html |
| Downloads |
NeurIPS-2020-se3-transformers-3d-roto-translation-equivariant-attention-networks-Paper
(Accepted author manuscript)
|
| Supplementary materials | |
| Permalink to this page | |
