Bruno Sauvalle, Arnaud de La Fortelle
We introduce a new architecture for unsupervised object-centric representation learning and multi-object detection and segmentation, which uses a translation-equivariant attention mechanism to predict the coordinates of the objects present in the scene and to associate a feature vector to each object. A transformer encoder handles occlusions and redundant detections, and a convolutional autoencoder is in charge of background reconstruction. We show that this architecture significantly outperforms the state of the art on complex synthetic benchmarks.
| Task | Dataset | Metric | Value | Model |
|---|---|---|---|---|
| Instance Segmentation | ShapeStacks | ARI-FG | 0.82 | AST |
| Instance Segmentation | ObjectsRoom | ARI-FG | 0.87 | AST |
| Unsupervised Object Segmentation | ShapeStacks | ARI-FG | 0.82 | AST |
| Unsupervised Object Segmentation | ObjectsRoom | ARI-FG | 0.87 | AST |