If you missed our demo at #3DV2024 , our setup will be presented at the #ICRA mobile manipulation workshop and we just finished to cut it into a nice video:

Spot-Compose: A Framework for Open-Vocabulary Object Retrieval and Drawer Manipulation in Point Clouds
๐Ÿ“„ Paper: https://arxiv.org/abs/2404.12440
๐Ÿ”— Webpage: https://spot-compose.github.io/
๐Ÿ“น Video: https://youtu.be/4BRKIzyaT3o?si=XB9SF4aTFXbBEv7f

#robotics #BostonDynamcis #spot #eth_zurich @cvg #computervision

Spot-Compose: A Framework for Open-Vocabulary Object Retrieval and Drawer Manipulation in Point Clouds

In recent years, modern techniques in deep learning and large-scale datasets have led to impressive progress in 3D instance segmentation, grasp pose estimation, and robotics. This allows for accurate detection directly in 3D scenes, object- and environment-aware grasp prediction, as well as robust and repeatable robotic manipulation. This work aims to integrate these recent methods into a comprehensive framework for robotic interaction and manipulation in human-centric environments. Specifically, we leverage 3D reconstructions from a commodity 3D scanner for open-vocabulary instance segmentation, alongside grasp pose estimation, to demonstrate dynamic picking of objects, and opening of drawers. We show the performance and robustness of our model in two sets of real-world experiments including dynamic object retrieval and drawer opening, reporting a 51% and 82% success rate respectively. Code of our framework as well as videos are available on: https://spot-compose.github.io/.

arXiv.org

So happy to finally be presenting this work with @[email protected] et al.

In SUCRe, we use SfM/MVS dense reconstruction to recover colors from underwater images.

This work, funded by
@ifremer_fr will be presented at #3DV2024 in March.

#underwater #deepsea #ai #machinelearning #computervision

Today we release the code for LabelMaker, a tool to automatically create semantic annotation for indoor scenes. We produce labels in 2D and 3D by lifting an ensemble of SOTA models in a common label space. LabelMaker will be presented as spotlight at #3DV2024 in March.

๐Ÿ”— https://labelmaker.org
๐Ÿ› ๏ธ https://github.com/cvg/LabelMaker
๐Ÿ“„ https://arxiv.org/abs/2311.12174

#computervision #robotics #3DV

LabelMaker ๐ŸŽจ

LabelMaker