Ruben Martinez-Cantin

45 Followers
46 Following
14 Posts
Professor in System Engineering. Researcher in Robotics and Machine Learning at @unizar Advisor at @SigOpt Most tweets in English, few in Spanish
Webhttp://webdiis.unizar.es/~rmcantin/

Multi-label affordance mapping from egocentric vision
L. Mur-Labadia, J.J. Guerrero, R. Martinez-Cantin

Paper: https://arxiv.org/abs/2309.02120

Github (code/dataset): https://github.com/lmur98/epic_kitchens_affordances

(7/7)

Multi-label affordance mapping from egocentric vision

Accurate affordance detection and segmentation with pixel precision is an important piece in many complex systems based on interactions, such as robots and assitive devices. We present a new approach to affordance perception which enables accurate multi-label segmentation. Our approach can be used to automatically extract grounded affordances from first person videos of interactions using a 3D map of the environment providing pixel level precision for the affordance location. We use this method to build the largest and most complete dataset on affordances based on the EPIC-Kitchen dataset, EPIC-Aff, which provides interaction-grounded, multi-label, metric and spatial affordance annotations. Then, we propose a new approach to affordance segmentation based on multi-label detection which enables multiple affordances to co-exists in the same space, for example if they are associated with the same object. We present several strategies of multi-label detection using several segmentation architectures. The experimental results highlight the importance of the multi-label detection. Finally, we show how our metric representation can be exploited for build a map of interaction hotspots in spatial action-centric zones and use that representation to perform a task-oriented navigation.

arXiv.org
-Second, we show how affordances can be used for task-oriented navigation. (6/7)
We show the potential of our dataset and pipeline in two scenarios:
-First, we build affordance maps of the environment with interaction hotspots that enable high level reasoning for robots and assistive devices. (5/7)
Contrary to other segmentation problems, there might be multiple affordances associated with the same object/pixel. Therefore, we have extended and evaluated several segmentation algorithms in the multi-label scenario. (4/7)
We present the EPIC-Aff, the largest and first multi-label affordance segmentation dataset. (2/7)
I’m also glad to share our #ICCV2023 paper on automatic affordance mapping with multi-label segmentation. Affordances are fundamental for any interactive system (robotics, assistive devices, etc.) Affordances are the potential actions that the environment offers. (1/7)

LightDepth: Single-View Depth Self-Supervision from Illumination Decline
J.R. Puigvert, V.M. Batlle, J.M.M. Montiel, R. Martinez-Cantin, P. Fua, J.D. Tardós, J. Civera

Web: https://sites.google.com/unizar.es/lightdepth
Paper: https://arxiv.org/abs/2308.10525

(7/7)

Inicio

LightDepth: Single-View Depth Self-Supervision from Illumination Decline J. Rodríguez-Puigvert*, V.M. Batlle*, J.M.M. Montiel, R. Martinez-Cantin, P. Fua, J.D. Tardós, J. Civera ICCV 2023 [Paper] [Supp]

We literally use light intensity as depth supervision! That’s what I called a bright idea! (sorry for the dad joke 😅). (6/7)
We are able to achieve performance close to fully supervised learning, without using any label. Because we are using a self-supervised loss, we are also able to perform test-time refinement (TTR). (5/7)
In this work, we train a neural network to predict the depth and albedo of the image. Then, we compute the surface normals from the depth map which allows us to use a differentiable rendering equation as a self-supervised loss. (4/7)