<?xml version="1.0" encoding="UTF-8"?>
<collection>
<dc:dc xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:invenio="http://invenio-software.org/elements/1.0" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://www.openarchives.org/OAI/2.0/oai_dc/ http://www.openarchives.org/OAI/2.0/oai_dc.xsd"><dc:identifier>doi:10.1109/ICCV51070.2023.00483</dc:identifier><dc:language>eng</dc:language><dc:creator>Mur-Labadia, Lorenzo</dc:creator><dc:creator>Guerrero, José J.</dc:creator><dc:creator>Martínez-Cantín, Rubén</dc:creator><dc:title>Multi-label affordance mapping from egocentric vision</dc:title><dc:identifier>ART-2024-137283</dc:identifier><dc:description>Accurate affordance detection and segmentation with pixel precision is an important piece in many complex systems based on interactions, such as robots and assitive devices. We present a new approach to affordance perception which enables accurate multi-label segmentation. Our approach can be used to automatically extract grounded affordances from first person videos of interactions using a 3D map of the environment providing pixel level precision for the affordance location. We use this method to build the largest and most complete dataset on affordances based on the EPIC-Kitchen dataset, EPIC-Aff, which provides interaction-grounded, multi-label, metric and spatial affordance annotations. Then, we propose a new approach to affordance segmentation based on multi-label detection which enables multiple affordances to co-exists in the same space, for example if they are associated with the same object. We present several strategies of multi-label detection using several segmentation architectures. The experimental results highlight the importance of the multi-label detection. Finally, we show how our metric representation can be exploited for build a map of interaction hotspots in spatial action-centric zones and use that representation to perform a task-oriented navigation.</dc:description><dc:date>2024</dc:date><dc:source>http://zaguan.unizar.es/record/132130</dc:source><dc:doi>10.1109/ICCV51070.2023.00483</dc:doi><dc:identifier>http://zaguan.unizar.es/record/132130</dc:identifier><dc:identifier>oai:zaguan.unizar.es:132130</dc:identifier><dc:relation>info:eu-repo/grantAgreement/ES/DGA/T45-23R</dc:relation><dc:relation>info:eu-repo/grantAgreement/ES/MICINN-AEI/PID2021-125209OB-I00</dc:relation><dc:relation>info:eu-repo/grantAgreement/EUR/MICINN/TED2021-129410B-I00</dc:relation><dc:relation>info:eu-repo/grantAgreement/EUR/MICINN/TED2021-131150B-I00</dc:relation><dc:identifier.citation>Proceedings (IEEE International Conference on Computer Vision) 2023 (2024), 5215-5226</dc:identifier.citation><dc:rights>All rights reserved</dc:rights><dc:rights>http://www.europeana.eu/rights/rr-f/</dc:rights><dc:rights>info:eu-repo/semantics/openAccess</dc:rights></dc:dc>

</collection>