A New Localization Objective for Accurate Fine-Grained Affordance Segmentation under High-Scale Variations

8Citations
Citations of this article
11Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

Fine-grained affordance segmentation for object parts can greatly benefit robotics and scene understanding applications. In this work, we propose an instance-segmentation framework that can accurately localize functionality and affordance of individual object parts. We build on the standard Mask-RCNN framework and propose two novelties to the localization objective that can lead to improved part detection and affordance segmentation results. Specifically, we notice two problems with the conventional IOU based regression loss, (a) the small boxes, that are specially relevant for fine-grained detection, have a higher risk of being ignored during the optimization process and (b) a constant value of IOU for non-overlapping candidates means no supervision is available to encourage the reduction in loss function. To address these limitations, we propose a novel Angular Intersection Over Larger (AIOL) measure. Our experiments show consistent improvement over other baselines and state of the art localization loss functions for the fine-grained affordance segmentation task.

Cite

CITATION STYLE

APA

Hassanin, M., Khan, S., & Tahtali, M. (2020). A New Localization Objective for Accurate Fine-Grained Affordance Segmentation under High-Scale Variations. IEEE Access, 8, 28123–28132. https://doi.org/10.1109/ACCESS.2019.2958608

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free