Abstract
Object segmentation (OS) technology is a research hotspot in computer vision, and it has a wide range of applications in many fields. Cognitive vision studies have shown that human vision is highly sensitive to both global information and local details in scenes. To this end, we design a novel, efficient, and easy-to-use enhanced-alignment measure (Eξ) for evaluating the performance of the OS model. Eξ combines local pixel values with the image-level mean value, jointly evaluates the image-/pixel-level similarity between a segmentation result and a ground-truth (GT) result. Extensive experiments on the four popular benchmarks via five metameasures, i.e., application ranking, demoting generic, denying noise, human ranking, and recognizing GT, show significant relative improvement compared with existing widely-adopted evaluation metrics such as IoU and Fβ. By using the weighted binary cross-entropy loss, the enhanced-alignment loss, and the weighted IoU loss, we further design a hybrid loss function (Hybrid-Eloss) to guide the network to learn pixel-, object- and image-level features. Qualitative and quantitative results show further improvement in terms of accuracy when using our hybrid loss function in three different OS tasks.
Author supplied keywords
Cite
CITATION STYLE
Fan, D. P., Ji, G. P., Qin, X., & Cheng, M. M. (2021). Cognitive vision inspired object segmentation metric and loss function. Scientia Sinica Informationis, 51(9), 1475–1489. https://doi.org/10.1360/SSI-2020-0370
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.