Robust human-computer interaction for unstable camera systems

1Citations
Citations of this article
6Readers
Mendeley users who have this article in their library.

Abstract

A lot of vision systems have been embedded in devices around us, like mobile phones, vehicles and UAVs. Many of them still need interactive operations of human users. However, specifying accurate object information could be a challenging task due to video jitters caused by camera shakes and target motions. In this paper, we first collect practical hand drawn bounding boxes on real-life videos which are captured by hand-held cameras and UAV-based cameras. We give a deep look into human-computer interactive operations on unstable images. The collected data shows that human input suffers heavy deviations which are harmful to interaction accuracy. To achieve robust interactions on unstable platforms, we propose a target-focused video stabilization method which utilizes a proposal-based object detector and a tracking-based motion estimation component. This method starts with a single manual click and outputs stabilized video stream in which the specified target stays almost stationary. Our method removes not only camera jitters but also target motions simultaneously, therefore offering an comfortable environment for users to do further interactive operations. The experiments demonstrate that the proposed method effectively eliminates image vibrations and significantly increases human input accuracy.

Cite

CITATION STYLE

APA

Zhu, H., You, Q., & Chen, W. (2018). Robust human-computer interaction for unstable camera systems. In IEICE Transactions on Information and Systems (Vol. E101D, pp. 1915–1923). Institute of Electronics, Information and Communication, Engineers, IEICE. https://doi.org/10.1587/transinf.2017EDP7248

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free