Multimodal Deep Feature Fusion (MMDFF) for RGB-D Tracking

4Citations
Citations of this article
6Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

Visual tracking is still a challenging task due to occlusion, appearance changes, complex motion, etc. We propose a novel RGB-D tracker based on multimodal deep feature fusion (MMDFF) in this paper. MMDFF model consists of four deep Convolutional Neural Networks (CNNs): Motion-specific CNN, RGB- specific CNN, Depth-specific CNN, and RGB-Depth correlated CNN. The depth image is encoded into three channels which are sent into depth-specific CNN to extract deep depth features. The optical flow image is calculated for every frame and then is fed to motion-specific CNN to learn deep motion features. Deep RGB, depth, and motion information can be effectively fused at multiple layers via MMDFF model. Finally, multimodal fusion deep features are sent into the C-COT tracker to obtain the tracking result. For evaluation, experiments are conducted on two recent large-scale RGB-D datasets and results demonstrate that our proposed RGB-D tracking method achieves better performance than other state-of-art RGB-D trackers.

References Powered by Scopus

Caffe: Convolutional architecture for fast feature embedding

8231Citations
N/AReaders
Get full text

FlowNet: Learning optical flow with convolutional networks

3524Citations
N/AReaders
Get full text

Fully-convolutional siamese networks for object tracking

3499Citations
N/AReaders
Get full text

Cited by Powered by Scopus

Visual Object Tracking in RGB-D Data via Genetic Feature Learning

8Citations
N/AReaders
Get full text

Knowledge Synergy Learning for Multi-Modal Tracking

7Citations
N/AReaders
Get full text

Research on a Microexpression Recognition Technology Based on Multimodal Fusion

4Citations
N/AReaders
Get full text

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Jiang, M. X., Deng, C., Zhang, M. M., Shan, J. S., & Zhang, H. (2018). Multimodal Deep Feature Fusion (MMDFF) for RGB-D Tracking. Complexity, 2018. https://doi.org/10.1155/2018/5676095

Readers' Seniority

Tooltip

PhD / Post grad / Masters / Doc 4

100%

Readers' Discipline

Tooltip

Computer Science 2

50%

Economics, Econometrics and Finance 1

25%

Engineering 1

25%

Save time finding and organizing research with Mendeley

Sign up for free