Browse by author
Lookup NU author(s): Dr Yang Long
This is the authors' accepted manuscript of an article that has been published in its final definitive form by IEEE, 2018.
For re-use rights please refer to the publisher's terms and conditions.
IEEE Recognizing RGB images from RGB-D data is a promising application, which significantly reduces the cost while can still retain high recognition rates. However, existing methods still suffer from the domain shifting problem due to conventional surveillance cameras and depth sensors are using different mechanisms. In this paper, we aim to simultaneously solve the above two challenges: 1) how to take advantage of the additional depth information in the source domain? 2) how to reduce the data distribution mismatch between the source and target domains? We propose a novel method called adaptive Visual- Depth Embedding (aVDE) which learns the compact shared latent space between two representations of labeled RGB and depth modalities in the source domain first. Then the shared latent space can help the transfer of the depth information to the unlabeled target dataset. At last, aVDE models two separate learning strategies for domain adaptation (feature matching and instance reweighting) in a unified optimization problem, which matches features and reweights instances jointly across the shared latent space and the projected target domain for an adaptive classifier. We test our method on five pairs of datasets for object recognition and scene classification, the results of which demonstrates the effectiveness of our proposed method.
Author(s): Cai Z, Long Y, Shao L
Publication type: Article
Publication status: Published
Journal: IEEE Transactions on Image Processing
Print publication date: 01/05/2018
Online publication date: 16/02/2018
Acceptance date: 02/04/2016
Date deposited: 05/03/2018
ISSN (print): 1057-7149
ISSN (electronic): 1941-0042
Altmetrics provided by Altmetric