• DocumentCode
    25226
  • Title

    Unsupervised Adaptation Across Domain Shifts by Generating Intermediate Data Representations

  • Author

    Gopalan, Raghavan ; Ruonan Li ; Chellappa, Rama

  • Author_Institution
    Video & Multimedia Technol. Res. Dept., AT&T Labs.-Res., Middletown, NJ, USA
  • Volume
    36
  • Issue
    11
  • fYear
    2014
  • fDate
    Nov. 1 2014
  • Firstpage
    2288
  • Lastpage
    2302
  • Abstract
    With unconstrained data acquisition scenarios widely prevalent, the ability to handle changes in data distribution across training and testing data sets becomes important. One way to approach this problem is through domain adaptation, and in this paper we primarily focus on the unsupervised scenario where the labeled source domain training data is accompanied by unlabeled target domain test data. We present a two-stage data-driven approach by generating intermediate data representations that could provide relevant information on the domain shift. Starting with a linear representation of domains in the form of generative subspaces of same dimensions for the source and target domains, we first utilize the underlying geometry of the space of these subspaces, the Grassmann manifold, to obtain a `shortest´ geodesic path between the two domains. We then sample points along the geodesic to obtain intermediate cross-domain data representations, using which a discriminative classifier is learnt to estimate the labels of the target data. We subsequently incorporate non-linear representation of domains by considering a Reproducing Kernel Hilbert Space representation, and a low-dimensional manifold representation using Laplacian Eigenmaps, and also examine other domain adaptation settings such as (i) semi-supervised adaptation where the target domain is partially labeled, and (ii) multi-domain adaptation where there could be more than one domain in source and/or target data sets. Finally, we supplement our adaptation technique with (i) fine-grained reference domains that are created by blending samples from source and target data sets to provide some evidence on the actual domain shift, and (ii) a multi-class boosting analysis to obtain robustness to the choice of algorithm parameters. We evaluate our approach for object recognition problems and report competitive results on two widely used Office and Bing adaptation data sets.
  • Keywords
    Hilbert spaces; data acquisition; data structures; differential geometry; eigenvalues and eigenfunctions; object recognition; Bing; Grassmann manifold; Laplacian Eigenmaps; Office; data distribution; discriminative classifier; domain adaptation; domain shifts; fine-grained reference domains; geodesic path; intermediate cross-domain data representations; intermediate data representations; kernel Hilbert space representation; labeled source domain training data; low-dimensional manifold representation; multiclass boosting analysis; object recognition problems; testing data sets; training data sets; two-stage data-driven approach; unconstrained data acquisition scenarios; unlabeled target domain test data; unsupervised adaptation; Adaptation models; Data models; Kernel; Manifolds; Object recognition; Training; Vectors; Domain adaptation; Grassmann manifold; Object recognition; Unsupervised; object recognition; unsupervised;
  • fLanguage
    English
  • Journal_Title
    Pattern Analysis and Machine Intelligence, IEEE Transactions on
  • Publisher
    ieee
  • ISSN
    0162-8828
  • Type

    jour

  • DOI
    10.1109/TPAMI.2013.249
  • Filename
    6684145