• DocumentCode
    32949
  • Title

    Parallel Randomly Compressed Cubes : A scalable distributed architecture for big tensor decomposition

  • Author

    Sidiropoulos, Nicholas ; Papalexakis, Evangelos E. ; Faloutsos, Christos

  • Author_Institution
    Electr. & Comput. Eng., Univ. of Minnesota, Minneapolis, MN, USA
  • Volume
    31
  • Issue
    5
  • fYear
    2014
  • fDate
    Sept. 2014
  • Firstpage
    57
  • Lastpage
    70
  • Abstract
    This article combines a tutorial on state-of-the-art tensor decomposition as it relates to big data analytics, with original research on parallel and distributed computation of low-rank decomposition for big tensors, and a concise primer on Hadoop?MapReduce. A novel architecture for parallel and distributed computation of low-rank tensor decomposition that is especially well suited for big tensors is proposed. The new architecture is based on parallel processing of a set of randomly compressed, reduced-size replicas of the big tensor. Each replica is independently decomposed, and the results are joined via a master linear equation per tensor mode. The approach enables massive parallelism with guaranteed identifiability properties: if the big tensor is of low rank and the system parameters are appropriately chosen, then the rank-one factors of the big tensor will indeed be recovered from the analysis of the reduced-size replicas. Furthermore, the architecture affords memory/storage and complexity gains of order for a big tensor of size of rank F with No sparsity is required in the tensor or the underlying latent factors, although such sparsity can be exploited to improve memory, storage, and computational savings.
  • Keywords
    matrix decomposition; signal processing; tensors; Hadoop; MapReduce; linear equation; parallel randomly compressed cubes; scalable distributed architecture; tensor decomposition; Big data; Complexity theory; Data storage; Information analysis; Matrix decomposition; Scalability; Tensile stress; Tensors; Tutorials;
  • fLanguage
    English
  • Journal_Title
    Signal Processing Magazine, IEEE
  • Publisher
    ieee
  • ISSN
    1053-5888
  • Type

    jour

  • DOI
    10.1109/MSP.2014.2329196
  • Filename
    6879586