Tensor-tensor algebra for optimal representation and compression of multiway data

被引:58
|
作者
Kilmer, Misha E. [1 ]
Horesh, Lior [2 ]
Avron, Haim [3 ]
Newman, Elizabeth [4 ]
机构
[1] Tufts Univ, Dept Math, Medford, MA 02155 USA
[2] IBM Res, Math AI, Yorktown Hts, NY 10598 USA
[3] Tel Aviv Univ, Sch Math Sci, IL-6997801 Tel Aviv, Israel
[4] Emory Univ, Dept Math, Atlanta, GA 30322 USA
关键词
tensor  compression  multiway data  SVD  rank; FACTORIZATION;
D O I
10.1073/pnas.2015851118
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
With the advent of machine learning and its overarching pervasiveness it is imperative to devise ways to represent large datasets efficiently while distilling intrinsic features necessary for subsequent analysis. The primary workhorse used in data dimensionality reduction and feature extraction has been the matrix singular value decomposition (SVD), which presupposes that data have been arranged in matrix format. A primary goal in this study is to show that high-dimensional datasets are more compressible when treated as tensors (i.e., multiway arrays) and compressed via tensor-SVDs under the tensor-tensor product constructs and its generalizations. We begin by proving Eckart-Young optimality results for families of tensor-SVDs under two different truncation strategies. Since such optimality properties can be proven in both matrix and tensor-based algebras, a fundamental question arises: Does the tensor construct subsume the matrix construct in terms of representation efficiency? The answer is positive, as proven by showing that a tensor-tensor representation of an equal dimensional spanning space can be superior to its matrix counterpart. We then use these optimality results to investigate how the compressed representation provided by the truncated tensor SVD is related both theoretically and empirically to its two closest tensor based analogs, the truncated high-order SVD and the truncated tensor-train SVD.
引用
收藏
页数:12
相关论文
共 50 条
  • [21] Learning from Multiway Data: Simple and Efficient Tensor Regression
    Yu, Rose
    Liu, Yan
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 48, 2016, 48
  • [22] Optimizing Compression Schemes for Parallel Sparse Tensor Algebra
    Xu, Helen
    Schardl, Tao B.
    Pellauer, Michael
    Emer, Joel S.
    2023 IEEE HIGH PERFORMANCE EXTREME COMPUTING CONFERENCE, HPEC, 2023,
  • [23] Optimizing Compression Schemes for Parallel Sparse Tensor Algebra
    Xu, Helen
    Schardl, Tao B.
    Pellauer, Michael
    Emer, Joel S.
    2023 DATA COMPRESSION CONFERENCE, DCC, 2023, : 373 - 373
  • [24] The tensor structure on the representation category of the Wp triplet algebra
    Tsuchiya, Akihiro
    Wood, Simon
    JOURNAL OF PHYSICS A-MATHEMATICAL AND THEORETICAL, 2013, 46 (44)
  • [25] Analysis of the tensor-tensor type scalar tetraquark states with QCD sum rules
    Wang, Zhi-Gang
    Zhang, Jun-Xia
    EUROPEAN PHYSICAL JOURNAL C, 2016, 76 (12):
  • [26] Bayesian Robust Tensor Ring Decomposition for Incomplete Multiway Data
    Huang, Zhenhao
    Qiu, Yuning
    Chen, Xinqi
    Sun, Weijun
    Zhou, Guoxu
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2024, 54 (07): : 4005 - 4018
  • [27] On Tensor algebra
    Waelsch, E
    MATHEMATISCHE ZEITSCHRIFT, 1924, 20 : 188 - 192
  • [28] Applications of tensor (multiway array) factorizations and decompositions in data mining
    Morup, Morten
    WILEY INTERDISCIPLINARY REVIEWS-DATA MINING AND KNOWLEDGE DISCOVERY, 2011, 1 (01) : 24 - 40
  • [29] NeuralCP: Bayesian Multiway Data Analysis with Neural Tensor Decomposition
    Bin Liu
    Lirong He
    Yingming Li
    Shandian Zhe
    Zenglin Xu
    Cognitive Computation, 2018, 10 : 1051 - 1061
  • [30] Kernelization of Tensor-Based Models for Multiway Data Analysis
    Zhao, Qibin
    Zhou, Guoxu
    Adali, Tulay
    Zhang, Liqing
    Cichocki, Andrzej
    IEEE SIGNAL PROCESSING MAGAZINE, 2013, 30 (04) : 137 - 148