Figures
Abstract
How can we accurately and efficiently decompose a tensor stream? Tensor decomposition is a crucial task in a wide range of applications and plays a significant role in latent feature extraction and estimation of unobserved entries of data. The problem of efficiently decomposing tensor streams has been of great interest because many real-world data dynamically change over time. However, existing methods for dynamic tensor decomposition sacrifice the accuracy too much, which limits their usages in practice. Moreover, the accuracy loss becomes even more serious when the tensor stream has an inconsistent temporal pattern since the current methods cannot adapt quickly to a sudden change in data. In this paper, we propose DAO-CP, an accurate and efficient online CP decomposition method which adapts to data changes. DAO-CP tracks local error norms of the tensor streams, detecting a change point of the error norms. It then chooses the best strategy depending on the degree of changes to balance the trade-off between speed and accuracy. Specifically, DAO-CP decides whether to (1) reuse the previous factor matrices for the fast running time or (2) discard them and restart the decomposition to increase the accuracy. Experimental results show that DAO-CP achieves the state-of-the-art accuracy without noticeable loss of speed compared to existing methods.
Citation: Son S, Park Y-c, Cho M, Kang U (2022) DAO-CP: Data-Adaptive Online CP decomposition for tensor stream. PLoS ONE 17(4): e0267091. https://doi.org/10.1371/journal.pone.0267091
Editor: Ning Cai, Beijing University of Posts and Telecommunications, CHINA
Received: December 13, 2021; Accepted: April 1, 2022; Published: April 14, 2022
Copyright: © 2022 Son et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: The data underlying the results presented in the study are available from https://github.com/snudatalab/DAO-CP.
Funding: This work was partly supported by the National Research Foundation of Korea(NRF) funded by MSIT(2022R1A2C3007921) and MSIT(2019R1A2C2004990). This work was partly supported by Institute of Information & communications Technology Planning & Evaluation (IITP) grant funded by the Korea government(MSIT) [NO.2021-0-02068, Artificial Intelligence Innovation Hub (Artificial Intelligence Institute, Seoul National University)] and [NO.2021-0-01343, Artificial Intelligence Graduate School Program (Seoul National University)]. The Institute of Engineering Research and ICT at Seoul National University provided research facilities for this work. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
Competing interests: The authors have declared that no competing interests exist.
Introduction
Given a tensor stream, how can we decompose it accurately and efficiently? A multi-dimensional array or tensor has been a fundamental component for numerous applications including signal processing [1–3], computer vision [4–6], graph analysis [7, 8], and statistics [9]. Tensor decomposition is a generalization of the matrix decomposition, and plays an important role in latent feature discovery and estimation of unobservable entries [10–12].
A tensor is called dynamic if its size and value change over time (for example, time-evolving network traffic data, social networks, and so on), or static otherwise [13]. Analysis of dynamic tensors with tensor decomposition is a crucial task. However, the methods devised for static tensor decomposition cannot be easily applied for dynamic tensor analysis since such static methods perform many iterations of computation until convergence at every time step, which leads to prohibitively large time cost.
On the other hand, dynamic tensor decomposition methods aim to incrementally and quickly analyze tensors; however, the accuracy of existing methods is not satisfactory. Indeed, current dynamic methods (1) bypass the factor update for temporal mode to improve the speed [14], or (2) decompose the current tensor slice using prior factor matrices [15, 16]. Unfortunately, these methods suffer from poor accuracy when the tensor stream has an inconsistent temporal pattern because they cannot adapt quickly to abrupt changes in data [17, 18].
In this paper, we propose Data-Adaptive Online CP decomposition (DAO-CP), an accurate and efficient tensor stream decomposition algorithm which adapts to data changes. The main ideas of DAO-CP are to (1) detect change points of “themes” in a tensor stream by tracking local error norms, and (2) re-decompose the tensor stream whenever a new theme is discovered. DAO-CP automatically decides whether to reuse the previous results of decomposition or to discard them depending on how much changes are detected in the tensor stream. Consequently, it provides much more accurate and fast decomposition for real-world datasets even with inconsistent temporal patterns. Furthermore, we introduce complementary matrices in order to reduce the redundant computations in CP-ALS optimization. We also simplify the estimation loss function from DTD [15] by fixing the non-temporal modes. As a result, DAO-CP achieves lower time complexity than the existing methods without accuracy loss. Through experiments, we show that DAO-CP outperforms the current state-of-the-art algorithms in terms of accuracy with little sacrifice in running time. We also investigate the sensitivity and the effect of hyperparameters of our proposed method.
The main contributions are summarized as follows:
- Method. We propose DAO-CP, an accurate and efficient online method for tensor stream decomposition.
- Analysis. We theoretically analyze the computational complexity of DAO-CP and compare it to existing methods.
- Experiments. DAO-CP shows the state-of-the-art accuracy on both synthetic and real-world datasets without significant loss of speed (see Fig 1).
DAO-CP automatically detects a change of theme (for example, an object starts moving or a scene changes) and re-decomposes the data depending on the degree of changes. Note that DAO-CP results in much more clear images than the competitors with little sacrifice in speed.
The code and datasets are available at https://github.com/snudatalab/DAO-CP. The rest of this paper is organized as follows. We first demonstrate preliminaries of tensor decomposition algorithms. We then present our proposed method in detail. After showing experimental results, we discuss related works, and conclude the paper.
Preliminaries
We describe preliminaries of tensors and tensor decomposition algorithms. Table 1 summarizes the symbols used in this paper.
Tensors
Tensors are multi-dimensional arrays that generalize vectors (1-order tensors) and matrices (2-order tensors) to higher orders. We denote vectors with bold lowercase letters (a), matrices with bold capital letters (A), and tensors with bold calligraphic letters (). An N-th order tensor has N modes whose lengths are I1, ⋯, IN, respectively. A tensor can be unfolded or matricized along any of its modes [19], and the unfolded matrix of along the n-th mode is denoted by X(n). When a tensor is unfolded, its elements are reordered into a matrix form; the mode-n unfolding matrix of a tensor maps the (i1, ⋯, iN)-th element of to the (in, j)-th element of X(n), where (1) We define the Frobenius norm of a tensor using the notation ‖⋅‖ as follows: (2) In what follows, we briefly define several important matrix products. The Kronecker product A ⊗ B of matrices and is a matrix of size IK × JL and defined as follows: The Hadamard product A ⊛ B and Khatri-Rao product A ⊙ B are two essential matrix products used in tensor decomposition. The Hadamard product is simply the element-wise product of two matrices A and B of the same size. The Khatri-Rao product is a column-wise Kronecker product: (3) where {an} and {bn} denote the column vectors of and , respectively.
Tensor decomposition
CANDECOMP/PARAFAC (CP) decomposition is one of the most widely used methods for tensor decomposition, which is considered to be a key building block in many other variants [1, 20]. CP decomposition factorizes a tensor into a sum of rank-one tensors: (4) where the number R of rank-one tensor sets is called the rank of the resulting tensor. The factor matrices {A1, ⋯, AN} refer to the combination of the vectors from the rank-one components, i.e., (5) We express the CP decomposition result of a tensor using Kruskal operator 〚⋅〛 and the unfolding matrix, where the Kruskal operator provides a shorthand notation for the sum of outer products of the columns in factor matrices [21]: Then, CP decomposition aims to find the factor matrices that minimize the estimation error defined as follows: (6) CP alternating least squares (CP–ALS) has been extensively used for this optimization problem. The main idea of ALS is to divide the original problem into N sub-problems, where each sub-problem corresponds to updating one factor matrix while keeping all the others fixed [20]: (7)
Online tensor decomposition
Of particular interest in the problem of tensor decomposition is an efficient online algorithm for time-evolving tensors. We think of a tensor as a set of “slices” given at each time step. Given an N-order time-evolving tensor , we expand it as a form of , where is the previous tensor data and is a new tensor slice for one time step. Then, the goal is to efficiently decompose the tensor given the previous decomposition result : (8) where and . This is done by minimizing the estimation error defined as follows: (9)
Related works
Tensor stream decomposition is widely studied under CP decomposition [22, 23]. Existing works employ one of the following two ideas: they update (1) only the non-temporal factors with precomputed auxiliary matrices [14], or (2) whole factors considering prior decomposition results [15, 16]. We describe three main approaches (OnlineCP, SeekAndDestroy, and DTD) for dynamic tensor decomposition, and compare them with our proposed method.
OnlineCP
OnlineCP [14] preserves the previous temporal factor to efficiently decompose new tensor slices. After updates of non-temporal factors and the partial temporal factor, it simply appends a part of the temporal factor matrix to the previous matrix. OnlineCP avoids duplicated computations such as Khatri-Rao and Hadamard products by introducing auxiliary matrices. It computes complementary matrices before ALS iteration and yields a new decomposition. Despite its low computational cost, the approach cannot achieve an accurate decomposition due to the lack of consideration on the change of themes in data (see Fig 2). Note that DAO-CP solves this problem by tracking local error norms of the tensor stream and detecting a change point of themes, which enables an accurate decomposition even when the data have an inconsistent temporal pattern.
In this figure, the length of time factor (with horizontal axis) becomes larger for each time step. Contrary to static decomposition methods, OnlineCP reduces computational cost using approximation with an additional constraint: it updates only the non-temporal mode, reusing the same temporal factor for all time steps. However, this leads to a substantial loss of accuracy if an incoming tensor has a different theme compared to previous tensors (e.g., theme changes A → B, B′ → C, or C → D). Thus, OnlineCP cannot achieve an accurate decomposition due to the lack of consideration on the change of themes in data.
SeekAndDestroy
SeekAndDestroy [16] additionally uses rank estimation to discover latent concepts and detect concept drift in streaming tensors. The method estimates the rank of each incoming tensor slice, and updates the previous decomposition after alleviating concept drift. However, SeekAndDestroy requires extra computation due to the rank estimation for every time step, which causes a substantial loss of speed. Moreover, it consistently performs worse than OnlineCP when the initial rank of OnlineCP is fine-tuned. Note that our proposed method efficiently detects the change of theme in streaming data because it does not require estimating the actual rank numbers, but only tracks local error in order to rapidly capture the change points.
DTD
DTD [15] was originally introduced as a part of MAST which is a low-rank tensor completion method to fill in the missing entries of the incomplete multi-aspect tensor stream. The method manages to reduce the time complexity by reusing the previous decomposition that approximates the tensor stacked until new slices come in. Specifically, for an N-th order tensor stream, DTD partitions the data into 2N sub-tensors for each time step and uses binary tuples (i1, ⋯, iN) ∈ Θ = {0, 1}N to denote the sub-tensors. Then, given that approximates , one can reformulate the estimation error of online tensor decomposition as follows (see the notations from Preliminaries): (10) where μ ∈ [0, 1] is the forgetting factor which alleviates the influence of the previous decomposition error. Although DTD is an efficient method, it suffers from poor accuracy when an incoming tensor has an entirely different pattern compared to previous tensors, as it still tries to reuse the prior decomposition result. Our proposed method addresses the problem by using “re-decomposition” process and adapting quickly to sudden changes in data, and significantly increases the accuracy of decomposition.
Proposed method
We propose DAO-CP, an accurate and efficient online algorithm for tensor stream decomposition.
Overview
DAO-CP is a time and memory efficient algorithm for accurate online CP–ALS tensor decomposition which adapts to data changes. The challenge of decomposing time-evolving tensors is to improve accuracy without sacrificing speed and memory usage. Considering that the themes of data change over time, we propose detecting the change points of themes and using different strategies depending on the degree of change. The main challenges are as follows:
- Reduce computational cost. How can we reduce the arithmetic cost for updating decomposition factors of tensor streams?
- Identify themes in data streams. How can we capture the latent themes in tensor streams and detect the change points of them?
- Increase decomposition accuracy. How can we exploit the detected change points of themes and increase the decomposition accuracy?
To address the above challenges, we propose the following approaches.
- Build an updatable framework for tensor stream. We use complementary matrices and previous decomposition results recursively, where the complementary matrices are updated only when there is a change in non-temporal factors, thus reducing the redundant operations.
- Detect data changes by tracking error norms. We continuously track the error norms of incoming data slices in the tensor stream, detecting a sudden accuracy drop based on z-score analysis, which we regard as a change point of themes.
- Re-decompose the tensor stream when a new theme is detected. Once a sudden change in theme is detected, we choose whether to refine or split the tensor stream depending on the degree of changes. We also introduce memory rate to improve the refinement process. These techniques determine how much information from the previous decomposition should be retained, balancing the trade-off between accuracy and speed.
Update rules for DAO-CP
Let be an N-order time-evolving tensor, where is the previous tensor data and is a new tensor slice; we assume that the first mode is the temporal mode. We design our update rules to efficiently decompose the tensor , given the previous decomposition result . We partition the temporal factor matrix A1 into old and new parts as using and , where R is the decomposition rank.
In order to consider the degree of change in themes, we introduce the memory rate ρ ∈ [0.5, 1] which determines how much weight to assign to the decomposition of the previous tensor data. We define the estimation error as a restricted form of the one from DTD [15], where the non-temporal modes of the tensor stream are fixed: (11) The optimization of the estimation error is based on CP-ALS [14, 20]. Note that we simplify the estimation error from DTD by setting the changes in non-temporal modes to zeros because there is a change only in the temporal mode for our problem. The update rules to minimize in (11) for each factor matrix are derived as follows: Note that we also update the prior temporal factors to further increase the accuracy of decomposition. If the previous temporal factors are not updated, they harm the accuracy of method whenever there is a change of theme because they are optimized only for the previous theme of the data. However, it is computationally demanding to directly apply these recursive processes. To address the problem, we introduce two complementary matrices G and H, (12) where G and H are updated only when there is a change in non-temporal factors, thus reducing the redundant computations. This leads to the following modified update rules: (13) (14) (15) The overall update process is outlined in Algorithm 1.
Algorithm 1: DAO-CP Alternating Least Square (DAO-CP-ALS)
Input: Factors from old tensors , new tensor slice , memory rate ρ, and number of ALS iterations niter
Output: Updated factors 〚A1, ⋯, AN〛
1 Initialize complementary matrices G and H by (12)
2 for n ← 1 to niter do
3 Update using (14) // latter part of temporal factor A1
4 for Ai≠1 ∈ non-temporal factors do
5 Update Ai, G and H using (15)
6 Update using (13) // former part of temporal factor A1
Change points detection with local error norm
The key of our proposed method is to detect change points of themes in tensor streams and thereby adapt quickly to abrupt changes in data. To do this, we continuously track the decomposition error of tensor streams and detect a sudden accuracy drop which we regard as a change point of themes. Such an accuracy drop is captured by measuring local error norm for the new tensor slice and its decomposition result: (16) We assume that follows a normal distribution and keep track of its mean and variance to detect outliers. Note that we should update the mean and variance in an online manner. This is achievable by using Welford’s algorithm [24, 25], which provides accurate estimates of mean and variance without the necessity of keeping the entire data. Moreover, the method requires only one pass of given data in order to compute their sample mean and variance. Using Welford’s algorithm, we detect outliers in the current local error norm by z-score analysis of the following criterion: (17) where L is a threshold of anomaly. Changing the value of L, one can fine-tune the criterion on whether a new tensor slice is similar to the previous tensors or not.
Re-decomposition process
Once a sudden change of theme is detected by z-score analysis, DAO-CP exploits this information to increase the decomposition accuracy. As a new tensor slice is stacked for each time step, DAO-CP updates the factor matrices following the optimization scheme described in Algorithm 1. It then computes the z-score in the local error norm distribution and performs “re-decomposition” depending on the score. Tracking the distribution of local error norm and setting the z-score criteria enable DAO-CP to automatically choose the best strategy between split and refinement processes depending on the degree of changes. Fig 3 illustrates the intuition of the two processes, and Table 2 shows the criterion for each process.
When the z-score exceeds the split threshold Ls (e.g., change from theme A to B), DAO-CP re-initializes the new tensor slice using the static CP decomposition. When the z-score is between Lr and Ls (e.g, change from theme B to B′), the refinement process determines how much information from the previous factors should be retained. Consequently, DAO-CP provides both fast and accurate decomposition for tensor streams even with inconsistent temporal patterns.
Ls is the threshold of splitting and initializing the decomposition, and Lr is the threshold of refining the previous decomposition. By changing the two hyperparameters, we can fine-tune the re-decomposition process to balance the trade-off between accuracy and speed.
Split process.
What if an incoming tensor slice has an entirely different theme compared to previous tensors? In this case, reusing the prior results of decomposition will cause a substantial loss of accuracy. To address the problem, we design a split process which divides the streaming tensor into separate tensors of different themes, using a threshold Ls. Despite extra costs of space and time due to re-initialization, the split process enables DAO-CP to successfully avoid the unexpected accuracy drop (lines 8-13 in Algorithm 2).
Refinement process.
The refinement process is used to update the decomposition result when there is only a modest difference in the theme from the previous tensor. We use the hyperparameter Lr to fine-tune the refinement criterion on whether an incoming tensor slice is similar to the previous tensors or not. We use the memory rate 1 − ρ because we need to focus more on the new slice (note that 1 − ρ ⩽ ρ since ρ ∈ [0.5, 1]). The ALS operation also takes the z-score and is performed extra more times accordingly, because a higher z-score implies that there is a more abrupt change in data. As a result, these techniques determine how much information from the previous decomposition should be retained, balancing the trade-off between accuracy and running time (lines 14-16 in Algorithm 2). The full computation of DAO-CP is outlined in Algorithm 2.
Algorithm 2: Data-Adaptive Online CP Decomposition (DAO-CP)
Inpit: Tensor stream , memory rate ρ, and number of ALS iterations niter
Output: Decomposition factor set
1 new slice from
2 Initialize using CP decomposition of
3 Calculate error norm between and
4 Initialize Welford with
5 for from do
6 〚A1, ⋯, AN〛 ← DAO-CP-ALS
7 Calculate error norm between and 〚A1, ⋯, AN〛
/* Split Process */
8 if Welford z-score then
9 Store the previous factors to
10 Initialize using CP decomposition of
11 Calculate error norm between and
12 Initialize Welford with
13 continue
/* Refinement Process */
14 if then
15 〚A1, ⋯, AN〛 ← DAO-CP-ALS
16 Calculate error norm between and 〚A1, ⋯, AN〛
17 Update Welford with
18
19 Store the previous factors to
Theoretical analysis
We analyze the computational complexity of DAO-CP. The following symbols are used for the analysis: N (dimensionality), R (rank), (time length of the new data slice), (time length of the formerly stacked data), Ii≠1 (mode length of the non-temporal i-th mode), and niter (number of ALS iterations).
Table 3 summarizes the comparison of DAO-CP to existing tensor decomposition methods. We find that DAO-CP has the lowest arithmetic cost among the methods except OnlineCP. Note that even though OnlineCP has lower complexity than our proposed method, it suffers from poor accuracy due to the lack of consideration on temporal change of data, which limits its usage in practice (see Fig 4).
Since Full–CP is not an online method, we evaluate its fitness whenever a new slice is added. Detecting the change points of theme, DAO-CP successfully increases the accuracy of decomposition, which is even higher than that of Full–CP.
Lemma 1. The time complexity of initializing the complementary matrices G and H by (12) is O(R2 ∑k≠1 Ik).
Proof. Because the operands and in Hadamard operations are R × R matrices, it takes O(R2 ⋅ Ik) for multiplication of and . Thus, the total arithmetic complexity of computing the matrices G and H is given by O(R2 ⋅ ∑k ≠ 1 Ik).
Lemma 2. The time complexity of updating the factor matrix by (13) is .
Proof. Finding the pseudo-inverse matrix H† of H takes O(R3), and the matrix multiplication of and requires operations.
Lemma 3. The time complexity of updating the factor matrix by (14) is .
Proof. The Khatri-Rao product consumes O(R ⋅ ∏k≠1 Ik). We compute the matrix multiplication in the order of appearance, and its time complexity is . This, combined with the complexity O(R3) of finding H†, yields the desired result.
Lemma 4. The time complexity of updating Ai≠1, G and H by (15) is .
Proof. Computing , and require O(R2 ⋅ Ii), and time costs, respectively. Constructing the pseudo-inverse matrix costs O(R3) since its size is R × R. For the first part before the pseudo-inverse, the first term is calculated in O(R2 ⋅ Ii + R2 ⋅ Iold) time, and the second term is computed similarly to Lemma 3 with complexity. Combining those results, we conclude the proof.
Theorem 5 (Complexity of DAO-CP-ALS). The time complexity of DAO–CP–ALS (Algorithm 1) is .
Proof. The computational cost of updating all the non-temporal factor matrices {Ai ≠ 1} is by Lemma 4, which can be written as . Combining the complexities of updating {Ai} by Lemmas 2 and 3 gives the following arithmetic cost for a single ALS iteration: . Thus, including the initialization cost from Lemma 1, we obtain the desired result.
Experiments
In this section, we experimentally evaluate DAO-CP to answer the following questions.
- Q1. Reconstruction error. How accurately does DAO-CP decompose real-world tensor streams compared to existing methods?
- Q2. Time cost. How much does DAO-CP improve the speed of tensor stream decomposition compared to existing online methods?
- Q3. Effect of thresholds Lr and Ls. How do the different choices of Lr and Ls for re-decomposition criteria affect the performance of DAO-CP?
- Q4. Refinement and split processes. How does each of the re-decomposition processes affect the performance of DAO-CP on real-world datasets?
In the following, we describe the experimental settings and answer the questions with the experimental results.
Experimental settings
All the experiments are conducted in a workstation with a single CPU (Intel(R) Xeon(R) CPU E5-2630 v4 @ 2.20GHz).
Datasets.
We use four real-world tensor streams and a synthetic tensor stream summarized in Table 4, where the first mode of each tensor corresponds to the temporal mode. We construct a tensor stream by splitting an original tensor into slices along the time mode (first mode); e.g., we make 41 slices from Sample Video dataset, where each slice is of dimension (5, 240, 320, 3).
- Sample Video dataset is a series of animation frames with RGB values. For this dataset, we expect that changes of theme occur when an object starts moving or a scene changes.
- Stock Price includes data on 140 stocks listed on the Korea Stock Price Index 200 (KOSPI 200) from Jan 2, 2008 to June 30, 2020, where each stock contains five features: adjusted opening price, closing price, highest price, lowest price, and trading volume. Changes of theme may occur when real-world events affect the rise or fall of stock markets.
- Airport Hall is a video recorded in an airport, initially used to verify OLSTEC [17, 27]. We expect that a sudden change of theme occurs when a crowd of people surges toward the airport during flight departure or arrival time.
- Korea Air Quality dataset consists of daily air pollutant levels for various locations in Seoul, South Korea from Sep 1, 2018 to Sep 31, 2019. The themes may continuously change depending on weather environment.
- Synthetic is made of concatenated tensors, which is the summation Tmain + Ttheme + Tnoise of three tensors {Tmain, Ttheme, Tnoise}, each referring to a , , normally distributed randomized tensor, respectively, of size (1000, 10, 20, 30). Note that we simulate the changes of theme using the tensor Ttheme.
Competitors.
We compare DAO-CP with existing dynamic tensor decomposition methods including OnlineCP [14] and DTD [15], as well as with the static CP decomposition method, Full–CP [20]. All the methods are implemented in Python3 using the TensorLy library.
Parameters.
The parameters Ls and Lr of DAO-CP are set to the values listed in Table 4. The section “Effect of Thresholds Ls and Lr” is an exception, where we vary the two values to investigate the effect of different thresholds. We set the memory rate as ρ = 0.8 for all the experiments.
Evaluation measure.
To evaluate our proposed method, we use local and global error norms and , as well as the corresponding “fitness” scores and , which are defined as follows: denotes the fitness for an incoming data slice at each time step, while is the fitness for whole tensors. They are the normalized versions of error norms with respect to data size, designed to compare the decomposition accuracy for multiple datasets with different sizes.
Reconstruction error
We compare DAO-CP to its competitors in terms of fitness, varying the decomposition rank in Fig 4. The average of local fitness is the mean of that is computed at every time step. Note that DAO-CP shows higher fitness than the existing methods in most cases, regardless of ranks.
Running time
DAO-CP allows an accurate tensor factorization by exploiting the characteristic of data and detecting change points. However, this results in a slightly longer running time due to the re-decomposition process. Fig 5 shows the running times of DAO-CP and other methods for various ranks. Note that DAO-CP has moderate running times between the static and dynamic decomposition methods, showing promising speeds comparable to the other dynamic algorithms (DTD and OnlineCP) and significantly faster than the static method (Full–CP).
Since Full–CP is not an online method, we evaluate its fitness whenever a new slice is added. Note that DAO-CP results in a promising speed comparable to DTD and OnlineCP with much more accurate decomposition, and significantly faster than Full–CP.
Effect of thresholds Lr and Ls
We change the values of Ls and Lr to investigate the effect of split and refinement processes. Table 5 shows the results, where the number of refinement or split points changes as Lr and Ls vary. Note that both the processes lead to more accurate decomposition with extra time costs, and among them the split process has bigger trade-offs because it requires re-initialization. As a result, the more parts the tensor is split into (smaller Ls), the more accurate decomposition DAO-CP yields with extra costs of time. More refinement processes (smaller Lr) also have a similar effect, although the trade-offs are relatively small. In contrast to the split process, it requires more memory to store intermediate data such as auxiliary matrices G and H. In a practical standpoint, these observations are very useful because one can benefit from the hyperparameter tuning when there is a particular importance in one of accuracy, speed, or memory usage.
The memory usage means the summation of byte allocation to store intermediate data to calculate next decomposition results (e.g. auxiliary matrices G and H). We use Korea Air Quality dataset with rank 20, and change Lr and Ls to investigate the effect of refinement and split processes. Note that the lower the thresholds is set, the more frequently the re-decomposition processes are executed. Thus, one can benefit from this observation when there is a particular importance in one of accuracy, speed, or memory usage depending on target tasks.
Refinement and split processes
Recall that the split process is used to start a new decomposition when an entirely different theme is detected, while the refinement process is used when there is only a modest difference from the previous decomposition. Fig 1 validates the importance of these intuitions, showing that DAO-CP results in remarkable performance for the video datasets with different scenes and object movements.
To further investigate the effects of split and refinement processes, we consider the following question: for each time step (or data slice), how does the re-decomposition process affect the running time and local error norm? With Sample Video dataset, we compare the running time and local error norm of DAO-CP to its competitors in Fig 6. We observe that both split and refinement processes significantly reduces the local error norm with only a modest sacrifice of running time.
Each re-decomposition process (at split point) significantly reduces the local error norm with only a modest sacrifice of running time (e.g., vertical line connecting Pprev, Pnext, Qprev, and Qnext). Note that DAO-CP runs slower than the other dynamic methods (OnlineCP and DTD) only when one of split or refinement processes is performed to increase the accuracy (horizontal line R: average running time of competitor methods).
Conclusions
In this paper, we propose DAO-CP, an efficient algorithm for decomposing time-evolving tensors. DAO-CP automatically detects a change point of theme in tensor streams and decides whether to re-decompose the tensors or not. Experimental results show that the proposed DAO-CP outperforms the current state-of-the-art methods on both synthetic and real-world datasets. We also investigate the effect of hyperparameters of our proposed method and demonstrate the advantages of trading-off between accuracy, speed, and memory usage. Future works include extending our method for simultaneously decomposing many related time-evolving tensors.
References
- 1. Sidiropoulos ND, De Lathauwer L, Fu X, Huang K, Papalexakis EE, Faloutsos C. Tensor decomposition for signal processing and machine learning. IEEE Transactions on Signal Processing. 2017;65(13):3551–3582.
- 2. Cichocki A, Mandic D, De Lathauwer L, Zhou G, Zhao Q, Caiafa C, et al. Tensor decompositions for signal processing applications: From two-way to multiway component analysis. IEEE signal processing magazine. 2015;32(2):145–163.
- 3. Cong F, Lin QH, Kuang LD, Gong XF, Astikainen P, Ristaniemi T. Tensor decomposition of EEG signals: a brief review. Journal of neuroscience methods. 2015;248:59–69. pmid:25840362
- 4.
Govindu VM. A tensor decomposition for geometric grouping and segmentation. In: 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05). vol. 1. IEEE; 2005. p. 1150–1157.
- 5.
Shakeri M, Zhang H. Moving Object Detection Under Discontinuous Change in Illumination Using Tensor Low-Rank and Invariant Sparse Decomposition. In: IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2019, Long Beach, CA, USA, June 16-20, 2019. Computer Vision Foundation / IEEE; 2019. p. 7221–7230. Available from: http://openaccess.thecvf.com/content_CVPR_2019/html/Shakeri_Moving_Object_Detection_Under_Discontinuous_Change_in_Illumination_Using_Tensor_CVPR_2019_paper.html.
- 6.
Ye J, Wang L, Li G, Chen D, Zhe S, Chu X, et al. Learning Compact Recurrent Neural Networks With Block-Term Tensor Decomposition. In: 2018 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2018, Salt Lake City, UT, USA, June 18-22, 2018. IEEE Computer Society; 2018. p. 9378–9387. Available from: http://openaccess.thecvf.com/content_cvpr_2018/html/Ye_Learning_Compact_Recurrent_CVPR_2018_paper.html.
- 7.
Kolda TG, Bader BW, Kenny JP. Higher-order Web link analysis using multilinear algebra. In: Fifth IEEE International Conference on Data Mining (ICDM’05); 2005. p. 8 pp.–.
- 8. Kolda T, Bader B. The TOPHITS model for higher-order web link analysis. In: Workshop on link analysis, counterterrorism and security. vol. 7; 2006. p. 26–29.
- 9. Acar E, Dunlavy DM, Kolda TG, Mørup M. Scalable tensor factorizations for incomplete data. Chemometrics and Intelligent Laboratory Systems. 2011;106(1):41–56.
- 10.
Kang U, Papalexakis E, Harpale A, Faloutsos C. Gigatensor: scaling tensor analysis up by 100 times-algorithms and discoveries. In: Proceedings of the 18th ACM SIGKDD international conference on Knowledge discovery and data mining; 2012. p. 316–324.
- 11.
Cai Y, Tong H, Fan W, Ji P, He Q. Facets: Fast comprehensive mining of coevolving high-order time series. In: Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining; 2015. p. 79–88.
- 12. Sedghi H, Sabharwal A. Knowledge completion for generics using guided tensor factorization. Transactions of the Association for Computational Linguistics. 2018;6:197–210.
- 13.
Sun J, Tao D, Faloutsos C. Beyond streams and graphs: dynamic tensor analysis. In: Proceedings of the 12th ACM SIGKDD international conference on Knowledge discovery and data mining; 2006. p. 374–383.
- 14.
Zhou S, Vinh NX, Bailey J, Jia Y, Davidson I. Accelerating online cp decompositions for higher order tensors. In: Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining; 2016. p. 1375–1384.
- 15.
Song Q, Huang X, Ge H, Caverlee J, Hu X. Multi-aspect streaming tensor completion. In: Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining; 2017. p. 435–443.
- 16.
Pasricha R, Gujral E, Papalexakis EE. Identifying and alleviating concept drift in streaming tensor decomposition. In: Joint European Conference on Machine Learning and Knowledge Discovery in Databases. Springer; 2018. p. 327–343.
- 17.
Kasai H. Online low-rank tensor subspace tracking from incomplete data by CP decomposition using recursive least squares. In: 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE; 2016. p. 2519–2523.
- 18. Mardani M, Mateos G, Giannakis GB. Subspace learning and imputation for streaming big data matrices and tensors. IEEE Transactions on Signal Processing. 2015;63(10):2663–2677.
- 19. Oh S, Park N, Jang JG, Sael L, Kang U. High-performance tucker factorization on heterogeneous platforms. IEEE Transactions on Parallel and Distributed Systems. 2019;30(10):2237–2248.
- 20. Kolda TG, Bader BW. Tensor decompositions and applications. SIAM review. 2009;51(3):455–500.
- 21. Kolda TG. Multilinear operators for higher-order decompositions. Sandia National Laboratories; 2006.
- 22.
Smith S, Huang K, Sidiropoulos ND, Karypis G. Streaming tensor factorization for infinite data sources. In: Proceedings of the 2018 SIAM International Conference on Data Mining. SIAM; 2018. p. 81–89.
- 23.
Najafi M, He L, Philip SY. Outlier-Robust Multi-Aspect Streaming Tensor Completion and Factorization. In: IJCAI; 2019. p. 3187–3194.
- 24. Welford B. Note on a method for calculating corrected sums of squares and products. Technometrics. 1962;4(3):419–420.
- 25. Ling RF. Comparison of several algorithms for computing sample means and variances. Journal of the American Statistical Association. 1974;69(348):859–866.
- 26. Battaglino C, Ballard G, Kolda TG. A practical randomized CP tensor decomposition. SIAM Journal on Matrix Analysis and Applications. 2018;39(2):876–901.
- 27. Zhang S, Guo B, Dong A, He J, Xu Z, Chen SX. Cautionary tales on air-quality improvement in Beijing. Proceedings of the Royal Society A: Mathematical, Physical and Engineering Sciences. 2017;473(2205):20170457. pmid:28989318