A Survey of Multilinear Subspace Learning For Tensor Data
A Survey of Multilinear Subspace Learning For Tensor Data
A Survey of Multilinear Subspace Learning For Tensor Data
University of Toronto, 10 King’s College Road, Toronto, ON, M5A 3G4, Canada
c Department of Electrical and Computer Engineering, Ryerson University, 350
Abstract
Please cite:
Haiping Lu, K. N. Plataniotis and A. N. Venetsanopoulos, “A Survey of Mul-
tilinear Subspace Learning for Tensor Data”, Pattern Recognition, vol. 44, no.
7, pp. 1540-1551, Jul. 2011.
BibTeX:
http://www.dsp.utoronto.ca/~haiping/BibTeX/MSLSurvey2011.bib
With the advances in data collection and storage capabilities, massive multidi-
mensional data are being generated on a daily basis in a wide range of emerg-
ing applications, and learning algorithms for knowledge extraction from these
data are becoming more and more important. Two-dimensional (2D) data in-
clude gray-level images in computer vision and pattern recognition [1–4], mul-
tichannel EEG signals in biomedical engineering [5, 6], and gene expression
data in bioinformatics [7]. Three-dimensional (3D) data include 3D objects in
generic object recognition [8], hyperspectral cube in remote sensing [9], and
gray-level video sequences in activity or gesture recognition for surveillance or
human-computer interaction (HCI) [10, 11]. There are also many multidimen-
sional signals in medical image analysis [12], content-based retrieval [1, 13],
and space-time super-resolution [14] for digital cameras with limited spatial
and temporal resolution. In addition, many streaming data and mining data
are frequently organized as third-order tensors [15–17]. Data in environmen-
tal sensor monitoring are often organized in three modes of time, location,
and type [17]. Data in social network analysis are usually organized in three
modes of time, author, and keywords [17]. Data in network forensics are often
organized in three modes of time, source, and destination, and data in web
graph mining are commonly organized in three modes of source, destination,
and text [15].
Due to the challenges in emerging applications above, there has been a press-
ing need for more effective dimensionality reduction schemes for massive mul-
tidimensional data. Recently, interests have grown in multilinear subspace
learning (MSL) [2, 21–26], a novel approach to dimensionality reduction of
multidimensional data where the input data are represented in their natural
multidimensional form as tensors. Figure 1 shows two examples of tensor data
2
representations for a face image and a silhouette sequence. MSL has the po-
tential to learn more compact and useful representations than LSL [21,27] and
it is expected to have potential future impact in both developing new MSL
algorithms and solving problems in applications involving massive multidi-
mensional data. The research on MSL has gradually progressed from heuristic
exploration to systematic investigation [28] and both unsupervised and super-
vised MSL algorithms have been proposed in the past a few years [2, 21–26].
It should be noted that MSL belongs to tensor data analysis (or tensor-based
computation and modeling), which is more general and has a much wider
scope. Multilinear algebra, the basis of tensor data analysis, has been studied
in mathematics for several decades [29–31] and there are a number of recent
survey papers summarizing recent developments in tensor data analysis. E.g.,
Qi et al. review numerical multilinear algebra and its applications in [32]. Muti
and Bourennane [33] survey new filtering methods for multicomponent data
modelled as tensors in noise reduction for color images and multicomponent
seismic data. Acar and Yener [34] surveys unsupervised multiway data anal-
ysis for discovering structures in higher-order data sets in applications such
as chemistry, neuroscience, and social network analysis. Kolda and Bader [35]
provide an overview of higher-order tensor decompositions and their appli-
cations in psychometrics, chemometrics, signal processing, etc. These survey
papers primarily focus on unsupervised tensor data analysis through factor
decomposition. In addition, Zafeiriou [36] provides an overview of both unsu-
pervised and supervised nonnegative tensor factorization (NTF) [37, 38] with
NTF algorithms and their applications in visual representation and recogni-
tion discussed.
3
Table 1
List of Important Acronyms.
Acronym Description
This section first reviews the notations and some basic multilinear opera-
tions [30, 31, 41] that are necessary in defining the MSL problem. The im-
portant concepts of multilinear projections are then introduced, including el-
ementary multilinear projection (EMP), tensor-to-vector projection (TVP),
and tensor-to-tensor projection (TTP), and their relationships are explored.
Table 2 summarizes the important symbols used in this paper for quick refer-
ence.
4
Table 2
List of Important Notations
Notation Description
In the (input) dimensionality of the n-mode
M the number of training samples
N the order of a tensor object, the number of indices/modes
P the number of EMPs in a TVP
Pn the n-mode dimensionality in the projected space of a TTP
U(n) the nth projection matrix
(n)T
up the n-mode projection of the pth EMP
vec(·) the vectorized representation
X an input tensor sample
Y the projection of X on {U(n) }
(n)T
y(p) the projection of X on {up , n = 1, ..., N }
k · kF the Frobenius norm
2.1 Notations
As in [30, 31, 41], an N th-order tensor is denoted as: A ∈ RI1 ×I2 ×...×IN , which
is addressed by N indices in , n = 1, ..., N , with each in addressing the n-mode
of A. The n-mode product of a tensor A by a matrix U ∈ RJn ×In , denoted as
A ×n U, is a tensor with entries [30]:
X
(A ×n U)(i1 , ..., in−1 , jn , in+1 , ..., iN ) = A(i1 , ..., iN ) · U(jn , in ). (1)
in
5
(a) (b) (c) (d)
Fig. 2. Visual illustration of n-mode vectors: (a) a tensor A ∈ R8×6×4 , (b) the
1-mode vectors, (c) the 2-mode vectors, and (d) the 3-mode vectors.
The scalar product of two tensors A, B ∈ RI1 ×I2 ×...×IN is defined as:
XX X
< A, B >= ... A(i1 , i2 , ..., iN ) · B(i1 , i2 , ..., iN ) (2)
i1 i2 iN
i.e.,
A(i1 , i2 , ..., iN ) = u(1) (i1 ) · u(2) (i2 ) · ... · u(N ) (iN ) (5)
for all values of indices. Unfolding A along the n-mode is denoted as
where the column vectors of A(n) are the n-mode vectors of A. Figures 2(b),
2(c), and 2(d) illustrate the 1-mode, 2-mode, and 3-mode vectors of a tensor
A in Fig. 2(a), respectively. Figure 3 shows the 1-mode unfolding of the tensor
A in Fig. 2(a).
6
The distance between tensors A and B can be measured by the Frobenius
norm [2]:
dist(A, B) =k A − B kF . (7)
Although this is a tensor-based measure, it is equivalent to a distance measure
of corresponding vector representations, as proven in [42]. Let vec(A) be the
vector representation (vectorization) of A, then
This implies that the distance between two tensors as defined in (7) equals to
the Euclidean distance between their vectorized representations.
y = UT x = x ×1 UT . (9)
[30, 43], the tensor space can be viewed as the Kronecker product of N vector
(linear) spaces RI1 , RI2 , ..., RIN [43]. To project a tensor X in a tensor space
RI1 RI2 ... RIN to another tensor Y in a lower-dimensional tensor space
N N
7
RP1 RP2 ... RPN , where Pn ≤ In for all n, N projection matrices {U(n) ∈
N N
It can be done in N steps, where in the nth step, each n-mode vector is
projected to a lower dimension Pn by U(n) , as shown in Fig. 5(a). Figure 5(b)
demonstrates how to project a tensor in 1-mode using a 1-mode projection
matrix, which projects each 1-mode vector of the original tensor to a low-
dimensional vector.
8
(a) Projection of a tensor in all modes.
where k · k is the Euclidean norm for vectors. It can be written in the scalar
product (2) as:
y =< X , u(1) ◦ u(2) ◦ ... ◦ u(N ) > . (12)
Denote U = u(1) ◦u(2) ◦...◦u(N ) , then y =< X , U >. This multilinear projection
T T T
{u(1) , u(2) , ..., u(N ) } is named as an EMP, the projection of a tensor on a
single line (resulting a scalar), with one projection vector in each mode. Figure
6(b) illustrates an EMP of a tensor.
9
(a) Tensor-to-vector projection.
where the pth component of y is obtained from the pth EMP as:
T T T
y(p) = X ×1 u(1)
p ×2 u(2)
p ... ×N u(N
p
)
. (14)
The following examines the relationships between the three basic multilinear
projections. It is easy to verify that VVP is the special case of TTP and
TVP with N = 1, and EMP is a degenerated version of TTP with Pn = 1
for all n. On the other hand, each projected element in TTP can be viewed
as the projection of an EMP formed by taking one column from each of the
10
projection matrices. Thus, the projected tensor in TTP is effectively obtained
through N
Q
n=1 Pn interdependent EMPs, while in TVP, the P EMPs obtained
sequentially are not interdependent in general.
T T T
Furthermore, the projection using an EMP {u(1) , u(2) , ..., u(N ) } can be writ-
ten as [28]
Table 3
Number of parameters to be estimated by three multilinear projections.
10 × 10 4 400 80 40 (Pn = 2)
100 × 100 4 40,000 800 400 (Pn = 2)
100 × 100 × 100 8 8,000,000 2,400 600 (Pn = 2)
Q4
n=1 100 16 1,600,000,000 6,400 800 (Pn = 2)
11
3 The Multilinear Subspace Learning Framework
This section formulates a general MSL framework. It defines the MSL problem
in a similar way as LSL, as well as tensor and scalar scatter measures for opti-
mality criterion construction. It also outlines a typical solution and discusses
related issues.
LSL algorithms [18,44] solve for a linear projection satisfying some optimality
criteria, given a set of training samples. The problem can be formulated as
follows.
Among various LSL algorithms, principal component analysis (PCA) [19] and
linear discriminant analysis (LDA) [44] are two most widely used ones in a
broad range of applications [49, 50]. PCA is an unsupervised algorithm that
does not require labels for the training samples, while LDA is a supervised
method that makes use of class specific information. Other popular LSL al-
gorithms include independent component analysis (ICA) [51] and canonical
correlation analysis (CCA) [52].
12
Mathematically, the MSL problem can be written in a general form as
{U(n) } = arg max Φ {U(n) }, {Xm } (16)
{U(n) }
or
T T
{u(n)
p }Pp=1 = arg max
T
Φ {u(n)
p }Pp=1 , {Xm } , (17)
(n)
{up }P
p=1
1, ..., M }, respectively.
Two key components for MSL are the multilinear projection employed and the
objective criterion to be optimized. The projection to be solved can be any of
the three types of basic multilinear projections discussed in Sec. 2.3. Thus, the
well-studied LSL can be viewed as a special case of MSL where the projection
to be solved is a VVP. Thus, the focus of this paper will be on MSL through
TTP and TVP. This general formulation of MSL is important for evaluating,
comparing, and further developing MSL solutions.
In analogy to the definition of various scatters for vectorial features in LSL [44],
tensor-based and scalar-based scatters in MSL are defined here.
The n-mode total scatter matrix of these samples is then defined as:
M T
(n) X
STA = Am(n) − Ā(n) Am(n) − Ā(n) , (20)
m=1
13
RI2 ... RIN . The between-class scatter and the within-class scatter of these
N
and the n-mode within-class scatter matrix of these samples is defined as:
M T
(n) X
SWA = Am(n) − Ācm (n) Am(n) − Ācm (n) , (24)
m=1
The tensor scatters defined above are for MSL based on TTP. For MSL based
on TVP, scalar-based scatters are defined, which can be viewed as degenerated
versions of the vector-based or tensor-based scatters.
14
between-class scatter of these scalars is defined as:
C
Mc (āc − ā)2 ,
X
SBa = (27)
c=1
where
1 X
āc = am . (29)
Mc m,cm =c
While a linear projection (VVP) in LSL often has closed-form solutions, this
is not the case for TTP and TVP in MSL. Instead, these two tensor-based
projections have N sets of parameters to be solved, one in each mode, and
the solution to one set often depends on the other sets (except when N = 1,
the linear case), making their simultaneous estimation extremely difficult, if
not impossible. Therefore, a suboptimal, iterative procedure originated from
the alternating least square (ALS) algorithm [53–55] is usually employed to
solve the tensor-based projections by alternating between solving one set of
parameters (in one mode) at a time. Specifically, the parameters for each
mode are estimated in turn separately and are conditioned on the parameter
values for the other modes. At each step, by fixing the parameters in all the
modes but one mode, a new objective function depending only on the mode
left free to vary is optimized and this conditional subproblem is linear and
much simpler to solve through unfolding tensors to matrices. The parameter
estimations for each mode are obtained in this way sequentially and iteratively
until convergence. This process is described in Fig. 7(a) and also illustrated
in Fig. 7(b).
Consequently, the issues due to the iterative nature of the solution, such as ini-
tialization, the order of solving the projections, termination, and convergence,
need to be addressed. In addition, for MSL through TTP, a mechanism is of-
ten needed to determine the desired subspace dimensionality {P1 , P2 , ..., PN }.
This is because it is costly to exhaustively test the large number of possible
combinations of the N values, P1 , P2 , ..., PN , for a specific amount of dimen-
sionality reduction, especially for higher-order tensors. In contrast, the value
P is relatively easier to determine for MSL through TVP.
To end this section, Table 4 summarizes the key differences between MSL
15
Input: A set of tensor samples {Xm ∈ RI1 ×I2 ×...×IN , m = 1, ..., M }, the
desired tensor subspace dimensionality.
Algorithm:
and LSL. In the table, massive data refers to data with its dimensionality
beyond the processing power of common computational hardwares when LSL
algorithms are used, such as high-resolution face images or standard gait sil-
houette sequences. A key difference between MSL and LSL is the number of
parameters to be estimated for the same dimensionality reduction problem,
as illustrated in Table 3. Consequently, as pointed out in [22], MSL is less
susceptible to the SSS problem and overfitting compared to LSL.
16
Table 4
Linear versus multilinear subspace learning.
This section reviews several important MSL algorithms under the MSL frame-
work. Due to the fundamentality and importance of PCA and LDA, the focus
is on the multilinear extensions of these two classical linear algorithms. Figures
8 and 9 depict taxonomies for these algorithms, one for multilinear extensions
of PCA and the other for multilinear extensions of LDA, respectively, which
will be discussed in detail in the following. The historical basis of MSL will
also be described.
17
4.1 Unsupervised Multilinear Subspace Learning Algorithms
18
rent subspaces analysis (CSA) is formulated in [27] for general tensor objects
as a generalization of GLRAM for higher-order tensors. The CSA algorithm
solves for a TTP minimizing a reconstruction error metric, however, how to
determine the tensor subspace dimensionality is not addressed in this work.
Whereas GLRAM and CSA advanced unsupervised MSL, they are both for-
mulated with the objective of optimal reconstruction or approximation of ten-
sors. Therefore, they ignored an important centering step in unsupervised
subspace learning algorithms developed for recognition, such as the classical
PCA, where the data is centered first before obtaining the subspace projec-
tion. It should be pointed out that for the reconstruction or approximation
problem, centering is not essential, as the (sample) mean is the main focus
of attention [21]. However, in recognition applications where the solutions in-
volve eigenproblems, non-centering (in other words, an average different from
zero) can potentially affect the eigen-decomposition in each mode and lead to
a solution that captures the variation with respect to the origin rather than
capturing the true variation of the data (with respect to the data center) [21].
19
which is important when the underlying data factors have physical or psycho-
logical interpretation. The solution for NMPCA is developed by exploiting the
structure of the Grassmann manifold [62].
In comparison to the review above, there are much fewer unsupervised MSL
algorithms based on TVP. The tensor rank-one decomposition (TROD) al-
gorithm introduced in [64] is TVP-based and it is formulated only for image
matrices. This algorithm looks for a second-order TVP that projects an image
to a low-dimensional vector while minimizing a least-square (reconstruction)
error measure. Hence, the input data is not centered before learning. The solu-
tion of TROD relies on a heuristic procedure of successive residue calculation,
i.e., after obtaining the pth EMP, the input image is replaced by its residue.
None of the above unsupervised MSL algorithms takes into account the cor-
relations among features and shares an important property with PCA, i.e.,
zero-correlation among extracted features. It is well-known that PCA derives
uncorrelated features, which contain minimum redundancy and ensure linear
independence among features. Uncorrelated features can also greatly simplify
the subsequent classification task and they are highly desirable in recogni-
tion applications. An uncorrelated MPCA (UMPCA) algorithm is proposed
in [25], which extracts uncorrelated multilinear features through TVP while
capturing most of the variation in the original data input (Definition 4). The
UMPCA solution consists of sequential iterative steps for successive variance
maximization. The work in [25] has also derived a systematic way to deter-
mine the maximum number of uncorrelated multilinear features that can be
extracted by the method.
20
4.2 Supervised Multilinear Subspace Learning Algorithms
Like GLRAM and GPCA, the 2D LDA (2DLDA) introduced in [65] solves
for two linear transformations that project an image to a low-dimensional
matrix, but with a different objective criterion. For the input image samples,
the between-class and within-class scatter measures are defined for matrix
representations (Definition 2). A matrix-based discrimination criterion is then
defined as the scatter ratio, which is to be maximized in 2DLDA. Unlike the
unsupervised MSL algorithms reviewed above, 2DLDA does not converge over
iterations.
In this category, the first algorithm is the tensor rank-one discriminant analysis
(TR1DA) algorithm proposed in [46, 47], derived from the TROD algorithm
[64]. The TR1DA algorithm is formulated for general tensor objects and it
looks for a TVP that projects a tensor to a low-dimensional vector while
maximizing the scalar scatter difference criterion (Definition 5). Therefore, the
criterion is also dependent on the coordinate system and there is no way to
determine the optimal tuning parameter either. Furthermore, as in TROD, this
1Here, the name used when the algorithm was first proposed is adopted as it is
more commonly referred to in the literature.
21
algorithm also relies on the repeatedly-calculated residues, originally proposed
in [68] for tensor approximation. The adoption of this heuristic procedure here
lacks theoretical explanation for a discriminative criterion.
Similar to the case of unsupervised MSL, the supervised MSL algorithms dis-
cussed so far do not take the correlations among features into account and
they do not derive uncorrelated features as in the classical LDA [69, 70]. As
mentioned in Sec. 4.1.2, uncorrelated features are highly desirable in many ap-
plications [70]. An uncorrelated multilinear discriminant analysis (UMLDA)
algorithm is formulated in [26]. UMLDA aims to extract uncorrelated dis-
criminative features directly from tensorial data through solving a TVP to
maximize a scalar scatter ratio criterion (Definition 5). The solution consists
of sequential iterative processes and incorporates an adaptive regularization
procedure to enhance the performance in the small sample size scenario. Fur-
thermore, an aggregation scheme is adopted to combine differently initialized
and regularized UMLDA recognizers for enhanced generalization performance
while alleviating the regularization parameter selection problem. This exten-
sion is called regularized UMLDA with aggregation (R-UMLDA-A) [26].
Multilinear algebra, the extension of linear algebra, has been studied in math-
ematics around the middle of the 20th century [29]. It built on the concept of
tensors and developed the theory of tensor spaces.
22
iteratively projected into a lower dimensional tensor space. The application
of HOSVD truncation and the best Rank-(R1 , R2 , ..., RN ) approximation to
dimensionality reduction in ICA was discussed in [75].
The work in [30,31] led to the development of new multilinear algorithms and
the exploration of new application areas for tensor data analysis. Multilinear
analysis of image data is pioneered by the TensorFace method [76, 77], which
employs the multilinear algorithms proposed in [30, 31] to analyze the factors
involved in the formation of facial images. Similar analysis has also been done
for motion signatures [78] and gait sequences [79]. However, in these multiple
factor analysis work, input data such as images or video sequences are still
represented as vectors. These vectors are arranged into a tensor according
to multiple factors involved in their formation for subsequent analysis. Such
tensor formation needs a large number of training samples captured under
various conditions, which is often impractical and may have the missing-data
problem. Furthermore, the tensor data size is usually huge, leading to high
memory and computational demands.
In the last few years, several methods were proposed for direct learning of a
subspace from tensorial data [1,2,21,23,25–27,47,56,64,65]. Besides the MSL
algorithms reviewed above, there is also a multilinear extension of the CCA
algorithm named as tensor CCA in [80]. In addition, solutions are proposed
in [81, 82] to rearrange elements within a tensor to maximize the correlations
among n-mode vectors for better dimensionality reduction performance. Fur-
thermore, systematic treatment of this topic has appeared in [28, 42]. Besides
the multilinear extensions of LSL algorithms, multilinear extensions of lin-
ear graph-embedding algorithms were also introduced in [43, 48, 83–86], in a
similar fashion as the existing MSL algorithms reviewed in this paper.
Due to the advances in sensor and storage technology, MSL is becoming in-
creasingly popular in a wide range of application domains involving tensor-
structured data sets. This section will summarize several applications of MSL
algorithms in real-world applications, including face recognition and gait recog-
nition in biometrics, music genre classification in audio signal processing, EEG
signal classification in biomedical engineering, anomaly detection in data min-
ing, and visual content analysis in computer vision. Other MSL applications
include handwritten digit recognition [27, 57], image compression [1, 27, 64],
image/video retrieval [1, 87], and object categorization and recognition [46].
For more general tensor data applications, [35], [88], [89] and [90] are good
references.
23
5.1 Face Recognition and Gait Recognition
Face and gait are two typical physiological and behavioral biometrics, respec-
tively. Compared with other biometric traits, face and gait have the unique
property that they facilitate human recognition at a distance, which is ex-
tremely important in surveillance applications. Face recognition has a large
number of commercial security and forensic applications, including video surveil-
lance, access control, mugshot identification, and video communications [10,
91]. Gait is a person’s walking style and it is a complex spatio-temporal bio-
metric [10, 91]. The interest in gait recognition is strongly motivated by the
need for an automated human identification system at a distance in visual
surveillance and monitoring applications in security-sensitive environments,
e.g., banks, parking lots, malls, and transportation hubs such as airports and
train stations [21]. Many MSL algorithms are first applied to appearance-
based learning for face recognition [2, 27, 56, 57, 86, 92, 93] and/or gait recog-
nition [21, 23, 94–97], where the input face images or binary gait silhouette
sequences are treated as tensorial holistic patterns as shown in Fig. 1.
24
in the spatial-spectral-temporal domain. GTDA is then applied to obtain low-
dimensional tensors, from which discriminative features are selected for SVM
classification. Motor imagery experiments on three data sets demonstrate that
the proposed scheme can outperform many other existing EEG signal classifi-
cation schemes, especially when there is no prior neurophysiologic knowledge
available for EEG signal preprocessing.
25
natural multidimensional representation: tensors. This survey covers multi-
linear projections, MSL framework, typical MSL solutions, MSL taxonomies,
and MSL applications. MSL is a new field with many open issues to be ex-
amined further. The rest of this section outlines several research topics that
worth further investigation. Two main directions have been identified. One is
towards the development of new MSL solutions, while the other is towards the
exploration of new MSL applications.
Besides the applications reviewed in Sec. 5, there are a wide range of appli-
cations dealing with real-world tensor objects where MSL may be useful, as
mentioned in the Introduction of this paper. Examples include high-resolution
and 3D face detection and recognition [108–111], clustering or retrieval of im-
ages or 3D objects [8,112], space-time analysis of video sequences [11,113], and
space-time super-resolution [14]. In addition, massive streaming data are fre-
quently organized as multidimensional objects, such as those in social network
analysis, web data mining, and sensor network analysis [17]. Some tensor-based
techniques have been developed in these fields [15,16] and further investigation
of MSL for these applications under the framework presented in this paper can
be fruitful.
26
Acknowledgment
The authors would like to thank the anonymous reviewers for their insightful
comments, which have helped to improve the quality of this paper.
References
[1] J. Ye, R. Janardan, Q. Li, GPCA: An efficient dimension reduction scheme for
image compression and retrieval, in: The Tenth ACM SIGKDD Int. Conf. on
Knowledge Discovery and Data Mining, 2004, pp. 354–363.
27
[11] R. D. Green, L. Guan, Quantifying and recognizing human movement patterns
from monocular video images-part II: applications to biometrics, IEEE
Transactions on Circuits and Systems for Video Technology 14 (2) (2004)
191–198.
[13] X. He, Incremental semi-supervised subspace learning for image retrieval, in:
ACM conference on Multimedia 2004, 2004, pp. 2–8.
[15] J. Sun, D. Tao, C. Faloutsos, Beyond streams and graphs: dynamic tensor
analysis, in: Proc. the 12th ACM SIGKDD int. conf. on Knowledge discovery
and data mining, 2006, pp. 374–383.
[16] J. Sun, Y. Xie, H. Zhang, C. Faloutsos, Less is more: Sparse graph mining
with compact matrix decomposition, Statistical Analysis and Data Mining
1 (1) (2008) 6–22.
28
[25] H. Lu, K. N. Plataniotis, A. N. Venetsanopoulos, Uncorrelated multilinear
principal component analysis for unsupervised multilinear subspace learning,
IEEE Transactions on Neural Networks 20 (11) (2009) 1820–1836.
[27] D. Xu, S. Yan, L. Zhang, S. Lin, H.-J. Zhang, T. S. Huang, Reconstruction and
recognition of tensor-based objects with concurrent subspaces analysis, IEEE
Transactions on Circuits and Systems for Video Technology 18 (1) (2008) 36–
47.
[32] L. Qi, W. Sun, Y. Wang, Numerical multilinear algebra and its applications,
FRONTIERS OF MATHEMATICS IN CHINA 2 (4) (2007) 501–526.
29
[39] R. Bro, Multi-way analysis in the food industry - models, algorithms and
applications, Ph.D. thesis, University of Amsterdam, The Netherlands (1998).
URL http://www.models.kvl.dk/sites/default/files/brothesis_0.pdf
[40] A. K. Smilde, R. Bro, P. Geladi, Multi-way Analysis, John Wiley and Sons,
2004.
[41] B. W. Bader, T. G. Kolda, Algorithm 862: Matlab tensor classes for fast
algorithm prototyping, ACM Trans. on Mathematical Software 32 (4) (2006)
635–653.
[42] H. Lu, Multilinear subspace learning for face and gait recognition, Ph.D. thesis,
University of Toronto (2008).
URL https://tspace.library.utoronto.ca/handle/1807/16750
[43] X. He, D. Cai, P. Niyogi, Tensor subspace analysis, in: Advances in Neural
Information Processing Systems 18 (NIPS), 2005.
[47] D. Tao, X. Li, X. Wu, S. J. Maybank, Elapsed time in human gait recognition:
A new approach, in: Proc. IEEE Int. Conf. on Acoustics, Speech and Signal
Processing, Vol. 2, 2006, pp. 177–180.
30
[54] R. A. Harshman, Foundations of the parafac procedure: Models and conditions
for an “explanatory” multi-modal factor analysis, UCLA Working Papers in
Phonetics 16 (1970) 1–84.
[64] A. Shashua, A. Levin, Linear image coding for regression and classification
using the tensor-rank principle, in: Proc. IEEE Conference on Computer Vision
and Pattern Recognition, Vol. I, 2001, pp. 42–49.
[66] Q. Liu, X. Tang, H. Lu, S. Ma, Face recognition using kernel scatter-difference-
based discriminant analysis, IEEE Transactions on Neural Networks 17 (4)
(2006) 1081–1085.
31
[68] T. G. Kolda, Orthogonal tensor decompositions, SIAM Journal of Matrix
Analysis and Applications 23 (1) (2001) 243–255.
[81] D. Xu, S. Yan, S. Lin, T. S. Huang, S.-F. Chang, Enhancing bilinear subspace
learning by element rearrangement, IEEE Transactions on Pattern Analysis
and Machine Intelligence 31 (10) (2009) 1913–1920.
32
[83] J. Zhang, J. Pu, C. Chen, R. Fleischer, Low-resolution gait recognition, IEEE
Transactions on Systems, Man, and Cybernetics—Part B: Cybernetics 40 (4)
(2010) 986–996.
[86] D. Xu, S. Lin, S. Yan, X. Tang, Rank-one projections with adaptive margins for
face recognition, IEEE Transactions on Systems, Man, and Cybernetics—Part
B: Cybernetics 37 (5) (2007) 1226–1236.
[87] X. Gao, X. Li, J. Feng, D. Tao, Shot-based video retrieval with optical flow
tensor and HMMs, Pattern Recognition Letters 30 (2) (2010) 140–147.
[90] Workshop on Algorithms for Modern Massive Data Sets (2010, 2008, 2006).
URL http://www.stanford.edu/group/mmds/
33
[97] H. Lu, K. N. Plataniotis, A. N. Venetsanopoulos, Boosting discriminant
learners for gait recognition using mpca features, EURASIP Journal
on Image and Video Processing 2009, article ID 713183, 11 pages,
doi:10.1155/2009/713183.
[99] J. Wen, X. Gao, Y. Yuan, D. Tao, J. Li, Incremental tensor biased discriminant
analysis: A new color-based visual tracking method, Neurocomputing 73 (4-6)
(2010) 827–839.
[101] X. Gao, Y. Yang, D. Tao, X. Li, Discriminative optical flow tensor for video
semantic analysis, Computer Vision and Image Understanding 113 (3) (2009)
372–383.
[109] S. Z. Li, C. Zhao, M. Ao, Z. Lei, Learning to fuse 3D+2D based face recognition
at both feature and decision levels, in: Proc. IEEE Int. Workshop on Analysis
and Modeling of Faces and Gestures, 2005, pp. 43–53.
34
[111] P. J. Phillips, P. Flynn, T. Scruggs, K. Bowyer, J. Chang, K. Hoffman,
J. Marques, J. Min, W. Worek, Overview of the face recognition grand
challenge, in: Proc. IEEE Conference on Computer Vision and Pattern
Recognition, Vol. 1, 2005, pp. 947–954.
35