Two accelerated optimization algorithms are presented for computing
approximate Tucker tensor decompositions (ATDs). The first is a
nonlinearly preconditioned conjugate gradient (NPCG) algorithm, wherein a
nonlinear preconditioner generates a direction replacing the gradient in
the nonlinear conjugate gradient iteration. The second is a nonlinear
GMRES (N-GMRES) algorithm, in which a linear combination of iterates
generated by a nonlinear preconditioner is minimized to produce an
improved search direction. The Euclidean versions of these methods are
extended to the manifold setting, where optimization on Grassmann
manifolds is used to handle orthonormality constraints and to allow
isolated minimizers. The higher order orthogonal iteration (HOOI), the
workhorse algorithm for computing ATDs, is used as the nonlinear
preconditioner in NPCG and N-GMRES. Four options are provided for the
update parameter
in NPCG. Two strategies for approximating the
Hessian operator applied to a vector are provided for N-GMRES. NPCG and
N-GMRES are compared to HOOI, NCG, limited memory BFGS, and a manifold
trust region algorithm using synthetic data and real life tensor data
arising from handwritten digit recognition. Numerical results show that
all four NPCG variants and N-GMRES using a difference of gradients
Hessian approximation accelerate HOOI significantly for large tensors,
noisy data, and when high accuracy results are required. For these
problems, the proposed methods converge faster and more robustly than
HOOI and the state-of-the-art methods considered.