• No results found

Perturbation theory for the tensor rank decomposition

N/A
N/A
Protected

Academic year: 2021

Share "Perturbation theory for the tensor rank decomposition"

Copied!
42
0
0

Loading.... (view fulltext now)

Full text

(1)

Perturbation theory for

the tensor rank decomposition

Nick Vannieuwenhoven

FWO / KU Leuven

(2)

Introduction Conditioning Motivation Definition Examples Conclusions

Overview

1 Introduction 2 Conditioning 3 Motivation 4 Definition 5 Examples 6 Conclusions

(3)

Overview

1 Introduction 2 Conditioning 3 Motivation 4 Definition 5 Examples 6 Conclusions

(4)

Introduction Conditioning Motivation Definition Examples Conclusions

Tensor rank decomposition

Hitchcock (1927) introduced the tensorrank decomposition:1

A= r X i=1 a1i ⊗ · · · ⊗adi A = +· · ·+

Therankof a tensor is the minimum number of rank-1 tensors of which it is a linear combination.

1

(5)

Identifiability

A rank-1 tensor isuniquely determined up to scaling:

a⊗b⊗c= (αa)⊗(βb)⊗(α−1β−1c).

Kruskal (1977) proved thatthe rank-1 terms appearing in

A=

r X

i=1

a1i ⊗a2i ⊗ · · · ⊗adi

(6)

Introduction Conditioning Motivation Definition Examples Conclusions

Generic identifiability

It is expected2 by [BCO13, COV14] that a random complex tensor rank decomposition A= r X i=1 a1i ⊗a2i ⊗ · · · ⊗adi,

of strictly subgeneric rank, i.e.,

r ≤ n1n2· · ·nd n1+· · ·+nd−d + 1 −1,

isidentifiable with probability 1, provided that it is not one of the exceptional cases where (n1,n2, . . . ,nd) is

(n1,n2), or

(4,4,3), (4,4,4), (6,6,3), (n,n,2,2), (2,2,2,2,2), or

n1>Qdi=2ni −Pdi=2(ni−1) (unbalanced).

2

(7)

Generic symmetric identifiability

It was proved by [COV16] that a random complex symmetric tensor rank decomposition

A=

r X

i=1

ai ⊗ai⊗ · · · ⊗ai,

of strictly subgeneric rank, i.e.,

r ≤ & n−1+d d n ' −1,

isidentifiable with probability 1, provided that the space is not one of the following exceptional cases:

symmetric matrices, i.e., S2Cn, S3C6,

S4C4, or S6C3.

(8)

Introduction Conditioning Motivation Definition Examples Conclusions

Overview

1 Introduction 2 Conditioning 3 Motivation 4 Definition 5 Examples 6 Conclusions

(9)

Perturbations and conditioning

In applications it is uncommon to work with the “true” tensorP. Usually we only have somePb for which

kP −Pkb =

with small.

The error can originate from many sources: measurement errors,

model errors, and

(10)

Introduction Conditioning Motivation Definition Examples Conclusions

Perturbations and conditioning

A true decomposition P = r X i=1 a1i ⊗a2i ⊗ · · · ⊗adi

is nice, but I only knowPb. I can compute an approximation

b P ≈ r X i=1 b a1ibad2 ⊗ · · · ⊗badi ≈ P

but what does it tell me aboutP?

Is P’s decomposition unique?

Are the terms in Pb’s decomposition related to those ofP?

(11)

Condition number

The condition number gives the first-order answer to the question: How sensitive isf to perturbations to the input?

• x • y • f(x) • f(y) κ

(12)

Introduction Conditioning Motivation Definition Examples Conclusions

Condition number

Definition

Therelative condition numberof a functionf :X →Y atx ∈X is

κ= lim

→0k∆xkmaxβ=

kf(x)−f(x+ ∆x)kα/kf(x)kα k∆xkβ/kxkβ ,

for some normsk · kα andk · kβ.

The condition number is the property of aproblem, not of a particular algorithm solving the problem.

(13)

Condition number

Definition

Therelative condition numberof a functionf :X →Y atx ∈X is

κ= lim

→0k∆xkmaxβ=

kf(x)−f(x+ ∆x)kα/kf(x)kα k∆xkβ/kxkβ ,

for some normsk · kα andk · kβ.

The condition number is the property of aproblem, not of a particular algorithm solving the problem.

(14)

Introduction Conditioning Motivation Definition Examples Conclusions

Overview

1 Introduction 2 Conditioning 3 Motivation 4 Definition 5 Examples 6 Conclusions

(15)

An experiment in Tensorlab

LetA= a1 a2 a3 a4 a5,B=b1 b2 b3 b4 b5, and C= c+c1 c+c2 c+c3 c+c3 c+c5 .

be 5×5 matrices withai,bi, andci random vectors.

Consider a sequence of tensors

T = 5 X i=1 ai ⊗bi ⊗(c+ci) →0 −−→ 5 X i=1 ai ⊗bi ⊗c. Then, T is 5-identifiable if 6= 0, while

(16)

Introduction Conditioning Motivation Definition Examples Conclusions

An experiment in Tensorlab

Let’s compute the unique decomposition ofT in Tensorlab using

an algebraic algorithm:

% Generate tensor.

T_eps = cpdgen({A,B,C_eps}); % Compute decomposition. [U, out] = cpd_gevd(T_eps,5); % Measure relative backward error.

(17)
(18)

Introduction Conditioning Motivation Definition Examples Conclusions

An experiment in Tensorlab

A small backward error is but the beginning in a data-analysis application. We need a smallforward error!

LetAb,Bb,Cb be the computed factor matrices of T. Then, we

really seek small errors

kA−AbPk kAk , kB−BbPk kBk , and kC−CbPk kCk ,

(19)

An experiment in Tensorlab

In our toy problem, we can compute the forward error using:

% [U, out] = cpd_gevd(T_eps,5); U_true = {A,B,C_eps};

(20)

Introduction Conditioning Motivation Definition Examples Conclusions

(21)

Introduction Conditioning Motivation Definition Examples Conclusions

An experiment in Tensorlab

Except for toy problems, we don’t know the true decomposition! So how do we assess the forward error??

By the definition of the condition number, we have:

(relative) forward error

(22)

Introduction Conditioning Motivation Definition Examples Conclusions

An experiment in Tensorlab

Except for toy problems, we don’t know the true decomposition! So how do we assess the forward error??

The condition number

By the definition of the condition number, we have:

(relative) forward error

(23)
(24)

Introduction Conditioning Motivation Definition Examples Conclusions

Overview

1 Introduction 2 Conditioning 3 Motivation 4 Definition 5 Examples 6 Conclusions

(25)

Terracini’s matrix

For every

pi =a1i ⊗a2i ⊗ · · · ⊗adi ∈Cn1n2···nd,

we define the matrix

Ti = In1⊗a 2 i ⊗ · · · ⊗adi · · · a1i ⊗ · · · ⊗a d−1 i ⊗Ind .

Then,Terracini’s matrixis given by

Tp1,...,pr =

T1 T2 · · · Tr

(26)

Introduction Conditioning Motivation Definition Examples Conclusions

Terracini’s matrix is singular

Working in affine space,

Terracini’s matrix is neverof maximal rank r(n1+n2+· · ·+nd).

Let us write the singular values ofTp1,...,pr as

σ1 ≥σ2≥ · · · ≥σN ≥0≥0≥ · · · ≥0

withN =r(n1+· · ·+nd−d + 1). Then, I claim that

σN−1

(27)

A condition number

Theorem (V, 2016)

Let Tp1,...,pr be Terracini’s matrix associated with the rank-1

tensors pi ∈Cn1⊗ · · · ⊗Cnd. Let N =r(n

1+· · ·+nd−d+ 1). Then,

κA =σN−1

is an absolute condition number of the rank decomposition problem atA=p1+p2+· · ·+pr.

(28)

Introduction Conditioning Motivation Definition Examples Conclusions

Overview

1 Introduction 2 Conditioning 3 Motivation 4 Definition 5 Examples 6 Conclusions

(29)

A well-conditioned example

Let’s take the “Amino acids” data set from

http://www.models.life.ku.dk/nwaydata, which is “a nice set for fitting PARAFAC.”

“The data set consists of five simple laboratory-made samples. Each sample contains different amounts of tyrosine,tryptophanandphenylalanine dissolved in phosphate buffered water. The samples were measured by fluorescence (excitation 250-300 nm, emission 250-450 nm, 1 nm intervals) on a spectrofluorometer. The array to be decomposed is hence 5×51×201. Ideally, these data should be describable with three PARAFAC components.”

(30)

Introduction Conditioning Motivation Definition Examples Conclusions

A well-conditioned example

Armed with the condition number, we can quantify precisely how “nice” this tensor is.

>> load ’real_data/amino.mat’; T = X; % Compute approximate decomposition. >> U = cpd(T,3);

% Compute condition number.

>> condNumber = cpdcondest(U,’relative’) ans = 3.1159e+00

The condition number isO(1), so the backward error is hardly amplified!

(31)

A well-conditioned example

% Compute backward error.

>> relBErr = frob(T - cpdgen(U))/frob(T); ans = 2.5049e-02

% Estimate forward error.

>> relFErr = relBErr * condNumber ans = 7.8049e-02

If you can bound the measurement errors byrelMErr, then one can even obtain the forward error w.r.t. the true decomposition:

% Estimate forward error w.r.t. true decomposition. >> relFErr = (relBErr + relMErr) * condNumber

(32)

Introduction Conditioning Motivation Definition Examples Conclusions

(33)
(34)

Introduction Conditioning Motivation Definition Examples Conclusions

(35)

Ill-posedness and ill-conditioning

The classic example from [dSL08] is the rank-3 tensor

u⊗u⊗v+u⊗v⊗u+v⊗u⊗u,

which is a limit of identifiable rank-2 tensors:

lim →0 1 u⊗u⊗u− 1 (u+v)⊗(u+v)⊗(u+v) .

[V16] proves that “a sequence of diverging components” entails that the relative condition numberκ→ ∞.

Formally: As you move towards an open part of the boundary of ther-secant variety of a Segre variety, the relative condition number becomes unbounded.

(36)

Introduction Conditioning Motivation Definition Examples Conclusions

(37)

Overview

1 Introduction 2 Conditioning 3 Motivation 4 Definition 5 Examples 6 Conclusions

(38)

Introduction Conditioning Motivation Definition Examples Conclusions

Conclusions

Take-away messages:

Tensors are conjectured to be identifiable [COV14].

Symmetric tensors were proved to be identifiable [COV16]. Forward errors matter.

The condition number multiplied with the backward error bounds the forward error “to first order.”

The condition number of a decomposition can be computed practically.

(39)
(40)

Introduction Conditioning Motivation Definition Examples Conclusions

Further reading

Main references:

Vannieuwenhoven, A condition number for the tensor rank decomposition, 2016. (In preparation)

Chiantini, Ottaviani, and V, An algorithm for generic and low-rank specific identifiability of complex tensors, SIAM J. Matrix Anal. Appl., 2014.

(41)

References

General introduction

[H1927] Hitchcock,The expression of a tensor or a polyadic as a sum of products, J. Math. Phys., 1927.

[K1977] Kruskal,Three-way arrays: rank and uniqueness of trilinear decompositions, with application to arithmetic complexity and statistics, Lin. Alg. Appl., 1977.

Conditioning

[BC13] B¨urgisser and Cucker, Condition: The Geometry of Numerical Algorithms, Springer, 2013.

[dSL08] de Silva and Lim,Tensor rank and the ill-posedness of the best low-rank approximation problem, SIAM J. Matrix Anal. Appl., 2008.

[V16] Vannieuwenhoven,A condition number for the tensor rank decomposition. (In preparation)

(42)

Introduction Conditioning Motivation Definition Examples Conclusions

References

Generic identifiability

[BCO13] Bocci, Chiantini, and Ottaviani,Refined methods for the identifiability of tensors, Ann. Mat. Pura Appl., 2013.

[CO12] Chiantini and Ottaviani,On generic identifiability of 3-tensors of small rank, SIAM J. Matrix Anal. Appl., 2013. [COV14] Chiantini, Ottaviani, and Vannieuwenhoven,An algorithm for generic and low-rank specific identifiability of complex tensors, SIAM J. Matrix Anal. Appl., 2014.

[COV16] Chiantini, Ottaviani, and Vannieuwenhoven,On generic identifiability of symmetric tensors of subgeneric rank, Trans. Amer. Math. Soc., 2016. (Accepted)

[DdL15] Domanov and De Lathauwer,Generic uniqueness

conditions for the canonical polyadic decomposition and INDSCAL, SIAM J. Matrix Anal. Appl., 2015.

[HOOS15] Hauenstein, Oeding, Ottaviani, and Sommese,Homotopy techniques for tensor decomposition and perfect identifiability, arXiv, 2015.

References

Related documents

The segment moving heuristic has some important advantages: (i) it can be used for any rectangular layout, (ii) one iterative step for all maximal segments takes O (n) time, (iii)

In this part, we analyze the optimal solution of the proposed system under the following three assumptions, which will provide an optimal bound to evaluate future solutions able

Firstly, since the dynamics is derived from a first principles treatment of thermal field theory, which drives inflation from the natural dynamics of a scalar field, slow roll is

In view of these issues, the following objectives are identified in the present study: verification of the nurs- ing workload required by cancer patients undergoing palliative

Superconducting magnet costs (M$) versus induction times field volume (T-m -3 ) for solenoid magnets (closed circles), dipole and Quadruple magnets (open squares) and toroid

Factors that could cause MPLX's or MWE's actual results to differ materially from those implied in the forward-looking statements include: risk that the synergies from the

6) Clear discussion chapter setting out the main findings of the dissertation linking your literature reviews with the research findings so that a clear theme can be identified

The DNA templated metal sulfide nanowire gas sensor showed a generally fast response and recovery time especially for room temperature gas sensing even though less than