nonnegative matrix and tensor factorizations: models ...a2 linear mixing model. a3 sources are...

53
Nonnegative Matrix and Tensor Factorizations: Models, Algorithms and Applications Andersen Man Shun Ang Supervisor: Nicolas Gillis Homepage: angms.science October 14, 2020

Upload: others

Post on 25-Dec-2020

9 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

Nonnegative Matrix and Tensor Factorizations:Models, Algorithms and Applications

Andersen Man Shun AngSupervisor: Nicolas Gillis

Homepage: angms.science

October 14, 2020

Page 2: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

Overview

1 Introduction

2 Part I: Minimum volume NMF

3 Part II: NuMF

4 Part III: HER

5 Summary of contributions

2 / 53

Page 3: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

About myself, and acknowledgments

I spent 9 years learning how to know you don’t know.

I acknowledge:

I My boss Nicolas Gillis

I Jury membersI Xavier Siebert (UMONS), Thierry Dutoit (UMONS), Laurent Jacques

(UCL), Francois Glineur (UCL), Lieven De Lathauwer (KUL)

I Colleagues and friendsNicolas, Pierre, Hien, Arnaud, Valentin, Maryam, Tim, FrancoisPunit, Jeremy, Junjun, Christophe, Flavia

3 / 53

Page 4: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

About this “talk”

I 3 main parts to cover the 8 chapters of the thesis.I Ch1 IntroductionI Ch2 NMF with minimum volume: minvol NMFI Ch3 Minvol NMF on hyperspectral unmixingI Ch4 Minvol NMF on audio source separationI Ch5 NMF with unimodality: NuMFI Ch6 Tensor algebra and factorizationI Ch7 Heuristic extrapolation with restartsI Ch8 Conclusion

I Highly compressed.→ details in the thesis

I Only core ideas and eye-catching items.

4 / 53

Page 5: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

Overview

1 Introduction

2 Part I: Minimum volume NMF

3 Part II: NuMF

4 Part III: HER

5 Summary of contributions

5 / 53

Page 6: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

(New) It is cool.6 / 53

Page 7: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

Thunder-fast review of NMF

7 / 53

Page 8: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

Thunder-fast review of NMF

8 / 53

Page 9: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

NMF

I Numerical optimization

minimizeW,H

1

2‖M−WH‖2F subject to W ≥ 0 and H ≥ 0.

I Regularized model

minW,H

1

2‖M−WH‖2F+g(W) s.t. W ≥ 0,H ≥ 0 and other constraints.

I Notation: hide W ≥ 0, H ≥ 0.

9 / 53

Page 10: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

NMF

I Numerical optimization

minimizeW,H

1

2‖M−WH‖2F subject to W ≥ 0 and H ≥ 0.

I Regularized model

minW,H

1

2‖M−WH‖2F+g(W) s.t. W ≥ 0,H ≥ 0 and other constraints.

I Notation: hide W ≥ 0, H ≥ 0.

10 / 53

Page 11: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

The details I skipped

I Popularity of NMF in research

I The nonnegative rank

I How NMF is used in application

I How NMF problems are solved

I HALS

I Solution space of NMF

I NMF is NP-hard

I Separable NMF

I How to solve Separable NMF

I Successive Projection Algorithm

... see chapter 1 of the thesis!

11 / 53

Page 12: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

Geometry of M = WH

00.2

Coordinate 1

0.40.6

NMF

0.811

0.8

0.6

Coordinate 2

0.4

0.2

0.6

0.8

0

0.2

0.4

1

0

Coor

din

ate

3

Columns of M

Columns of W

Columns of I3

00.2

Coordinate 1

0.4

Separable NMF

0.60.8

11

0.8

0.6

Coordinate 2

0.4

0.2

0.8

0.6

0.4

1

0.2

0

0

Coor

din

ate

3

Figure: Nested cone geometry of NMF: blue cone ⊆ red cone ⊆ green cone.

12 / 53

Page 13: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

Minimum volume NMF

minW,H

1

2‖M−WH‖2F + λV(W) subject to some constraints.

13 / 53

Page 14: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

What is “volume”?

I Lemma 2.1.1 (p.17 in thesis) If W ∈ Rm×r+ is full rank, then√det(W>W)

is the volume of conv([0 W]

)in the column space of W, up to a

constant factor.

Proof idea: Gram-Schmidt orthogonalization and SVD.

I Figure illustration.

14 / 53

Page 15: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

What is “volume”?

det(W>W) = ‖w1‖22 ·∥∥∥P⊥1 w2

∥∥∥22

∥∥∥P⊥1,2w3

∥∥∥22. . .∥∥∥P⊥1,...,r−1wr

∥∥∥22,

where P⊥a = I− aa>

‖a‖22is the projector on span⊥(a),

P⊥1 = P⊥a1, a1 = w1

P⊥1,2 = P⊥1 P⊥a2, a2 = P⊥1 w2

P⊥1,2,3 = P⊥1,2P⊥a3, a3 = P⊥1,2w3

......

P⊥1,...,r−1 = P⊥1,2,...,r−2P⊥ar−1

, ar−1 = P⊥1,2,...,r−2wr−1.

15 / 53

Page 16: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

Some volume functions for minW,H

1

2‖M−WH‖2F + λV(W)

Name Definition In g ◦ σ(W)

Determinant (det) Vdet(W) = det(W>W

) r∏i=1

σ2i

log-determinant (logdet) Vlogdet(W) = log det(W>W + δIr

) r∑i=1

log(σ2i + δ

)Frobenius norm squared VF(W) = ‖W‖2F

r∑i=1

σ2i

Nuclear norm V∗(W) = ‖W‖∗r∑i=1

σi

Smooth Schatten-p norm Vp,δ(W) = Tr(W>W + δIr

) p2

r∑i=1

(σ2i + δ)

p2

16 / 53

Page 17: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

New theory developed in the thesis:Identifiability of a minvol (N)MF

I Theorem 2.1.1 Let M = WH where H ∈ Rr×n+ satisfies the SSC,W ∈ Rm×r satisfies W>1m = 1r and rank (M) = r. Then the(exact) solution of

argminW,H

Vdet(W) % Det-volume

s.t. M = WH % Matrix factorizationH ≥ 0 % NonnegativityW>1m = 1r % Normalization of col. of W

is essentially unique.

I Significance: justify the use of minvol NMF in application.

17 / 53

Page 18: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

Illustration

http://angms.science/eg_SNPA_ini.gif

18 / 53

Page 19: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

The details I skipped

I Details of the formulation of the minvol NMF

I Motivation of minvol NMF

I Details on the volume regularizer

I Comparing the volume regularizer

I Parameter tuning

I Rank deficiency case

I The proof of the identifiability theorem

I How to solve minvol NMF

I Computational cost of the algorithm

I Convergence analysis of the algorithm

... see chapter 2 of the thesis!

19 / 53

Page 20: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

Hyperspectral Unmixing

Band number20 40 60 80 100 120 140 160 180

Re.ectance

0

0.1

0.2

0.3

0.4

0.5

0.6

Endmember spectral pro-les of the Jasper Ridge HSI data

Vegetation

Water

Dirt

Road

20 / 53

Page 21: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

Minvol NMF on Hyperspectral UnmixingImage w.r.t. wavelength `5'

x-coordinate20 60 100

y-coordinate 20

40

60

80

100

Vectorized image

Spatial coordinate (x and y)3000 6000 9000

Data matrix M

Spatial coordinate3000 6000 9000

Wavelength

50

100

150

H

Spatial coordinate3000 6000 9000

2

4

One row of H

x-coordinate20 60 100

y-coordinate 20

40

60

80

100

W

2 4

Wavelength

50

100

150

Wavelength 0

100

2000 0.5

Plot of W

NMF

Matricization

Vectorization

h3

w1

hj

21 / 53

Page 22: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

22 / 53

Page 23: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

50 100 1500

0.5

1Vegetation type-1

50 100 1500

0.5

1Dirt type-1

50 100 1500

0.5

1Vegetation type-2

50 100 1500

0.5

1Vegetation type-3

50 100 1500

0.5

1Dirt type-2

50 100 1500

0.5

1Water

50 100 1500

0.5

1Vegetation type-4

50 100 1500

0.5

1Vegetation

50 100 1500

0.5

1Dirt

50 100 1500

0.5

1Water

50 100 1500

0.5

1Ref. Vegetation

50 100 1500

0.5

1Ref. Dirt

50 100 1500

0.5

1Ref. Water

23 / 53

Page 24: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

The details I skipped

I Details of the hyperspectral image

I Pure-pixel assumption

I Performance metrics for ranking algorithms

I Parameter tuning and experiment setup

I Details of many experimental resultsI minvol NMF � SPA, MVC-NMF and RVolMin on synthetic and real

datasets.I minvol NMF with Vdet and Vlogdet are the top two methods among the

tested methods.

... see chapter 3 of the thesis!

24 / 53

Page 25: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

Audio source separation

4

4&

E4

Ma

D4

-ry

C4

had

D4

a

E4

lit-

E4

tle

E4

lamb

œœ

œœ

œ œ ˙

-4

-3

-2

-1

0

1

-3

-2

-1

0

1

2

3

0 1 2 3 4

E4C4D4Hammer noise

redundant component

0 0.6 1.2 1.8 2.4 3 3.6 4.2

C4

Hammer noiseRedundant component

D4

E4

25 / 53

Page 26: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

Illustration

https://www.youtube.com/watch?v=1BrpxvpghKQ

26 / 53

Page 27: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

Correct identification of the note sequence

11

7

3

2

5

4

8

6

1

2

3

4

5 8

7

6

2 21

3

4

5 8

7

6 3

4

5 8

7

6

1

27 / 53

Page 28: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

How does it work?

X(f,t)x(t) Ψ Amplitude V(f,t)

Phase

w1h1

NMF

[WH](f,t)

wrhr

y1(t)

yr(t) Ψ-1

y(t)

+Ψ-1

...... ...

...

Time domain Time-frequency domain

Y1(f,t)

Yr(f,t)

...

Complex-valued Real-valued

Input

Estimate

And the data fitting term is changed from1

2‖M−WH‖2F to the

β-divergence.28 / 53

Page 29: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

Assumptions on applying minvol NMF on audio data

On using minvol NMF to estimate the source s:

A1 No time delay in the mixing.

A2 Linear mixing model.

A3 Sources are balanced.

A4 Additive mixture V =∑

i |Si|.

A5 |Si| are well approximated by nonnegative rank-1 matrices.

A6 Reconstructed components are consistent.

A7 The data has no outlier.

Violating any one of these assumptions leads to errors, or ... new researchopportunities!

29 / 53

Page 30: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

On the consistency assumption

𝒙

ℝ𝑇 ℂ𝐹×𝑇′

𝐗 = 𝐕exp 𝑖𝚽

Time domain Time-frequency domain

𝚿

𝐘 = 𝐖1𝐇1exp 𝑖𝚽

𝐙 = 𝐖2𝐇2exp 𝑖𝚽

𝐍𝐌𝐅

𝐍𝐌𝐅

𝒚

𝒛

𝚿−1

𝚿−1

Figure: Illustrating the consistency issue of decomposing a time-frequency matrix.Here the NMF result W1H1 is consistent as the matrix Y while the resultW2H2 is not. 30 / 53

Page 31: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

The details I skipped

I Details of the audio source separation process

I Time frequency transform

I β-divergence and β-divergence NMF

I Assumptions on applying NMF on audio data

I Parameter tuning and experiment setup

I Details of experimental results

... see chapter 4 of the thesis!

31 / 53

Page 32: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

Overview

1 Introduction

2 Part I: Minimum volume NMF

3 Part II: NuMF

4 Part III: HER

5 Summary of contributions

32 / 53

Page 33: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

Nonnegative unimodality

I A vector x is nonnegative unimodal ∃p ∈ [m] such that

0 ≤ x1 ≤ x2 ≤ · · · ≤ xp and xp ≥ xp+1 ≥ · · · ≥ xm ≥ 0.

Um+ : the set of Nu vectors in RmUm,p+ : the set of Nu vectors in Um+ with tonicity change at p.

I Examples

33 / 53

Page 34: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

Characterizing the Nu set

x ∈ Um+︸ ︷︷ ︸“x is unimodal”

⇐⇒ ∃p s.t. x ∈ Um,p+ ∪ Um,p+1+︸ ︷︷ ︸

a convex set

⇐⇒

0 ≤ x1x1 ≤ x2

...xp−1 ≤ xpxp+1 ≥ xp+2

...xm−1 ≥ xmxm ≥ 0︸ ︷︷ ︸

Union of two systemsof monic inequalities.

34 / 53

Page 35: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

Characterizing the Nu set

0 ≤ x1x1 ≤ x2

...xp−1 ≤ xpxp+1 ≥ xp+2

...xm−1 ≥ xmxm ≥ 0

⇐⇒ Upx ≥ 0

Up =

1−1 1

. . .. . .

−1 1

p×p︸ ︷︷ ︸

D

0p×(m−p)

0(m−p)×p

1 −1

. . .. . .

1 −11

(m−p)×(m−p)︸ ︷︷ ︸

D

∈ {−1, 0, 1}m×m,

35 / 53

Page 36: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

NuMF

minW,H

1

2‖M−WH‖2F

subject to Upjwj ≥ 0 for all j ∈ [r],

w>j 1m = 1 for all j ∈ [r],

hi ≥ 0 for all i ∈ [r],

How to solve it?

I There are r integer unknowns p1, . . . , pr

I Idea to solve it: brute force

I Improvement: accelerated projected gradient, peak detection,multi-grid

36 / 53

Page 37: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

New theory developed in the thesis:Restriction operator preserves Nu

I Theorem 5.2.1 Let x ∈ Um+ and R ∈ Rm×m′ with the structure

R(a, b) =

a b

b a b. . .

. . .. . .

b a b

b a

, a > 0, b > 0, a+ 2b = 1,

then y = Rx ∈ Um′+ . Furthermore, if x ∈ Um,p+ where p is even, then

y ∈ Um,py+ with py ∈{ p

2− 1,

p

2,p

2+ 1}

.

I Significance: justify the use of Multi-grid as a dimension reductionstep in solving NuMF.

37 / 53

Page 38: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

New theory developed in the thesis:3 theorems related to identifiability of NuMF

I Theorem 5.3.1 Informally, if wi are Nu and have strictly disjointsupport, then the sol. of NuMF is essentially unique.I Specialty of the theorem: it works with n ≥ 1, even if r > n.

I Theorem 5.3.2 Informally, if wi are Nu and have strictly disjointsupport, H satisfies the independent sensing condition, then the(exact) sol. of NuMF is essentially unique.

I Theorem 5.3.3 Informally, given x,y ∈ Um+ s.t. supp(x)*+supp(y).

If x, y can be generated by two non-zero Nu vectors u, v asx = au+ bv and y = cu+ dv, then the only possibilities are eitheru = x, v = y or u = y, v = x.

38 / 53

AngMSAndersen
AngMSAndersen
adjacent
Page 39: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

39 / 53

Page 40: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

40 / 53

Page 41: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

The details I skipped

I Details of the accelerated projected gradient

I Details of the peak detection

I Details of multi-grid

I Details of the identifiability results

... see chapter 5 of the thesis!

41 / 53

Page 42: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

Overview

1 Introduction

2 Part I: Minimum volume NMF

3 Part II: NuMF

4 Part III: HER

5 Summary of contributions

42 / 53

Page 43: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

HER

I A framework for accelerating algo. for solving NMF, NTF and CPD.

a1

1

1

XXX a

a

+ … +ar

ar

ar

(1)

(2)

(3)

(1)

(3)

(2)

=

I Key ideasI Extrapolation with restart.I Cheap computation by making use of computed component in the

update.

I Many numerical evidences on the effectiveness of HER.

43 / 53

Page 44: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

Extrapolation? Why?

55

10

10

10

1520

0 2 4-1

0

1

2

3

4

5

2

2

4

4

6

6

8

8

8

10

10

10

12

12

12

1416

0 2 4-1

0

1

2

3

4

5

2

2

4

4

6

6

6

8

8

8

10

10

10

12

12

12 14 16

0 2 4-1

0

1

2

3

4

5

5

5

5

10

10

10 15 20 25

0 2 4-1

0

1

2

3

4

5

44 / 53

Page 45: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

45 / 53

Page 46: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

k-1

Up A1

Ex A1

Pr A1

Up A2

Ex A2

Pr A2

Up AN

Ex AN

Pr AN. . .

k

46 / 53

Page 47: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

Results

See thesis:Chapter 7, p.113- p.125.

47 / 53

Page 48: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

The details I skipped

I The discussion of HER in the original form for solving NMF problem.

I The details of HER on NMF and NTF problems.

I Other similar algorithms.

... see chapter 7 of the thesis!

48 / 53

Page 49: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

Overview

1 Introduction

2 Part I: Minimum volume NMF

3 Part II: NuMF

4 Part III: HER

5 Summary of contributions

49 / 53

Page 50: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

Summary of contributions: the modeling aspect

I Minvol NMFI It generalizes Separable NMF, an important class of NMF.

I Minvol NMF with Vdet is identifiable under the SSC condition.

I Minvol NMF is empirically superior than other approaches.

I Minvol NMF with nuclear norm: new model.

I NuMFI Proposed a brute-force heuristic to solve it, with acceleration by APG,

peak detection and a dimension reduction step based on a multi-gridmethod.

I The multi-grid method preserves unimodality.

I 3 identifiability theorems of NuMF in 3 special cases.

50 / 53

Page 51: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

Summary of contributions: the algorithm aspectI A generic framework named HER on accelerating both exact and

inexact BCD type of algorithms in solving NMF, NTF and CPD.

I Many empirical evidences on the effectiveness of HER.

I The benefits of HER:I it can accelerate any BCD algorithm

I it extrapolates the variable sequence without increasing theper-iteration computational cost of the algorithm

I the auxiliary extrapolation sequence it produces is always feasible.

I The main shortcomings of HER are: no theoretical convergenceguarantee; requires parameter tuning.

I Apart from HER, we provided efficient algorithms for solving minvolNMF and NuMF.

51 / 53

Page 52: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

Summary of contributions: the application aspect

I GeographyHyperspectral unmixing in remote sensing.

I MusicAudio blind source separation.

I ChemistryChromatography - mass spectrometry

52 / 53

Page 53: Nonnegative Matrix and Tensor Factorizations: Models ...A2 Linear mixing model. A3 Sources are balanced. A4 Additive mixture V = P i jS ij. A5 jS ijare well approximated by nonnegative

The end

I PhD Thesis:Nonnegative Matrix and Tensor Factorizations: Models, Algorithms and ApplicationsAvailable online: https://angms.science/doc/PhDThesis.pdf

Errata: https://angms.science/doc/PhDThesisErrata.pdf

I Works during the PhD period

green = journal paper, black = conference paper1. A., and Gillis, Volume regularized non-negative matrix factorizations, WHISPERS182. A., and Gillis, Algorithms and comparisons of nonnegative matrix factorizations with volume regularization for

hyperspectral unmixing, JSTAR, 193. A., and Gillis, Accelerating nonnegative matrix factorization algorithms using extrapolation, Neural

Computation, 194. A., Cohen and Gillis, Accelerating approximate nonnegative canonical polyadic decomposition using

extrapolation, GRETSI195. Leplat, Gillis, Siebert and A., Separation aveugle de sources sonores par factorization en matrices positives avec

penalite sur le volume du dictionnaire, GRETSI196. Leplat, A. and Gillis, Minimum-volume rank-deficient nonnegative matrix factorizations, ICASSP197. Leplat, Gillis and A., Blind Audio Source Separation with Minimum-Volume Beta-Divergence NMF, TSP, 208. A., Cohen, Le and Gillis, Extrapolated Alternating Algorithms for Approximate Canonical Polyadic

Decomposition, ICASSP209. A., Cohen, Gillis and Le, Accelerating Block Coordinate Descent for Nonnegative Tensor Factorization, NLAA,

under review, 2010. A., Gillis, Vandaele and De Sterck, Nonnegative Unimodal Matrix Factorization, submitted to ICASSP (NEW)

53 / 53

AngMSAndersen
slide: https://angms.science/doc/PhDPublicDefSlide.pdf