yan cui 2013.1.16
DESCRIPTION
A novel supervised feature extraction and classification framework for land cover recognition of the off-land scenario. Yan Cui 2013.1.16. The related work 2. The integration algorithm framework 3. Experiments. The related work Locally linear embedding - PowerPoint PPT PresentationTRANSCRIPT
![Page 1: Yan Cui 2013.1.16](https://reader036.vdocuments.net/reader036/viewer/2022062304/56812bb1550346895d8feb81/html5/thumbnails/1.jpg)
A novel supervised feature extraction and classification framework for land cover
recognition of the off-land scenario
Yan Cui
2013.1.16
![Page 2: Yan Cui 2013.1.16](https://reader036.vdocuments.net/reader036/viewer/2022062304/56812bb1550346895d8feb81/html5/thumbnails/2.jpg)
1. The related work
2. The integration algorithm
framework
3. Experiments
![Page 3: Yan Cui 2013.1.16](https://reader036.vdocuments.net/reader036/viewer/2022062304/56812bb1550346895d8feb81/html5/thumbnails/3.jpg)
The related work
Locally linear embedding
Sparse representation-based classifier
K-SVD dictionary learning
![Page 4: Yan Cui 2013.1.16](https://reader036.vdocuments.net/reader036/viewer/2022062304/56812bb1550346895d8feb81/html5/thumbnails/4.jpg)
Locally linear embedding
LLE is an unsupervised learning algorithm
that computers low-dimensional,
neighbor-hood-preserving embedding of
high-dimensional inputs.
![Page 5: Yan Cui 2013.1.16](https://reader036.vdocuments.net/reader036/viewer/2022062304/56812bb1550346895d8feb81/html5/thumbnails/5.jpg)
Specifically, we expect data point and
its neighbors to lie on or close to a
locally linear patch of the manifold and
the local reconstruction errors of these
patches are measured by
2
1 2( ) (1)
k
i ij ji je w x w x
![Page 6: Yan Cui 2013.1.16](https://reader036.vdocuments.net/reader036/viewer/2022062304/56812bb1550346895d8feb81/html5/thumbnails/6.jpg)
2
1 2( ) (2)
k
i ij ji je w y w y
![Page 7: Yan Cui 2013.1.16](https://reader036.vdocuments.net/reader036/viewer/2022062304/56812bb1550346895d8feb81/html5/thumbnails/7.jpg)
Sparse representation-based classifier
The sparse representation-based classifier
can be considered a generalization of
nearest neighbor (NN) and nearest
subspace (NS), it adaptively chooses the
minimal number of training samples
needed to represent each test sample.
![Page 8: Yan Cui 2013.1.16](https://reader036.vdocuments.net/reader036/viewer/2022062304/56812bb1550346895d8feb81/html5/thumbnails/8.jpg)
1
1 2
11 12 1 1 2 1 2
[ , , , ]
[ , , , , , , , , , , , , , ] i c
c
m nn i i in c c cn
A A A A
x x x x x x x x x R
(3)my A R
![Page 9: Yan Cui 2013.1.16](https://reader036.vdocuments.net/reader036/viewer/2022062304/56812bb1550346895d8feb81/html5/thumbnails/9.jpg)
( ) arg 00 min ,
s.t.
L
A y
(4)
![Page 10: Yan Cui 2013.1.16](https://reader036.vdocuments.net/reader036/viewer/2022062304/56812bb1550346895d8feb81/html5/thumbnails/10.jpg)
( ) arg 11 min ,
s.t.
L
A y
(5)
![Page 11: Yan Cui 2013.1.16](https://reader036.vdocuments.net/reader036/viewer/2022062304/56812bb1550346895d8feb81/html5/thumbnails/11.jpg)
)(ˆ ii Ay
![Page 12: Yan Cui 2013.1.16](https://reader036.vdocuments.net/reader036/viewer/2022062304/56812bb1550346895d8feb81/html5/thumbnails/12.jpg)
2 2
2 2ˆmin ( ) ( ) = (6) i i ii
r y y y y A
![Page 13: Yan Cui 2013.1.16](https://reader036.vdocuments.net/reader036/viewer/2022062304/56812bb1550346895d8feb81/html5/thumbnails/13.jpg)
K-SVD dictionaries learningThe original training samples have much
redundancy as well as noise and trivial information that can be negative to the recognition.
If the training samples are huge, the computation of SR will be time consuming, so an optimal dictionary is needed for the sparse representation and classification.
![Page 14: Yan Cui 2013.1.16](https://reader036.vdocuments.net/reader036/viewer/2022062304/56812bb1550346895d8feb81/html5/thumbnails/14.jpg)
![Page 15: Yan Cui 2013.1.16](https://reader036.vdocuments.net/reader036/viewer/2022062304/56812bb1550346895d8feb81/html5/thumbnails/15.jpg)
The K-SVD algorithm
2
02 0min . . ( 1, 2, , )0
ii i ix D s t T i n
![Page 16: Yan Cui 2013.1.16](https://reader036.vdocuments.net/reader036/viewer/2022062304/56812bb1550346895d8feb81/html5/thumbnails/16.jpg)
The dictionary update stage:
![Page 17: Yan Cui 2013.1.16](https://reader036.vdocuments.net/reader036/viewer/2022062304/56812bb1550346895d8feb81/html5/thumbnails/17.jpg)
![Page 18: Yan Cui 2013.1.16](https://reader036.vdocuments.net/reader036/viewer/2022062304/56812bb1550346895d8feb81/html5/thumbnails/18.jpg)
![Page 19: Yan Cui 2013.1.16](https://reader036.vdocuments.net/reader036/viewer/2022062304/56812bb1550346895d8feb81/html5/thumbnails/19.jpg)
![Page 20: Yan Cui 2013.1.16](https://reader036.vdocuments.net/reader036/viewer/2022062304/56812bb1550346895d8feb81/html5/thumbnails/20.jpg)
Let be the training data matrix,
is the -th class training samples matrix, a test data can be well approximated by the linear combination of the training data, i.e.
The integration algorithm for supervised learning
1 2[ , , , ] m ncB B B B R
1 2[ , , , ] ( 1,2, , ) ii
m ni i i inB x x x R i c i
1
n
i iiy x
mRy
![Page 21: Yan Cui 2013.1.16](https://reader036.vdocuments.net/reader036/viewer/2022062304/56812bb1550346895d8feb81/html5/thumbnails/21.jpg)
Let be the representation coefficient vector with respect to -class. To make SRC achieve good performance on all training samples, we expect the within class residual minimized, while the between class residual maximized, simultaneously. Therefore we redefine the following optimization problem:
22
2 12min ( ) ( ) i jj i
y B y B
( )i i
(15)
![Page 22: Yan Cui 2013.1.16](https://reader036.vdocuments.net/reader036/viewer/2022062304/56812bb1550346895d8feb81/html5/thumbnails/22.jpg)
22
2 12min ( ) ( )i jj i
y D y D
( ) , ) (k k i j
k
(16)
![Page 23: Yan Cui 2013.1.16](https://reader036.vdocuments.net/reader036/viewer/2022062304/56812bb1550346895d8feb81/html5/thumbnails/23.jpg)
Let is the representation coefficient vector with respect to -th class, so the optimization problem in Eq. (16) is turned to
( )i
i
22
2 12min ( ) ( )i iy D y D
(17)
![Page 24: Yan Cui 2013.1.16](https://reader036.vdocuments.net/reader036/viewer/2022062304/56812bb1550346895d8feb81/html5/thumbnails/24.jpg)
In order to obtain the sparse representation coefficients, we want to learn an embedding map to reduce the dimensionality of and preserve the spare reconstruction. So the optimization problem in Eq. (17) is turned to
1 2[ , , , ] m ddW w w w R
2 2
12 2,min ( ) ( )T T T T
i iWW y W D W y W D
![Page 25: Yan Cui 2013.1.16](https://reader036.vdocuments.net/reader036/viewer/2022062304/56812bb1550346895d8feb81/html5/thumbnails/25.jpg)
For a given test set , we can adaptivelylearn the embedding map, the optimal dictionary and the sparse reconstruction coefficients by the following optimization problem
1 2{ , , , }lU y y y
2 2
1,
ˆmin T T T T
FW FW U W D W U W D
![Page 26: Yan Cui 2013.1.16](https://reader036.vdocuments.net/reader036/viewer/2022062304/56812bb1550346895d8feb81/html5/thumbnails/26.jpg)
The feature extraction and classification algorithm
![Page 27: Yan Cui 2013.1.16](https://reader036.vdocuments.net/reader036/viewer/2022062304/56812bb1550346895d8feb81/html5/thumbnails/27.jpg)
Experiments for unsupervised learning
The effect of dictionary selection
Compare with pure feature extraction
![Page 28: Yan Cui 2013.1.16](https://reader036.vdocuments.net/reader036/viewer/2022062304/56812bb1550346895d8feb81/html5/thumbnails/28.jpg)
Databases descriptions
UCI databases: the Gas Sensor Array Drift Data set and the Synthetic Control Chart Time Series Date Set.
![Page 29: Yan Cui 2013.1.16](https://reader036.vdocuments.net/reader036/viewer/2022062304/56812bb1550346895d8feb81/html5/thumbnails/29.jpg)
The effect of dictionary selection
![Page 30: Yan Cui 2013.1.16](https://reader036.vdocuments.net/reader036/viewer/2022062304/56812bb1550346895d8feb81/html5/thumbnails/30.jpg)
![Page 31: Yan Cui 2013.1.16](https://reader036.vdocuments.net/reader036/viewer/2022062304/56812bb1550346895d8feb81/html5/thumbnails/31.jpg)
Compare with pure feature extraction
![Page 32: Yan Cui 2013.1.16](https://reader036.vdocuments.net/reader036/viewer/2022062304/56812bb1550346895d8feb81/html5/thumbnails/32.jpg)
Experiments
The effect of dictionary selection
Compare with pure classification
Compare with pure feature extraction
![Page 33: Yan Cui 2013.1.16](https://reader036.vdocuments.net/reader036/viewer/2022062304/56812bb1550346895d8feb81/html5/thumbnails/33.jpg)
Databases descriptions
![Page 34: Yan Cui 2013.1.16](https://reader036.vdocuments.net/reader036/viewer/2022062304/56812bb1550346895d8feb81/html5/thumbnails/34.jpg)
![Page 35: Yan Cui 2013.1.16](https://reader036.vdocuments.net/reader036/viewer/2022062304/56812bb1550346895d8feb81/html5/thumbnails/35.jpg)
The effect of dictionary selection
![Page 36: Yan Cui 2013.1.16](https://reader036.vdocuments.net/reader036/viewer/2022062304/56812bb1550346895d8feb81/html5/thumbnails/36.jpg)
![Page 37: Yan Cui 2013.1.16](https://reader036.vdocuments.net/reader036/viewer/2022062304/56812bb1550346895d8feb81/html5/thumbnails/37.jpg)
Compare with pure classification
![Page 38: Yan Cui 2013.1.16](https://reader036.vdocuments.net/reader036/viewer/2022062304/56812bb1550346895d8feb81/html5/thumbnails/38.jpg)
![Page 39: Yan Cui 2013.1.16](https://reader036.vdocuments.net/reader036/viewer/2022062304/56812bb1550346895d8feb81/html5/thumbnails/39.jpg)
![Page 40: Yan Cui 2013.1.16](https://reader036.vdocuments.net/reader036/viewer/2022062304/56812bb1550346895d8feb81/html5/thumbnails/40.jpg)
Compare with pure feature extraction
![Page 41: Yan Cui 2013.1.16](https://reader036.vdocuments.net/reader036/viewer/2022062304/56812bb1550346895d8feb81/html5/thumbnails/41.jpg)
Thanks!
![Page 42: Yan Cui 2013.1.16](https://reader036.vdocuments.net/reader036/viewer/2022062304/56812bb1550346895d8feb81/html5/thumbnails/42.jpg)
Question & suggestion?