acknowledgements: barry kalmanstan kwasny koong-nah chungjohn heuser

29
Acknowledgements: Barry Kalman Stan Kwasny Koong-Nah Chung John Heuser Detection of Intracellular Organelles in Digitized Electronmicroscopic Images Using Wavelets and Neural Networks Final Report for CS513 John Olsen and Wei Yan

Upload: tender

Post on 18-Jan-2016

21 views

Category:

Documents


0 download

DESCRIPTION

Detection of Intracellular Organelles in Digitized Electronmicroscopic Images Using Wavelets and Neural Networks Final Report for CS513 John Olsen and Wei Yan. Acknowledgements: Barry KalmanStan Kwasny Koong-Nah ChungJohn Heuser. The Experiment. - PowerPoint PPT Presentation

TRANSCRIPT

Page 1: Acknowledgements: Barry KalmanStan Kwasny Koong-Nah ChungJohn Heuser

Acknowledgements:

Barry Kalman Stan Kwasny

Koong-Nah Chung John Heuser

Detection of Intracellular Organelles in Digitized Electronmicroscopic Images Using Wavelets and Neural Networks

Final Report for CS513John Olsen and Wei Yan

Page 2: Acknowledgements: Barry KalmanStan Kwasny Koong-Nah ChungJohn Heuser

The Experiment

• We used Kalman & Kwasny’s neural network tool to create a set of trained feedforward neural networks (FFNNs) that detect particular intracellular organelles, called caveolae, in scanning EM images.

Page 3: Acknowledgements: Barry KalmanStan Kwasny Koong-Nah ChungJohn Heuser
Page 4: Acknowledgements: Barry KalmanStan Kwasny Koong-Nah ChungJohn Heuser
Page 5: Acknowledgements: Barry KalmanStan Kwasny Koong-Nah ChungJohn Heuser
Page 6: Acknowledgements: Barry KalmanStan Kwasny Koong-Nah ChungJohn Heuser

Overview of Experiment

Training Set

Test Set

32 N 32 P

99 N100 P

WaveletTransform

WaveletTransform

Unsupervised,Recurrent Net:LOSRAAM

Supervised,Feed-ForwardNeural Net

Unsupervised,Recurrent Net:LOSRAAM

5 TrainedFFNNs

Final DecisionBy Vote

Page 7: Acknowledgements: Barry KalmanStan Kwasny Koong-Nah ChungJohn Heuser

Performance is 84% correct

0

5

10

15

20

25

30

35

Test1 Test2 Test3 Test4 Total Perfect

Negatives

Positives

Page 8: Acknowledgements: Barry KalmanStan Kwasny Koong-Nah ChungJohn Heuser

Images

Components: Vectors of waveletcoeffs(resolution, dimension)

Vectors ofLOSRAAM StateTransitions

K Centers ofAttraction

K x K FuzzyTransitionMatrix

Hold-outSet

FuzzyFeatureVectors

Reduced,CombinedFFV

PACSet

TestSet

TrainedFFNNs

FinalDecision

Data Flow Diagram

Page 9: Acknowledgements: Barry KalmanStan Kwasny Koong-Nah ChungJohn Heuser

Wavelet Transform

• Equivalent to the 1st order derivative of a smoothing function.

• Analogous to the Fourier Transform.

• Multiple levels of Resolution.

• 2-Dimensions

• Image component: The set of wavelet coefficients for a particular dimension at a particular resolution.

• Image components: Multiple representations of the entire image at different resolutions.

Page 10: Acknowledgements: Barry KalmanStan Kwasny Koong-Nah ChungJohn Heuser

Image L1 L2 L3 L4

320 x 240 160 x 120 80 x 60 40 x 30 20 x 15

1 4 16 64 256

76,800 19,200 4,800 1200 300

- 2 2 2 3

- 38,400 9,600 2,400 900

Level

Image size(grains)

Grain size(pixels)

Total Grains

No. of Comps (x, y, r)

No. of Coeffs

Four-level, 2-D Wavelet Hierarchy

• Maximum no. of coeffs at any one level, direction = no of grains in image. For our data, the max no. of coefficients/image = 51, 300.

• A Component consists of the set of wavelet coefficients for a particular direction and level of resolution.

• Total of 9 components: 2 for each level of decomposition, 1 for undecomposed components at the last level (DC, very low frequency).

Page 11: Acknowledgements: Barry KalmanStan Kwasny Koong-Nah ChungJohn Heuser

Why Wavelets ?

• Enhance contrast edges.

• Multi-Resolution emphasizes features of different sizes.

• Data reduction from thresholding.

Page 12: Acknowledgements: Barry KalmanStan Kwasny Koong-Nah ChungJohn Heuser

Wavelet Coefficient Thresholding

• For all images, dropped coefficients with magnitudes < 0.4

• 16 million coefficients reduced to 1.6 million.

• Threshold determined by experience and experiment: Set of coefficients after thresholding reverse transformed to image, and this image is visually compared to the original.

Page 13: Acknowledgements: Barry KalmanStan Kwasny Koong-Nah ChungJohn Heuser
Page 14: Acknowledgements: Barry KalmanStan Kwasny Koong-Nah ChungJohn Heuser

LOSRAAM Training Set

• Each component is made a separate vector: 9 components per image * 199 images = 1791 components.

• 1.6 M coefficients distributed over 1791 components.

• Each coefficient represented by a doublet: [magnitude, normalized index into component array].

• Coefficients fed into LOSRAAM Neural Network one by one.

Page 15: Acknowledgements: Barry KalmanStan Kwasny Koong-Nah ChungJohn Heuser

The LOSRAAM ANN

• Linear Output Recurrent Recursive Auto-Associative Memory

• AAM: Targeting the outputs to be the same as the inputs.

• Recurrence: a portion of the input is the activation pattern from the hidden layer of the previous iteration.

• Unsupervised Learning: A criterion for judging outputs is determined. The ANN learns a mapping of inputs to outputs that fits the criterion.

Page 16: Acknowledgements: Barry KalmanStan Kwasny Koong-Nah ChungJohn Heuser

Output Units

Linear State (“Context”)

Input Units

HiddenUnits

Coefficient Index

6 input units 4 hidden units, determined by experiment 6 output units 10 bias units 74 adjustable weights, initially all 0. Recurrent connections from hidden units to

input state units

6-4-6 LOSRAAM Neural Network

Page 17: Acknowledgements: Barry KalmanStan Kwasny Koong-Nah ChungJohn Heuser

LOSRAAM Training Statistics

• Running time 55 hours.• Sun Sparc, 4 parallel processors.• Error function: initial value = 1,800,000

final value = 401.

Page 18: Acknowledgements: Barry KalmanStan Kwasny Koong-Nah ChungJohn Heuser

Clustering LOSRAAM State Vectors

• Image structure produces a sequence of activation patterns of the 4 unit hidden layer.

• Activation patterns represented by a sequence of vectors in 4-D space.

• The trajectory of these vectors is represented by a series of point values in 4-D space.

• These points are clustered into “centers of attraction”.

• Clustering is a “fuzzy” process.

• Our LOSRAAM data yielded 4 centers of attraction.

Page 19: Acknowledgements: Barry KalmanStan Kwasny Koong-Nah ChungJohn Heuser

Fuzzy Transition Matricesand Fuzzy Feature Vectors

• An FTM is a 4 x 4 matrix of transitions between centers of attraction.

• FTMs were computed for each component, for each image.

• Thus, each image is represented by 9 FTMs, each containing 16 elements, 144 elements total.

• Linearization of FTMs gives 144 element FFVs, one per image.

Page 20: Acknowledgements: Barry KalmanStan Kwasny Koong-Nah ChungJohn Heuser

Singular Valued Decomposition of FFVs

• SVD is a data reduction and conditioning technique.• Combine FFVs for all images into a 144 x 199 matrix.• Using SVD, identified 9 of the 144 columns that accounted

for > 99% of the variance in the entire training data set.• These 1791 values used to train the Feed Forward ANNs.

144 elements

199 FFVs 199 reduced FFVs

9 elements

Page 21: Acknowledgements: Barry KalmanStan Kwasny Koong-Nah ChungJohn Heuser

Feed-Forward Neural Network

• Supervised Learning: Input and output are provided. The FFNN learns mapping by example.

• We used a 9-1-1 FFNN.

Page 22: Acknowledgements: Barry KalmanStan Kwasny Koong-Nah ChungJohn Heuser

Image’s 9 element

FFV

Input Units

9-1-1 Feed-Forward Neural Network

Output Unit

Hidden UnitFully skip-connected

Page 23: Acknowledgements: Barry KalmanStan Kwasny Koong-Nah ChungJohn Heuser

FFNN Training

• Input 199 training patterns (9 values each), one at a time.• Periodically check performance with PAC set. (A set of 16

images, 8 positive, 8 negative).• Harvest weights if performance on training set and on PAC are

both > 85%.• If PAC test failed, weights are discarded.• Trained 5 FFNNs to criterion performance on training set and

PAC.

Page 24: Acknowledgements: Barry KalmanStan Kwasny Koong-Nah ChungJohn Heuser

FFNN Training Issues

• Overtraining. First time training with a two-hidden-unit FFNN led to 100% performance for both training and PAC, but only 65% performance on the test set.

• Attempt to raise harvest criterion to 88% for one-hidden-unit FFNN failed. No trained networks were produced.

Page 25: Acknowledgements: Barry KalmanStan Kwasny Koong-Nah ChungJohn Heuser

Images

Components: Vectors of waveletcoeffs(resolution, dimension)

Vectors ofLOSRAAM StateTransitions

K Centers ofAttraction

K x K FuzzyTransitionMatrix

Hold-outSet

FuzzyFeatureVectors

Reduced,CombinedFFV

PACSet

TestSet

TrainedFFNNs

FinalDecision

Data Flow Diagram

Page 26: Acknowledgements: Barry KalmanStan Kwasny Koong-Nah ChungJohn Heuser

Can performance be improved?

• Human performance ~ 100%.

• ANN must handle several different sources of variation: position, view angle, size, number, flatness.

• Increase the size and depth of the training set, e.g., 500 images.

• Add positional ‘hint’ units, number hint units, etc.

Page 27: Acknowledgements: Barry KalmanStan Kwasny Koong-Nah ChungJohn Heuser

Significance of Results to Kalman and Kwasny’s ANN

• Scanning EM images are rich in detail, perhaps more so than mammograms.

• Absolute performance on EM images is better than performance on mammograms, particularly with respect to specificity.

• The results suggest that the ANN is capable of using the rich detail found in EM images to achieve a low false positive rate.

Mammograms EMs

False Pos. 44% 14%

False Neg. 25% 14%

Page 28: Acknowledgements: Barry KalmanStan Kwasny Koong-Nah ChungJohn Heuser
Page 29: Acknowledgements: Barry KalmanStan Kwasny Koong-Nah ChungJohn Heuser

Potential Applications

• Automated scanning of EM images.• Large scale screening of histological slides for

abnormal cell morphologies.