probabilistic distance measures for prototype-based rules włodzisław duch department of...

Download PROBABILISTIC DISTANCE MEASURES FOR PROTOTYPE-BASED RULES Włodzisław Duch Department of Informatics, Nicolaus Copernicus University, Poland, School of

Post on 17-Dec-2015




1 download

Embed Size (px)


  • Slide 1
  • PROBABILISTIC DISTANCE MEASURES FOR PROTOTYPE-BASED RULES Wodzisaw Duch Department of Informatics, Nicolaus Copernicus University, Poland, School of Computer Engineering, Nanyang Technological University Singapore. Marcin Blachnik, Tadeusz Wieczorek Department of Electrotechnology Faculty of Materials Engineering & Metallurgy, The Silesian University of Technology, Poland
  • Slide 2
  • ICONIP 2005 Taiwan Outline Type of rules What are prototype rules? Heterogeneous distance function Probability density function (PDF) estimation Results Conclusions
  • Slide 3
  • ICONIP 2005 Taiwan Types of rules Crisp logical rules. Rough sets and logic. Fuzzy rules (F-rules). Prototype rules (P-rules) most general? P-rules with additive similarity functions may be converted into the neurofuzzy rules with natural membership functions, including nominal features. P-rules do not need the feature space. There are many neurofuzzy programs, but no P-rules so far.
  • Slide 4
  • ICONIP 2005 Taiwan Motivation Understanding data, situations, recognizing objects or making diagnosis people frequently use similarity to known cases, and rarely use logical reasoning, but soft computing experts use logic instead of similarity... Relations between similarity and logic are not clear. Q 1 : How to obtain the same decision borders in Fuzzy Logic systems and Prototype Rule Based systems? Q 2 : What type of similarity measure corresponds to a typical fuzzy functions and vice versa? Q 3 : How to transform one type of a system into another type preserving their decision borders? Q 4 : Are there any advantages of such transformations? Q 5 : Can we understand data better using prototypes instead of logical rules?
  • Slide 5
  • ICONIP 2005 Taiwan Example
  • Slide 6
  • ICONIP 2005 Taiwan Prototype rules - advantages Inspired by cognitive psychology: understanding data, situations, recognizing objects or making diagnosis people frequently use similarity to known cases, and rarely use logical reasoning. With Heterogeneous Distance Functions P-rules supports all types of attributes: continues, discrete, symbolic and nominal, while F-rules require numerical inputs. Locally linear decision borders to avoid overfitting. Many algorithms for prototype selection and optimization exist but they have not been applied to understand data. Applications of P-rules to real datasets give excellent results generating small number of prototypes.
  • Slide 7
  • ICONIP 2005 Taiwan Prototype rules - learning Learning process involves: select similarity or dissimilarity (distance) functions model optimization: the number and positions of prototypes Decision making task consist of: calculating distance (similarity) to each prototype assigning P-rule to calculate the output class as a rule Nearest Neighbour rule: If P=argmin p (D(X,P)) Then Class(X)=Class(P) Threshold rule: If D(X,P)d p Then Class(X)=Class(P) Taking D (X,P) - Chebychev distance crisp logic rules are obtained
  • Slide 8
  • ICONIP 2005 Taiwan Applications to real data (ICONIP2004) Gene expression data for 2 types of leukaemia (Golub et al, Science 286 (1999) 531-537 Description: 2 classes, 1100 features, 3 most relevant selected. Used methods: 1 prototype/class LVQ, DVDM similarity measure. Results (number of misclassified vectors): Data SetGolub et alP-rules Train30 Test53 Searching for Promoters in DNA strings Description: 2 classes, 57 features, all symbolic features. Used methods: 9 prototypes for promoters, 12 for nonpromoters, generated using C-means + LVQ, with VDM similarity measure. Results: 5 misclassified vectors in leave one out test.
  • Slide 9
  • ICONIP 2005 Taiwan Distance (similarity) functions Continuous attributes Probabilistic Metrics N number of attributes K number of classes Input vectors X=[x 1, x 2, , x N ]T Y=[y 1, y 2, , y N ]T q exponent value P(C i |x) - posterior probab. for symbolic features, estimated as P(C i |x)=n i /n
  • Slide 10
  • ICONIP 2005 Taiwan Heterogeneous distance function Combine contributions from symbolic and real-valued features to get the distance. or use only probabilistic measures
  • Slide 11
  • ICONIP 2005 Taiwan Probability density function estimation Problem: how to combine influence of nominal/symbolic? 1. Normalization continuous symbolic 2. Estimation continuous attributes => prob. If estimation, then several options to get probabilities: Discretization (DVDM) Discretization + Interpolation (IVDM) Gaussian kernel estimation (GVDM) Rectangular Parzen window (LVDM) Rectangular moving Parzen window (PVDM)
  • Slide 12
  • ICONIP 2005 Taiwan Discretization & Interpolation Moving Parzen windows. Gaussian kernel Rect. Parzen window 3 overlapping Gaussians in 4D, good parameters for estimation.
  • Slide 13
  • ICONIP 2005 Taiwan Discretization & Interpolation Gaussian kernel Rect. Parzen windowMoving Parzen wind. 3 overlapping Gaussians in 4D, bad parameters for estimation.
  • Slide 14
  • ICONIP 2005 Taiwan Testing and comparison procedure 6 real datasets with mixes symbolic/real features. Flags (UCI repository) Glass (UCI repository) Promoters (UCI repository) Wisconsin Brest Cancer, WBC (UCI repository) Pima Indians diabetes (UCI repository) Lancet ( from A.J. Walker, S.S. Cross, R.F. Harrison, Visualization of biomedical datasets by use of growing cell structure networks: a novel diagnostic classification technique. Lancet Vol. 354, pp. 1518-1522, 1999. ) For all tasks 10 fold CV test procedure is used. Two artificial datasets for testing, 2D 200 vectors/class uniform distribution 200 vectors/class normal distribution
  • Slide 15
  • ICONIP 2005 Taiwan Classification results Results on artificial datasets. Left: Gaussian distributed. Right: uniform distributed. Similar results, except for convergence problems. Datasets with all symbolic or discrete values. leave-one-out results.
  • Slide 16
  • ICONIP 2005 Taiwan Real datasets
  • Slide 17
  • ICONIP 2005 Taiwan Results & discussion Selection of appropriate parameters is very important. Incorrect values if one uses: too small sigma (Gaussian Estimation); too narrow window (Rectangular Parzen Window estimations) too many bins in discretization. Increased sensitivity of estimation methods => overfitting if too high sigma (Gaussian Estimation); too wide window (Rectangular Parzen Window estimations) Low number of bins in discretization. Decreased sensitivity of estimation methods leading to over- generalization. Middle values of parameters are best start points leading to good results ( 0.5, Parzen width 0.5, Parzen step 0.01)
  • Slide 18
  • ICONIP 2005 Taiwan Some conclusions First step in understanding relations between fuzzy and similarity- based systems. Prototype rules can be expressed using fuzzy rules and vice versa leading to new possibilities in both fields: new type of membership functions & new type of distance functions. Expert knowledge can be captured in any kind of rules, but sometimes it may be more natural to express knowledge as P-rules (similarity) or as F-rules (logical conditions). VDM measure used in P-rules leads to a natural shape of membership functions in fuzzy logic for symbolic data. There is no best choice of heterogeneous distance function type or PDF estimation method or probability metrics. Simplest methods may lead to good results. Selection of appropriate parameters is very important. P-systems should be as popular as neurofuzzy systems, although many open problems still remain, both theoretical and practical.
  • Slide 19
  • Thank you for lending your ears...


View more >