Professional Documents
Culture Documents
discussions, stats, and author profiles for this publication at: https://www.researchgate.net/publication/280038962
CITATIONS
READS
34
3 authors:
Soren Atmakuri Davidsen
Sreedevi Erry
8 PUBLICATIONS 3 CITATIONS
5 PUBLICATIONS 2 CITATIONS
SEE PROFILE
SEE PROFILE
Padmavathamma Mokkala
Sri Venkateswara University
68 PUBLICATIONS 36 CITATIONS
SEE PROFILE
All content following this page was uploaded by Soren Atmakuri Davidsen on 14 July 2015.
The user has requested enhancement of the downloaded file. All in-text references underlined in blue
are linked to publications on ResearchGate, letting you access and read them immediately.
Abstract. Fuzzy pattern classifiers are a recent type of classifiers making use of fuzzy membership functions and fuzzy aggregation rules, providing a simple yet robust classifier model. The fuzzy pattern classifier is
parametric giving the choice of fuzzy membership function and fuzzy aggregation operator. Several methods for estimation of appropriate fuzzy
membership functions and fuzzy aggregation operators have been suggested, but considering only fuzzy membership functions with symmetric
shapes found by heuristically selecting a middle point from the learning examples. Here, an approach for learning the fuzzy membership functions and the fuzzy aggregation operator from data is proposed, using
a genetic algorithm for search. The method is experimentally evaluated
on a sample of several public datasets, and performance is found to be
significantly better than existing fuzzy pattern classifier methods. This is
despite the simplicity of the fuzzy pattern classifier model, which makes
it interesting.
Keywords: approximate reasoning, pattern recognition, fuzzy classifier,
genetic algorithm
Introduction
Ideas from uncertain reasoning using fuzzy logic have found their way into classification, where new methods are continuously being developed, as their ability
to deal with the uncertainty, or fuzziness of training data is intuitive and useful[3]. In particular the fuzzy methods are useful for robustness, meaning how
stable they are in a testing situation as compared to training; and interpretability, meaning how easy it is to understand the learned model; while maintaining
good generalization, meaning how well is dealt with unseen examples. Many conventional classifiers are designed from the idea that a training example should
only belong to one class, and hence perform best in cases where classes are welldefined and separable by clear boundaries. This may not be the case in many
real world problems, where classes are hard to define, and feature-value regions
may often be overlapping. Fuzzy classifiers, by their fuzzy nature, assume that
training examples may have gradual membership to a class, and feature-value regions may be allowed to take on nearby values and still be considered relevant
for the class at hand.
Fuzzy pattern classifiers have recently been introduced as a new method for
approaching classification problems[7, 8]. A fuzzy pattern describes a prototype,
the look, of a class. The fuzzy pattern consists of a fuzzy membership function
for each feature, and a fuzzy aggregation operator used for inference. A fuzzy
pattern classifier is an ensemble of such fuzzy patterns. The ensemble classifies
new instances by selecting the fuzzy pattern with highest aggregated membership. The fuzzy pattern classifier is fuzzy not only because of using concepts
from fuzzy logic, but, due to the way fuzzy membership functions capture the
uncertain knowledge from the learning examples and describe the look of the
class in a way that unseen examples may still be recognized though values are
outside the boundaries of what was already seen[6].
Existing work on fuzzy pattern classifiers have proposed modifications along:
(1) which fuzzy membership function is most appropriate to describe a feature;
(2) which fuzzy aggregation operator is most appropriate for inferring the class.
However, the focus has been on providing a heuristic for constructing the fuzzy
membership functions[7, 8], or simply trying out new fuzzy aggregation operators[8].
The contributions of this paper are two new fuzzy pattern classifier methods
using genetic algorithms, which address the issue of constructing fuzzy membership functions and selecting fuzzy aggregation operators. The first proposed
method uses a global search where all parameters of the fuzzy pattern classifier are searched simultaneously and accuracy is used to measure fitness, based
on the idea that global search will provide appropriate separation of the classes.
The second proposed method searches fuzzy membership functions for each fuzzy
pattern, and an within-class similarity measure as fitness (local search). The local search method is based on the idea that learning shapes of fuzzy membership
functions (i.e. covering the training examples properly) will give a more robust
classifier.
Fuzzy logic has been applied to many types of fuzzy classifiers, and the idea
of using genetic algorithms to search for parameters is not new either. In the
area of genetic fuzzy systems, the use of genetic algorithms in fuzzy rule based
classifiers has been extensively researched. In fuzzy rule based classifiers, the
model is known as the knowledge base and consists of a data base and a rule
base. The data base defines how fuzzification is to be done, and the rules define
the inference. In genetic fuzzy systems, genetic algorithms have been used in two
contexts: (1) tuning the knowledge base, meaning, adjusting the membership
functions used for fuzzification and adjusting weights for each of the inference
rules; (2) learning the knowledge base, meaning, searching for membership functions useful for fuzzification and searching for inference rules. For a recent survey
and overview, see [2] and [9].
In fuzzy pattern classifiers however, to best knowledge of the authors, this
is the first example of using genetic algorithms to search for fuzzy membership
functions and fuzzy aggregation operators. The rule base in a fuzzy rule based
classifier can contain several rules for each class, depending on the training examples and how they cover the domain. Likewise, each rule in a fuzzy rule based
classifier can consider only a subset of the features. A fuzzy pattern classifier on
the other hand is analogous to a fuzzy rule based classifier, which learns only a
single rule (pattern) for each class, and the single rule considers all the features of
the training examples. Thus, selection of membership functions to represent the
rule and aggregation operator is important. In this paper the focus is selection
of membership functions, through use of genetic algorithms.
The rest of the paper is organized as follows: Section 2 introduces the classification setting, fuzzy pattern classifiers and genetic algorithms. In Section 3
the fuzzy pattern classifiers with searched membership functions are proposed,
while in Section 4 experimental analysis of the proposed classifiers are shown
for a sample of public datasets. Finally, concluding remarks and further work is
given in Section 5.
2
2.1
Preliminaries
Classification setting
The supervised classification problem can be defined as: Given training examples D with n examples, where each example contains m features xi =
(ai,1 , ai,2 , ..., ai,m ); and each example belongs to a class yi Y:
D = {(x1 , y1 ), (x2 , y2 ), ..., (xn , yn )} (X Y)n
(1)
select a hypothesis h from the classifiers hypothesis space H which minimizes the loss function l on the training examples:
Z
h arg min hH
l(h(x), y)
(2)
(x,y)D
The loss function must be selected, and typically a measure such as accuracy
is used (note, that the negation of accuracy should be used to make it measure
loss). Accuracy can be calculated as follows for a binary classification problem:
accuracy = (T P + T N )/(T P + F P + T N + F N )
(3)
where TP and TN are true positive and true negative (correctly classified)
examples, and FP and FN are false positives and false negatives (incorrectly
classified) examples.
2.2
Fuzzy pattern classifiers (FPCs) are popular because they are simple to construct, their model is easy to understand, they deals well with uncertainty of
unobserved examples, and the run-time evaluation of the model is very little.
The fuzzy pattern classifier is an ensemble of fuzzy patterns F , one for each
class, and a fuzzy aggregation operator used for inference. Thus, the model
is denoted (, F ). Defuzzification is done by selecting the class belonging to the
pattern with the highest inferred value. In literature, the fuzzy aggregation operator must be the same for all fuzzy patterns, and this restriction is upheld
here.
A fuzzy pattern is a model of a class, and consists of a set of fuzzy membership
functions (-functions), one for each feature of the example:
Fi = [i,1 , i,2 , ..., i,m ]
(4)
m
m1
2
((r
x)/(r
a))
p<xr
m
m1
(5)
(x) = 1 2
((x r)/(b r))
r<xq
m
m1
2
((b x)/(b r))
q<xb
0
Otherwise
where m > 0, a < r < b, p = 21 (r + a), and q = 12 (r + b).
For the j th feature of examples belonging to class k, determine -function
parameters r, p, q as follows:
rk,j = ai,j ,
yi = k
yi = k
(6)
AIWA
Symbol Definition
Qn
PROD
i ai
min
min(a1 , a2 , ..., an )
P
hMEAN n1 n
ai
Pn i
v
a(i) where a() is an ordered permutation
hOWA
i
i
of aand v a sequence of weights.
Pn
(1/)
hAIWA 1
, for p > 0.5, = (1/p) 1
i (1 ai )
Maximum
Algebraic sum
max
max(a1 , a2 , ..., an )
Q
alg.sum 1 n
i (1 ai )
Product
Minimum
Arithmetic mean
OWA
kY
(7)
after which defuzzication is simply selecting the pattern with the highest
inferred value:
y = arg max(Q1 , Q2 , ..., Qk )
(8)
kY
2.3
mutation probability). For boolean values, mutation simply flip the value, while
real-valued mutation adds a value drawn using a uniform random distribution
within an interval, for example [0.5, 0.5]. The crossover operator produce a new
chromosome provided two parent chromosomes. For selecting the mix of genes
from each parent, one or more crossover-points are used: At each cross-over point,
the childs genes will switch origin from one parent to the other. The cross-over
points can be selected randomly, or, from a specified set of crossover-points (this
is also called uniform crossover).
In order to use a GA, the following steps should be considered: (1) Decide
coding of chromosome into candidate solution; (2) Decide fitness function to
evaluate fitness of candidate solution; (3) Define which specific genetic operators
to use; (4) Define a stopping criteria.
g0
Pg Initialize population
fg Fitness Pg
repeat
g g+1
Pg Select fittest from Pg1
Crossover Pg
Mutate Pg
fg Fitness Pg
until g = G or fg fg1
Coding means defining a function for creating the candidate solution S from
the chromosome, f : S. The fitness function J, as already mentioned, evaluates how good a candidate solution is for the given problem typically measuring
by measuring some kind of loss, J : S [0, 1] where 0 means no loss (best
solution) and 1 complete loss (worst). Stopping criteria is for example a fixed
number of iterations, or reaching a specific minimum improvement threshold .
For a complete introduction to genetic algoritms please see [10]. For simplicity,
we denote use of an algorithm G : fitness , meaning, the genetic algorithm
is a function G, which provided a fitness function, yields a fittest chromosome
.
,(x, a, r, b)
,(x, a, r, b)
,(x, a, r, b)
crossover points
aggregation
feature-i
3.1
FPC-GA classifier
In the FPC-GA classifier, the GA considers all fuzzy patterns in the fuzzy pattern classifier simultaneously. This means the fitness of the selected chromosome/solution considers the separation of the classes, and not only covering the
feature-values of the training examples most effectively.
For FPC-GA, a chromosome is of length 5mC + 2, where 5 is the number
of genes used to represent a membership function and its parameters, m is the
number of features and C the number of classes. The last 2 are used for coding
of the aggregation operator and its parameters.
= 01
F1,1 = M02 (3 , 4 , 5 )
F1,2 = M06 (7 , 8 , 9 )
=
(9)
...
...
(10)
3.2
FPC-LGA classifier
For the second proposed classifier, FPC-LGA, the GA searches for fuzzy membership functions one fuzzy pattern at a time, i.e. the fitness is how well the fuzzy
membership functions cover a specific class, not the separation of the classes.
The method of FPC-LGA coding is similar to the FPC-GA coding, except a
chromosome represents a single fuzzy pattern, and the fuzzy aggregation operator is fixed:
F1 = M01 (i1 , i2 , i3 )
F = M 0 ( , , )
2
i4
i5
i6
2
=
(11)
...
predictions by the fuzzy pattern k as previously defined in Eq. 7 and ideal target
for fuzzy pattern k is:
vector Q
(
1 if yi = k
, for i = 1, .., n
(12)
Qk =
0 otherwise
This gives the fitness function:
JL = RMSE(Qk , Qk )
(13)
Experimental analysis
Datasets
Twelve available datasets were selected, available through the UCI dataset repository2 , except for Telugu Vowels, which is from the Indian Statistical Institute
website3 . All datasets had examples with missing values removed and values
feature-wise normalized to the unit interval [0, 1]. The 12 datasets were selected
based on the absence of nominal data, hence, only numeric value features (real
and/or integer), which are supported by the FPC. The datasets used are shown
in Table 2.
1
2
3
10
4.2
Iris
Pima Indians Diabetes
Wisconsin Breast Cancer
Bupa Liver Disorders
Wine
Telugu Vowels
Haberman
Indian Liver Patient
Vertebral Column
Glass
Ionosphere
Balance Scale
4
8
9
6
12
3
3
10
6
10
34
4
3
2
2
2
3
6
2
2
2
6
2
3
0.33
0.65
0.65
0.58
0.40
0.24
0.74
0.72
0.68
0.36
0.64
0.46
Method
For comparison, the FyLearn implementation of the two FPC classifiers FRR[7]
and MFPC[8] were included. The classifiers were configured with the parameters suggested by their authors. For FRR -membership functions and two
different fuzzy aggregation rules were selected; product and arithmetic mean,
FRRPROD and FRRMEAN respectively. For MFPC two aggregation operators
were selected; OWA and AIWA, MFPCOWA and MFPCAIWA respectively. MFPC
further allows for parameter tuning, which was done by allowing a grid-search
for the parameter values suggested in [8].
For the proposed FPC-GA and FPC-LGA configurations were created to
match FRR and MFPC. Two FPC-GA variants similar to FRRs parameters
were included: Both restricted to using a -function for membership function
and for fuzzy aggregation rule, one with the product and another with the
arithmetic mean, FPC-GAPROD and FPC-GAMEAN respectively. For comparison
with MFPC, two variants were configured with the -membership function the
fuzzy aggregation operator replaced with OWA and AIWA respectively. In FPCGA the genetic algorithm was allowed to set parameter values for OWA/AIWA,
while for FPC-LGA the parameters were fixed to those chosen by the corresponding MFPCs grid-search. For FPC-(L)GAs genetic algorithm, 100 chromosomes
and 100 iterations was used, together with a mutation probability of 0.3 and
tournament selection.
4.3
Figure 3 show an example of membership functions from FRR (heuristic approach), FPC-GA and FPC-LGA classifiers. Similar visualizations were created
for the other datasets, with similar results, but, omitted for clarity. The GA
searched fuzzy membership functions are not easily identifiable or interpretable
as the ones produced by the heuristic approach, however regions appear similar.
1.0
1.0
0.8
0.8
0.6
0.6
Membership
Membership
0.4
0.2
0.4
0.2
0.0
0.0
0.0
0.2
0.4
Value
0.6
0.8
1.0
0.0
0.2
0.4
Value
0.6
0.8
1.0
1.0
1.0
0.8
0.8
0.6
0.6
Membership
Membership
0.4
0.2
0.4
0.2
0.0
0.0
11
0.0
0.2
0.4
Value
0.6
0.8
1.0
0.0
0.2
0.4
Value
0.6
0.8
1.0
Fig. 3: Example of heuristic (solid) and learned (dashed) -functions of two features
from the Iris dataset for FPC-GAPROD and FPC-LGAPROD classifiers. Each class is a
separate color.
Accuracy
Reported results are mean accuracy and standard deviation from conducting the
experiment using a strategy of 10-fold stratified cross-validation and repeating
the classification 100 times (100 10 CV). The results are reported in two sets:
Table 3 reports comparison with the FRR classifier, and Table 4 reports the
same for MFPC.
Considering the accuracy, the FPC-(L)GA classifiers outperform the heuristic
FRR and MFPC classifiers on most datasets with a clear higher average ranking.
The Friedman test was used to compare pairs of classifiers [4]. The average
rankings are shown above in Table 3 and Table 4.
12
Table 3: Mean accuracy and standard deviation. Best classifier is marked with bold.
#
Avg rank
1
2
3
4
5
6
7
8
9
10
11
12
FRRprod
10.87500
FRRmean
7.58333
53.56(9.56)
67.71(1.45)
74.35(1.87)
44.41(4.68)
34.58(1.41)
29.23(8.83)
71.78(2.58)
70.91(1.61)
50.07(13.69)
32.71(0.26)
35.90(0.06)
10.03(2.36)
77.93(5.98)
70.96(1.86)
94.06(1.58)
56.72(2.34)
69.92(7.71)
49.29(3.45)
68.25(6.81)
70.56(1.93)
64.91(10.13)
41.18(1.94)
56.56(10.87)
53.30(4.71)
84.22(7.16)
70.89(2.34)
92.13(2.68)
55.14(2.20)
69.37(8.45)
60.93(6.00)
49.24(11.15)
66.25(2.51)
67.89(9.22)
56.65(10.95)
73.44(5.23)
50.43(13.21)
93.19(4.65)
72.77(2.24)
88.56(20.78)
55.98(3.64)
73.13(11.78)
58.49(11.01)
61.41(15.97)
66.55(4.63)
70.54(9.40)
64.50(10.10)
35.90(0.06)
50.05(4.35)
92.22(2.32)
70.24(1.41)
95.69(1.22)
59.80(3.99)
85.39(4.68)
62.62(6.43)
65.32(6.36)
71.31(0.83)
64.80(11.94)
56.00(7.47)
70.56(4.94)
70.93(7.62)
However, at borderline significance, p < 0.1, several other classifiers were different as well, favoring FPC-GA and FPC-LGA over FRR and MFPC. Full details
have been omitted for space constraints.
4.5
Comparing FPC-GA and FPC-LGA was also done by using Friedman and Nemenyi statistics. The results are shown in Table 5. At p < 0.05 the only difference is with different fuzzy aggregation operators, namely FPC-GAPROD and
FPC-LGAMEAN . However, considering borderline significance, also with different averaging there is difference between FPC-GA and FPC-LGA. However, in
ranking, FPC-LGA is a clear winner over FPC-GA. Again, the differences between the fuzzy aggregation operators suggest that it is an important parameter
to optimize by the GA as well.
4.6
Overfitting/underfitting
A potential problem with learning membership functions over the heuristic approach is that the classifier may adjust the membership functions too well to the
13
Table 4: Mean accuracy and standard deviation. Best classifier is marked with bold.
#
MFPCAIWA MFPCOWA FPC-GAAIWA FPC-GAOWA FPC-LGAAIWA FPC-LGAOWA
Avg rank 10.04167
8.95833
7.25000
5.75000
4.08333
3.33333
1
2
3
4
5
6
7
8
9
10
11
12
88.00(3.67)
63.59(6.48)
90.66(5.76)
52.01(3.23)
76.04(9.28)
53.55(6.57)
57.80(9.91)
58.02(6.16)
66.49(13.94)
31.96(8.06)
35.90(0.06)
37.58(6.82)
92.96(2.35)
67.90(4.84)
91.20(6.42)
54.72(3.93)
77.98(9.44)
57.07(6.87)
54.50(12.98)
60.83(5.76)
69.50(14.10)
31.80(8.03)
35.90(0.06)
40.94(8.15)
76.00(8.38)
84.74(8.14)
71.59(1.56)
71.67(2.37)
92.79(2.12)
93.67(1.78)
57.33(4.91)
56.75(2.82)
70.05(8.41)
71.28(7.96)
58.96(6.81)
57.64(7.61)
56.17(12.16) 54.55(15.41)
65.15(2.85)
65.07(3.37)
66.16(11.66) 69.46(6.16)
52.47(7.74) 57.89(8.13)
72.53(6.79)
73.82(3.72)
50.62(9.62)
54.53(7.45)
93.33(2.32)
72.15(1.03)
95.44(1.49)
58.84(2.62)
84.46(4.66)
52.62(7.07)
60.95(10.25)
70.98(1.22)
66.24(11.50)
57.76(8.80)
74.01(3.26)
58.91(6.02)
92.30(3.72)
71.48(1.21)
95.72(1.01)
60.45(2.87)
87.27(4.14)
64.75(1.98)
66.07(11.60)
71.12(0.98)
67.46(11.75)
55.41(7.28)
69.83(2.76)
68.68(7.58)
Table 5: Nemenyi statistic p-values and Friedman average ranks. Significant differences
(p 0.05) marked with bold.
FPC-GAPROD FPC-GAMEAN FPC-GAAIWA FPC-GAOWA FPC-LGAPROD FPC-LGAMEAN FPC-LGAAIWA Avg. rank
FPC-GAPROD
FPC-GAMEAN
FPC-GAAIWA
FPC-GAOWA
FPC-LGAPROD
FPC-LGAMEAN
FPC-LGAAIWA
FPC-LGAOWA
0.99990
0.99998
0.84991
0.76055
0.19522
0.16187
0.03318
1.00000
0.97458
0.94135
0.42572
0.37213
0.10798
0.96042
0.91676
0.37213
0.32179
0.08690
1.00000
0.96042
0.94135
0.65411
0.98457
0.97458
0.76055
1.00000
0.99781
0.99907
6.00000
5.58333
5.66667
4.58333
4.41667
3.50000
3.41667
2.83333
training examples, and hence give poor accuracy on the unseen testing examples;
this problem is known as overfitting. A common way to determine if a classifier
is overfitting, is to consider the difference in accuracy between the training data
and testing data. A larger difference means that the model is overfitting, while
a small difference means no overfitting [1]. Table 6 shows the mean difference
between test and training.
The FRR and MFPC classifiers have the lowest mean difference between the
test and training data, indicating least overfitting, however their classification
accuracy is consistently ranked below the FPC-(L)GA classifiers. The primary
parameter for deciding how accurate to learn the -functions, is the number of
iterations in the GA. Intuition is that a large number of iterations may learn
-functions too well, resulting in overfitting. To test this, an experiment with the
FPC-GA and FPC-LGA classifiers was conducted, where number of iterations for
the GA was {10, 50, 100, 500, 1000}. Figure 4 show the results for the classifiers
FPC-GAPROD and FPC-LGAPROD on all datasets. Again the experiment was
repeated 100 times and the mean accuracy used. Except for a single dataset
(Ionosphere) the classifiers reach a stable accuracy around 50 to 100 iterations,
and performance does not seem to decrease over number of iterations, indicating
the number of iterations can be chosen freely. The Ionosphere dataset has the
14
Table 6: Mean difference between test and training accuracy (all datasets)
Classifier
FRRprod
FRRmean
FPC-GAPROD
FPC-GAMEAN
FPC-LGAPROD
FPC-LGAMEAN
MFPCAIWA
MFPCOWA
FPC-GAAIWA
FPC-GAOWA
FPC-LGAAIWA
FPC-LGAOWA
Training
Testing
Difference
58.81(17.30)
73.28(8.99)
89.50(12.79)
94.88(5.05)
85.06(17.03)
88.05(8.23)
68.61(19.22)
73.75(20.65)
95.07(4.81)
94.80(5.06)
90.77(7.14)
89.09(8.04)
47.94(19.53)
64.47(13.54)
63.65(13.26)
66.38(12.46)
65.92(15.02)
72.07(12.06)
59.30(18.29)
61.27(18.98)
65.82(11.36)
67.59(11.92)
70.48(13.59)
72.54(12.03)
10.87(10.81)
8.81(7.64)
25.85(13.72)
28.50(11.50)
19.13(14.86)
15.98(10.06)
9.31(9.13)
12.48(10.20)
29.25(10.71)
27.21(10.50)
20.30(12.49)
16.55(10.51)
most features, which reflects in the number of genes required. This has effect on
the number of iterations needed for the search to complete.
1.0
0.9
0.9
0.8
0.8
0.7
1
2
3
4
5
6
7
8
9
10
11
12
0.6
0.5
0.4
0.3
0
200
400
Iterations
600
800
1000
Accuracy
Accuracy
1.0
0.7
1
2
3
4
5
6
7
8
9
10
11
12
0.6
0.5
0.4
0.3
0
200
400
Iterations
600
800
1000
Fig. 4: Accuracy for all datasets for two classifiers using a variable number of iterations.
15
Conclusions
In this paper new a new fuzzy pattern classifier with genetic learning of fuzzy
membership functions and fuzzy aggregation operator was proposed, based on
the existing recent work in fuzzy pattern classifiers.
The experimental results had both the Friedman and Nemenyi tests conducted to evaluate the significance of the proposed classifiers, and results were
encouraging the experiment present a potential usefulness of GA learned FPCs
and may give direction for what next steps could be. The experiment addressed
the overfitting problem where GA learned classifier slightly overfits the original
classifier, however, with a better accuracy in all cases.
As indicated by the average ranking, the choice of fuzzy aggregation operator
is important, but in the experiments does not allow FPC-(L)GA to search it, as
given in our initial limitations. As the membership functions are only one part
of tuning a FPC, this area is a future direction for research. Another interest
is the overfitting problem, where mentioned alternatives to reducing number of
iterations is to be done.
References
1. Goncalves, I., Silva, S.: Experiments on controlling overfitting in genetic programming. In: Proc. 15th Portuguese Conference on Artificial Intelligence. (2011)
152166
2. Herrera, F.: Genetic fuzzy systems: taxonomy, current research trends and
prospects. Evolutionary Intelligence 1(1) (2008) 2746
3. H
ullemeier, E.: Fuzzy sets in machine learning and data mining. Applied Soft
Computing 11(2) (2011) 14931505
4. Japkowicz, N., Shah, M.: Evaluating learning algorithms: A classification perspective. Cambridge University Press (2011)
5. Klir, G.J., Yuan, B.: Fuzzy Sets and Fuzzy Logic: Theory and Applications. Prentice Hall (1995)
6. Kuncheva, L.I.: Fuzzy classifier design. Volume 49 of Studies in Fuzziness and Soft
Computing. Springer (2000)
7. Meher, S.K.: A new fuzzy supervised classification method based on aggregation
operator. In: Proc. 3rd Int. IEEE Conf. on Signal-Image Technologies and InternetBased Systems. (2008) 876882
8. M
onks, U., Larsen, H.L., Lohweg, V.: Aggregation operator based fuzzy pattern
classifier design. In: Proc. Machine Learning in Real-Time Applications. (2009)
9. Rutkowski, L.: Flexible neuro-fuzzy systems: Structures, learning and performance
evaluation. Kluwer (2004)
10. Whitley, D.: A genetic algorithm tutorial. Statistics and Computing 4 (1994)
6585