Professional Documents
Culture Documents
HTTPS://SITES.GOOGLE.COM/SITE/JOURNALOFCOMPUTING/
WWW.JOURNALOFCOMPUTING.ORG 20
Abstract— In inductive inference one of the widely used practical methods is learning with decision tree. The method is used at
discrete structure objective function that is shown as the decision tree of the learning function. In this study ID3, fuzzy ID3
(F_ID3) and probabilistic fuzzy ID3 (PF_ID3) decision tree algorithms are used in a comparative manner for evaluation of
learning achievements of students. For this purpose, test difficulty (high, middle, low), assignment time (long, average, short),
and score (very good, good, average, bad, very bad) attributes are taken for the tests (A, B, C, D, E) which are applied to
randomly selected students from determined classroom. According to the academic average students are classified as
“success” and “fail” and from the obtained results it was observed that the algorithms with fuzzy and probabilistic structures
given better results.
—————————— ——————————
1 INTRODUCTION
entropy of such a source is zero: there is no uncertainty at 2.3 Probabilistic Fuzzy ID3 Algorithm
all. On the other hand, if a source can send n possible Another algorithm used to convert unstable classifiers
messages and each message occurs independently of the into stable structures is PF_ID3 which is made of the
preceding message with equal probability, then the un- combination of the ID3 and F_ID3 algorithms.
certainty of the receiver is maximized [6]. Probabilistic Fuzzy ID3 Algorithm [2]:
In order to measure the information which is obtained
1. Create a Root node that has a set of fuzzy data with
by dividing the database S according to test A, the ex-
membership value 1 that fits the condition of well-
pression given in (2) that is called the information gain
defined sample space.
measure is referred [9].
2. Execute the Fuzzy ID3 algorithm from step 2 to end.
Sv
G(S , A ) H ( S ) H (S v ) (2)
AndValues ( A )
S The entropy and the gain measure which are used in the
algorithm are shown in (5) and (6) respectively [3]. The
In order to decide which attribute to split upon, the criteria about vr and vn are also valid for this algorithm.
ID3 algorithm computes the information gain for each C
attribute, and selects the one with the highest gain [6]. H ob (S , A) E
c 1
AC ( x) log 2 ( E( AC ( x))) (5)
begins from the root node of the tree, and then branch
one edge to test the attribute of the sub node. Repeat the 3.2 Evaluating the Learning Achievement
testing until the leaf node is reached. The result of ID3 is In the study, only the compared results for the “A test” of
the class attached to the leaf node. the ID3, F_ID3 and PF_ID3 algorithms that are used to
The reasoning of the fuzzy decision trees is different. It evaluate the learning achievement are given. Similar re-
does not branch one edge, but all the edges of the tree. It sults are obtained from the other tests.
begins from the root node through the branches to the In Table 2, the corresponding normal and fuzzy values
leaf nodes until all the leaf nodes have been tested. Each for the levels of each quality for the A test are given. The
normal values are used for ID3 algorithm whereas the
leaf node has various proportions of all the classes. In
fuzzy values for the F_ID3 algorithm obtained with the
other words, each leaf node has own certainties of the
help of the value ranges in Table 1 and (7) are also used
classes. The result is the aggregation of the certainties at for the PF_ID3 algorithm without any need to change
all the leaf nodes. since the sum of all values on every level of every quality
TABLE 2
3 FINDINGS THE NORMAL AND FUZZY VALUES THAT CORRESPOND TO
3.1 Regulation of the Data Set THE LEVELS OF QUALITIES
F_ID3 and ID3 with the values 0,697 and 0,557 respective- thods for the A level of the SG quality is lower than 95%,
ly. The fuzziness control threshold value is taken as the expansion continues. The expansion stopped for the G
vr %95 and the leaf decision threshold value is taken as and VG levels since the ratio of being successful is 100%
vn %2. In other words, the expansion is stopped where TABLE 6
the ratio of the quality groups in the subset is smaller OBTAINED FINDINGS FOR THE SG QUALITY
than 95% and the number of samples in a node are small-
er than 2.
The examination is done only for the S and A levels of
the WT quality since they satisfy the specified criteria.
The obtained learning achievement results are given in
TABLE 4
OBTAINED FINDINGS FOR THE WT QUALITY FOR THE ID3,
F_ID3 AND PF_ID3 ALGORITHMS
in all three methods.
Therefore, the A level of the SG quality is selected after
the S level of the WT quality and the only remaining qual-
ity, TD, is examined. Only the M and H levels of the TD
quality are considered since they satisfy the leaf decision
threshold criterion. The results obtained are “fail” for the
M level and “success” for the H level.
It is observed from the results that the algorithms con-
Table 4. taining fuzzy and probabilistic structures give better re-
It is seen from the table that the ratio of successful sults. The decision tree reached for all three algorithms is
students for the S level of the WT quality is 90% in reality.
The methods that give the closest value to this ratio are
F_ID3 and PF_ID3 with a ratio of 92,3% and the ID3 me-
thod gives 73%. All the methods give with a ratio of 100%
the same results obtained in reality for the A level of the
WT quality. Therefore, the expansion is carried out since
the ratios obtained for the S level of the WT quality are
smaller than 95% whereas the expansion is stopped since
the ratios obtained for the A level are greater than 95%.
Hence, the entropy and gain values are calculated accord-
ing to the three methods only for the S level of the WT
quality. The obtained results are given in Table 5.
TABLE 5
OBTAINED FINDINGS FOR LEVEL S OF THE WT QUALITY Fig. 1. Final decision tree.
shown in Figure 1.
REFERENCES
[1] C.E. Shannon, “A mathematical theory of communication”, Bell
System Technical Journal, 27 pp.379–423 and 623–656, 1948.
[2] G. Liang, “A Comparative Study of Three Decision Tree Algo-
rithms: ID3, Fuzzy ID3 and Probabilistic Fuzzy ID3”, Bachelor
Thesis, Informatics & Economics, Erasmus University
Rotterdam, Rotterdam, the Netherlands, 2005.
[3] J.V. Berg and U. Kaymak, “On the Notion of Statistical Fuzzy
*: The quality that gives the best gain.
Entropy”, Soft Methodology and Random Information Systems,
Advances in Soft Computing, Heidelberg: Physica Verlag, pp.535–
542, 2004.
It is seen from the table that the best quality for all
[4] M. Umano, H. Okamoto, I. Hatono, H. Tamura, F. Kawachi, S.
three methods is the SG. The methods that give the best
Umedzu and J. Kinoshita, “Fuzzy Decision Trees by Fuzzy ID3
gain for the WT quality are PF_ID3 and ID3 with 0,306
algorithm and Its Application to Diagnosis Systems”, In
and F_ID3 method comes after them with the value 0,092.
Proceedings of the third IEEE Conference on Fuzzy Systems,
The expansion is carried out only for the A, G and VG
Orlando, vol. 3, pp.2113-2118,1994.
levels of the SG quality since they satisfy the vr and vn
[5] Q. Yun-Tao and X. Wei-Xin, “Building Fuzzy Neural Classifiers
criteria. Obtained results are given in Table 6.
By Fuzzy ID3 Algorithm”, Proceedings of ICSP, 1996.
Since the ratio of being successful in all the three me-
JOURNAL OF COMPUTING, VOLUME 2, ISSUE 12, DECEMBER 2010, ISSN 2151-9617
HTTPS://SITES.GOOGLE.COM/SITE/JOURNALOFCOMPUTING/
WWW.JOURNALOFCOMPUTING.ORG 24