Knowledge-Based Systems 116 (2017) 144–151
Contents lists available at ScienceDirect
Knowle dge-Base d Systems
journal homepage: www.elsevier.com/locate/knosys
Data classification using evidence reasoning rule
Xiaobin Xu
a
,
∗
, Jin Zheng
a
, Jian-bo Yang
b
, Dong-ling Xu
b
, Yu-wang Chen
b
a
Institute of System Science and Control Engineering, Hangzhou Dianzi University, Hangzhou 310018, Zhejiang, China
b
Decision and Cognitive Sciences Research Centre, The University of Manchester, Manchester M15 6PB, UK
a r t i c l e i n f o
Article history:
Received 16 June 2016
Revised 21 August 2016
Accepted 1 November 2016
Available online 2 November 2016
Keyword:
Date classification
Dempster–Shafer evidence theory (DST)
Evidential reasoning (ER) rule
Reliability and weight of evidence
Sequential linear programming (SLP)
a b s t r a c t
In Dempster–Shafer evidence theory (DST) based classifier design, Dempster’s combination (DC) rule is
commonly used as a multi-attribute classifier to combine evidence collected from different attributes.
The main aim of this paper is to present a classification method using a novel combination rule i.e., the
evidence reasoning (ER) rule. As an improvement of the DC rule, the newly proposed ER rule defines
the reliability and weight of evidence. The former indicates the ability of attribute or its evidence to
provide correct assessment for classification problem, and the latter reflects the relative importance of
evidence in comparison with other evidence when they need to be combined. The ER rule-based clas-
sification procedure is expatiated from evidence acquisition and estimation of evidence reliability and
weight to combination of evidence. It is a purely data-driven approach without making any assumptions
about the relationships between attributes and class memberships, and the specific statistic distributions
of attribute data. Experiential results on five popular benchmark databases taken from University of Cal-
ifornia Irvine (UCI) machine learning database show high classification accuracy that is competitive with
other classical and mainstream classifiers.
©2016 Elsevier B.V. All rights reserved.
1.
Introduction
Classification problem is one of the most important issues in
data mining and knowledge discovery [1] . Its purpose is to fall a
sample with unknown class into a basket with a label of specific
class where an appropriate classifier should be used to analyze the
attributes of this sample. Classification are fundamental to many
theoretical and practical applications, including pattern recogni-
tion [2–4] , fault diagnosis [5–7] , and image processing [8–10] , etc.
Many well-known methods have been proposed to solve classifica-
tion problems, including k nearest neighbors (k-NN) [11] , support
vector machine [12] , naive Bayes [13] , Bayes net [14] , decision tree
learner [15] , random forest [16] , and other latest techniques, such
as gravitational inspired classifier [17] , feature vector graph-based
classifier [18] , and Learning automata(LA)-based classifier [19] , and
so on.
From the perspective of uncertain information processing, the
imprecision or even incorrectness of classification results is likely
to be caused by the fact that the values of attributes of a sam-
ple cannot categorically point to a certain class, that is to say, the
boundaries of attributes among different classes are commonly im-
precise, or even overlapping [20–22] . As a result, Dempster–Shafer
evidence theory (DST) can provide an available mechanism to deal
∗
Corresponding author.
E-mail address: xuxiaobin1980@163.com (X. Xu).
with the classification imprecision. In detail, a frame of discern-
ment (FoD) needs to be firstly determined which includes all pre-
assigned class memberships. The next step is to obtain a basic be-
lief assignment (BBA), i.e., a belief distribution (BD) function, in
which the belief degrees are used to measure the extents to which
the attributes of a sample supports each class and the subsets of
the classes. Such a BBA or a BD can be also named as a piece of ev-
idence. There are different ways for generating BBAs from different
types of attribute information. The typical ways include core sam-
ple [20] , neural network [21] , k-NN [22] , expert system [23] and
so on. The final step is to use Dempster’s combination (DC) rule to
fuse these BBAs and then make a classification decision according
to the fused results. The aim of combination is to reduce the clas-
sification imprecision by fusing multi-source attribute information.
Recently, the evidential reasoning (ER) rule has been estab-
lished to advance the seminal Dempster–Shafer evidence theory
[24–28] and the original ER algorithm [29–32] . Compared with the
DC rule, the main advance of the ER rule is to propose a novel con-
cept of weighted evidence (WE) and extend to WE with Reliability
(WER) in order to characterize evidence in complement of BBA or
BD introduced in the DST. As a result, the implementation of the
orthogonal sum operation on WEs and WERs leads to the estab-
lishment of the new ER rule [33] . The most important property of
the ER rule is that it constitutes a generic conjunctive probabilis-
tic reasoning process, or a generalized Bayesian inference process
which can be implemented on the power set of FoD. It has been
http://dx.doi.org/10.1016/j.knosys.2016.11.001
0950-7051/© 2016 Elsevier B.V. All rights reserved.