Incremental Naïve Bayesian Learning Algorithm Based On .

3y ago
33 Views
3 Downloads
420.90 KB
8 Pages
Last View : 24d ago
Last Download : 3m ago
Upload by : Joanna Keil
Transcription

JOURNAL OF COMPUTERS, VOL. 9, NO. 8, AUGUST 20141967Incremental Naïve Bayesian Learning Algorithmbased on Classification Contribution DegreeShuxia Ren *, Yangyang LianSchool of Computer Science & Software Engineering, Tianjin Polytechnic University, Tianjin, Chinat rsx@126.com , sunshine lianyy@qq.comXiaojian ZouMilitary Transportation University, Tianjin, Chinaala 99@126.comAbstract—In order to improve the ability of graduallearning on the training set gotten in batches of NaiveBayesian classifier, an incremental Naïve Bayesian learningalgorithm is improved with the research on the existingincremental Naïve Bayesian learning algorithms. Aiming atthe problems with the existing incremental amendingsample selection strategy, the paper introduced the conceptof sample Classification Contribution Degree in the processof incremental learning, based on the comprehensiveconsideration about classification discrimination, noisy andredundancy of the new training data. The definition andtheoretical analysis of sample Classification ContributionDegree is given in this paper. Then the paper proposed theincremental Naïve Bayesian classification method based onthe Classification Contribution Degree. The experimentalresults show that the algorithm simplified the incrementallearning process, improved the classification accuracy ofincremental tion Degree, Naïve BayesianClassificationI. INTRODUCTIONData mining has been becoming more and morepopular with all walks of life, from manufacturing toservice industry, from medical diagnose to businessdecision, because its special advantage in discovering thepotential decision information and business model [1].Classification prediction is an important content in theresearch of data mining, which has been widely used inclassification prediction and pattern recognition.There are many classification methods such as decisiontree, Naïve Bayesian classifier, support vector machine,neural network and so on. Naïve Bayesian classificationmethod is widely used for its simple and efficiency. Itsclassification performance is comparative to neuralnetwork, decision tree and better than many otherclassifier in some cases [2]. The classification rulestrained on a large scale of dataset with strongcompleteness can accurately response model of thepractical problem. But in variety of applications such asmedical decision making, stock analysis, industrial 2014 ACADEMY PUBLISHERdoi:10.4304/jcp.9.8.1967-1974production, data are produced over time, and the trainingset are gained patched [3].In order to ensure the classifier gotten by the NaïveBayesian classification method more accuracy, theclassifier must be retrained again and again whenever thetraining set is updated as the time goes by. The problemsof wasting time and memory resources arise because ofrepeat training [4].Incremental learning can amend the current classifierwith knowledge learning from the new training set [5]. Itcan make full use of the previous trained classifier,making the previous work meaning. Meanwhile,incremental learning only needs new training set inmemory, making the amending process fast and efficient.Compared with training a new system, amending atrained system will save lots of resources like time, space,manpower and material. Moreover, incremental learningcan solve some problems of real-time application moreaccurately on the basis of small-scale study.Incremental learning for a mining algorithm, especiallythe classification mining algorithms, is a very importantability. Many studies of incremental learning ability weredown with many classification methods like RBF neuralnetwork, Support vector and k-Nearest Neighbor [6-9].And the applications of incremental classification focuson predicting bugs in software, pattern classification,mining of frequent time sequence, analysis of medicaldata features, collaborative filtering and so on [10-12].The research of incremental ability with naïveBayesian classifier was first proposed by Gong Xiujun.He presented the learning formula of incremental NaïveBayesian and selection strategy of amending samples [13].Many improving methods on the select range ofamending sample and sequence are made based on thispaper. There are three kinds of incremental NaïveBayesian classification algorithms.The first kind of algorithm uses all the samples in thenew training set for amending the current classifier, butprefer for samples whose sum loss of classification islesser, which was first proposed by Gong. The secondkind of algorithm predicts samples in the new training setwith the current classifier first, and uses samples that are

1968correctly classified for amending the current classifier[14]. The third kind of algorithm also predicts samples inthe new training set with the current classifier first, anduses samples that are wrongly classified to amend thecurrent classifier. In the process of amending, thesealgorithms still prefer for samples whose sum loss ofclassification is lesser [15]. Correctly classified samplesare samples whose predicted class labels are the samewith their own class labels. And wrongly classifiedsamples are samples whose predicted class labels aredifferent from their own class labels.The problems of three kinds of incremental NaïveBayesian classification algorithms are as follows:First, in the incremental learning process, thealgorithms do not fully consider the factor of datarepresentation, data redundancy and noise. Weakcharacterized data and redundant data will blur thecategory features. And the noisy data will decline theperformance of the classifier. Obviously, it is notadvisable to use all the new training samples to amendthe current classifier [16]. In fact, whether a new instanceis suit for amending the current classifier, we shouldconsider the data qualities that affect the predict result,the representativeness, discrimination degree ively, instead of whether the sample is rightlyor wrongly predicted.Second, when the order of samples is different, thelearning sequence obtained by the sum loss ofclassification is different with the same training set [17].There are two extreme cases for example. When samplesin new training set are descending according to theclassification sum loss, all samples will be used to amendthe current classifier. When samples in new training setare ascending according to the classification sum loss,only the first sample will be used to amend the currentclassifier.In order to overcome the above problems, the conceptof Classification Contribution Degree is presented rimination, redundancy and noisy of the data, andthen the Incremental Naïve Bayesian Learning Algorithmbased on Classification Contribution Degree (INB-CCD)is proposed. The algorithm analyzes the effect onclassification performance from the level of data quality,and avoids the dependence on sample order of learningresult.The rest of the paper is organized as follows. Section 2presents the Naïve Bayesian classification method andfurther gives the learning formula of incremental NaïveBayesian classification. In Section 3 the paper first givesthe definition and theoretical analysis of ClassificationContribution Degree, and then proposes the sampleselection strategy of Naïve Bayesian incremental learningand at last gives the medical data mining algorithm basedon Classification Contribution Degree. The effectivenessof the algorithm is studied in Section 4 throughcomparing its performance on five different medicaldatasets with the other three incremental learningalgorithms. Finally, the paper concludes in Section 5. 2014 ACADEMY PUBLISHERJOURNAL OF COMPUTERS, VOL. 9, NO. 8, AUGUST 2014II. LEARNING FORMULA OF INCREMENTAL NAÏVEBAYESIANA. Naïve Bayesian Classification MtheodNaïve Bayesian classification method is mainly basedon the Naïve Bayesian principle. So before theintroduction of incremental Naïve Bayesian formula, wefirst present the Naïve Bayesian theory here. It gives thecalculation of post probability according to the priorprobability. Prior probability is the probabilitydistribution of each class counted by the given datainformation. It contains two statistics: class priorprobability and conditional probability. Post probabilityis the probability of each unclassified sample belongingto some certain class.Suppose X { A1 , A2 ,., An } is a data sample with nfeatures, then post probability of sample X belonging toclass C is P (C X ) . The Naïve Bayesian formula ofcalculating the post probability is as follows:P( X C ) P(C )(1)P(C X ) P( X )Where, P(C) is the class prior probability of a samplebelonging to class C. P(X C) is conditional probability offeatures same with sample X among samples whose classlabel is C.The main idea of Naïve Bayesian classification methodis as follows:Suppose C1 , C2 ,., Cm represents the m different classes.For each test sample X, the classification methodcomputes the posterior probability P(C j X ) throughclass prior probability P (C C j ) and class conditionalprobability P( X C j ) . The value range of j is from 1 to m.And sample X belongs to the class whose posteriorprobability is the biggest of the all [18]. That is to say,Naïve Bayesian classification method classified thesample X as class Ci , when the sample X satisfied thefollowing expression:P(Ci X ) P(C j X ) , 1 i j m(2)We can learn from (1) that (2) equals to the followingexpression:P( X Ci ) P(Ci ) P( X C j ) P(C j )(3) P( X )P( X )Where, 1 i j mMoreover, (2) is equal to the following expression:(4)P( X Ci ) P(Ci ) P( X C j ) P(C j )Where, 1 i j mTherefore, the essence of training Naive Bayesianclassifier is to calculate two parameters using statisticalknowledge according to the training samples: priorprobability and conditional probability.B. Incremental Naïve Bayesian FormulaThe process of classification forecasting with NaïveBayesian classifier is as follows:

JOURNAL OF COMPUTERS, VOL. 9, NO. 8, AUGUST 2014Firstly, calculating the posterior of test samplebelonging to each class with the two parameters obtainedfrom the new training information using Naïve Bayesiantheory, and then dividing the test sample to the class withthe largest posterior.Therefore, the task of incremental learning of NaiveBayesian classifier is equal to how to determine the newprior probability and class conditional probabilityaccording to the prior information and new traininginformation. The incremental amending process is shownin Fig. 1.1969 m1when c p c j and Ai akθik j 1 m 1 m mwhen c p c j and Ai akθ ik j ' θik j 1 m θwhen c p c j ik j(6)m Ai count (c j ) Ai represents the numbers of values for attribute Ai ,count (c j ) represents number of samples whose classlabel is c j .Figure 1. Incremental Amending of Naïve Bayesian classifierIt can be learnt from the Fig. 1 that, correction processof the incremental learning of Naïve Bayesian classifier isactually a recursive Bayesian estimation of parameters.Its advantage is that it can preserve the information ininitial training data in the form of parameters. During theprocess of incremental learning, the system does not needto visit the raw data, but only needs to access the twosaved statistic parameters. According to the informationin new training set, the system amends and resaves thetwo statistic parameters. So, the incremental learningformula of Naïve Bayesian classifier is deduced by thestatistical knowledge.Suppose D is the current training set, T is the newtesting set, x p ( A1 , A2 ,. Ai ,., An ) T is the newinstance for amending, andcpmeans its class label.θ j P (c c j ) is class prior probability of class labelcjThen the incremental amending of Naïve Bayesianclassifier is completed after the modification of the twostatistical parameters according to the (5) and (6). Inorder to illustrate the incremental learning process ofNaive Bayesian classifier with two statistic parametersclearly, Fig. 2 shows incremental learning model ofNaive Bayesian classifier with the three probabilityvalues of Naive Bayesian theory.Figure 2. Incremental learning model of Naive Bayesian classifier. Then the amending formula of class priorprobability is:1 swhen c p c jθj 1 s(5)θ j' s 1s when c p c jθj s 1Where, s D T D represents number of samples in training set D, T represents number of samples in new testing set T.Suppose θ ik j P ( Ai ak c c j ) is class priorprobability of attribute Ai of value ak in class c j . Thenthe amending formula of class conditional probability is:III. INCREMENTAL NAÏVE BAYESIAN CLASSIFICATIONALGORITHMA. The Definition and Theory Analysis of ClassificationContribution DegreeThis paper defined the classification earnings obtainedby the amending sample, namely the degree of reductionbetween classification results and sample observations, asclassification contribution degree, which is denotedas con . Suppose T ( x1 , c1 xi , ci . x p , c p )is the new samples set. The testing class label of samplexiisci'obtained by the current classifier, andci''obtained by classifier amended with sample xk , ck .Then the classification contribution degree of new testingsample xk is: 2014 ACADEMY PUBLISHER

1970JOURNAL OF COMPUTERS, VOL. 9, NO. 8, AUGUST 2014conk (ci ci' ci ci'' )(7)i kcc' is distance between testing class label c ' and'Expression②:max{count (ci ci'' 0) count (ci ci' 0)} T 1 observation class label c , and the value of distance cc is: max{ ( 0, c c 'cc' ' 1, c c min{(8)Therefore, 1 ci ci' ci ci'' 1 0 ''i i''i iwhen c c 0 and c c 1when c c 1 and c c 0(9)when c c ci ci'' 0 or 1And the classification accuracy of classified amendedby sample xk , ck is:count (ci ci'' 0),1 i k T T 1 Then the classification earningsimprovement of sample xk , ck is:oraccuracycount (ci ci'' 0) count (ci ci' 0),1 i k T T 1 Obviously, the sample with the largest classificationearnings is most favorable for classification accuracyimprovements.The theoretical proof of classification contributiondegree is as follows:Expression①: max{count (ci ci'' 0) count (ci ci' 0)} T 1 count (ci ci'' 1)count (ci ci' 1)) (1 )} T 1 T 1 count (ci ci' 1) count (ci ci'' 1)} T 1 2014 ACADEMY PUBLISHERmax{ maxcount (ci ci' 0),1 i T T max{(1 count (ci ci' 0) count (ci ci'' 0)} T 1 Comprehensive expression ① and ②:'i i'i i'i iThrough the analysis of (9), there are three cases aboutthe new classifier amended by a sample. If the testingclass label changes from 0 to 1, then the classificationcontribution degree of this sample is -1, and the samplehas a negative influence to the improving of classificationaccuracy. If the testing class label turns from 1 to 0, thenthe classification contribution degree of this sample is 1,and the sample has a positive influence to the improvingof classification accuracy. If the testing class labelchanges from 0 to 0 or from 1 to 1, then the classificationcontribution degree of this sample is 0, and the samplemakes no difference on the improving of classificationaccuracy.It can be learnt from (8) that the classification accuracyof current classifier is:max{count (ci ci' 0) count (ci ci'' 0))} T 1 count (ci ci'' 0) count (ci ci' 0)} T 1 c c i k T max'i i (c c1 i k T min'i i c c i k T ''i i''i i c c ) max conkVisibly, the sample classification contribution degreecan response the improved degree of classification byusing the sample to amend the current classifier, namelythe classification earnings gotten by the amending sample.B. The Selection Strategy and Model of IncrementalLearningBecause the amending sample sequence gotten byclassification sum loss has a strong dependence onsample order in new training set, so the incrementallearning result with the same training set may be different.This section comprehensively analyzes the factorsaffecting the prediction result in order to get a betterselection strategy of amending samples. These factors arequalities of data, classification discrimination,representativeness, noisy and redundancy of traininginformation, mentioned in section I.There are two results using current classifier to test thenew testing sample set: rightly classified and wronglyclassified. The rightly classified samples demonstrate thecurrent training set contains the classification informationcarried by the rightly classified samples in the new testingset. However, there are two reasons for wronglyclassified samples. One result is that the datacompleteness of current training set is bad. The currenttraining set doesn’t contain the classification informationcarried by the wrongly classified samples. The otherresult is that the wrongly classified sample is noisy data.In order to avoid the decline of classification accuracyand feature weakening brought by data noisy andredundancy, enriching the current training set with thestrong classification representative samples, and avoidingthe dependence on sample order of learning result, thepaper amends the current classifier with two sampleswhose classification contribution degree is the largestamong all rightly classified samples and wronglyclassified samples separately.As is known from (7), the sample with the largestclassification contribution degree is satisfied with thefollowing condition.cont con p(10)

JOURNAL OF COMPUTERS, VOL. 9, NO. 8, AUGUST 2014Where, 1 t , p T , t pObviously, sample with the largest classificationcontribution degree can improve the classificationperformance of the classifier best among all the samples.That is to say, sample with the largest classificationcontribution degree has strong representativeness, carriesmore classification information among all the samples innew training set. And it must not be noisy andredundancy data. Therefore, enriching the current trainingset with this sample can enrich the classificationinformation of the training set effectively.The main idea of Incremental learning based onclassification contribution degree is that: testing the newtraining set with the current classifier first, puttinginstances whose class label are same with their testedlabel to the rightly classified sample set, and puttinginstances whose class label are different from their testedlabel to the wrongly classified sample set; then amendingthe current classifier with the samples whoseclassification contribution degree is the largest among allthe rightly tested samples and wrongly tested samplesseparately.Incremental learning model based on classificationcontribution degree is showed in Fig. ree coni of each sample xi Tright , putting sample t1with the largest classification contribution degree into butiondegree con j of each sample x j Terror , putting samplet2 with the largest classification contribution degree intoset ‘Update’Step5: amending the current classifier with samples in'set ‘update’, and getting the amended classifier nb .The work flow of the incremental Naive Bayesianclassification algorithm based on the classificationcontribution degree is shown in Fig. 4.Figure 3. Incremental learning model based on classificationcontribution degreeC. Incremental Naïve Bayesian Classification AlgorithmThe incremental Naive Bayesian classificationalgorithm based on the classification contribution degreefirst predicts samples in the new training set with thecurrent classifier, and adds the rightly classified samplesto set 'Tright', wrongly classified samples to set 'Terror'.Then the algorithm calculates classification contributiondegree of each sample, and puts the samples with thelargest classification contribu

Incremental learning for a mining algorithm, especially the classification mining algorithms, is a very important ability. Many studies of incremental learning ability were down with many classification methods like RBF neural network, Support vector and k-Nearest Neighbor [6-9]. And the applications of incremental classification focus

Related Documents:

Some other works on incremental learning and its applications include the incremental learning fuzzy neural (ILFN) network for fault detection and classification [5], incremental learning for multi-sensor data fusion [6], incremental genetic learning for data classification [7], incremental semi-supervised learn-ing [8], incremental learning .

44 Incremental Backups: John Snow; FOSDEM 2017 Life Cycle - First Incremental (The first step of our journey) Example 3: Create an incremental backup. Can be done via transaction or single QMP command. { "execute": "drive-backup", "arguments": {"device": "drive0", "bitmap": "bitmap0", "target": "inc.0.qcow2", "format": "qcow2", "sync": "incremental",

Data Mining: Practical Machine Learning Tools and Techniques (Chapter 6) 3 Implementation: Real machine learning schemes Numeric prediction Regression/model trees, locally weighted regression Bayesian networks Learning and prediction, fast data structures for learning Clustering: hierarchical, incremental, probabilistic Hierarchical, incremental, probabilistic, Bayesian

Computational Bayesian Statistics An Introduction M. Antónia Amaral Turkman Carlos Daniel Paulino Peter Müller. Contents Preface to the English Version viii Preface ix 1 Bayesian Inference 1 1.1 The Classical Paradigm 2 1.2 The Bayesian Paradigm 5 1.3 Bayesian Inference 8 1.3.1 Parametric Inference 8

value of the parameter remains uncertain given a nite number of observations, and Bayesian statistics uses the posterior distribution to express this uncertainty. A nonparametric Bayesian model is a Bayesian model whose parameter space has in nite dimension. To de ne a nonparametric Bayesian model, we have

Incremental learning refers to learning from streaming data, . and applications which emerged in the last years. . ent possibilities to assess the performance of incremental learning algorithms: (1) Incremental-vs- non-incremental: In particular in the absence of concept

Currently FUFP, pre-FUFP and IMBT algorithms have been developed that support incremental learning. The IMBT uses a binary tree data structure called an Incremental mining binary tree. This work proposes a novel incremental learning algorithm that makes use of a data structure called Item-Itemset(I-Is) tree that is a variation of B tree.

During the American Revolution both the American Continental Army and the British Army had spies to keep track of their enemy. You have been hired by the British to recruit a spy in the colonies. You must choose your spy from one of the colonists you have identified. When making your decisions use the following criteria: 1. The Spy cannot be someone who the Patriots mistrust. The spy should be .