Top Banner
DDD: A New Ensemble Approach for Dealing with Concept Drift Leandro L. Minku, Member, IEEE, and Xin Yao, Fellow, IEEE Abstract—Online learning algorithms often have to operate in the presence of concept drifts. A recent study revealed that different diversity levels in an ensemble of learning machines are required in order to maintain high generalization on both old and new concepts. Inspired by this study and based on a further study of diversity with different strategies to deal with drifts, we propose a new online ensemble learning approach called Diversity for Dealing with Drifts (DDD). DDD maintains ensembles with different diversity levels and is able to attain better accuracy than other approaches. Furthermore, it is very robust, outperforming other drift handling approaches in terms of accuracy when there are false positive drift detections. In all the experimental comparisons we have carried out, DDD always performed at least as well as other drift handling approaches under various conditions, with very few exceptions. Index Terms—Concept drift, online learning, ensembles of learning machines, diversity. Ç 1 INTRODUCTION O NLINE learning has been showing to be very useful for a growing number of applications in which training data are available continuously in time (streams of data) and/or there are time and space constraints. Examples of such applications are industrial process control, computer security, intelligent user interfaces, market-basket analysis, information filtering, and prediction of conditional branch outcomes in microprocessors. Several definitions of online learning can be found in the literature. In this work, we adopt the definition that online learning algorithms process each training example once “on arrival,” without the need for storage or reprocessing [1]. In this way, they take as input a single training example as well as a hypothesis and output an updated hypothesis [2]. We consider online learning as a particular case of incremental learning. The latter term refers to learning machines that are also used to model continuous processes, but process incoming data in chunks, instead of having to process each training example separately [3]. Ensembles of classifiers have been successfully used to improve the accuracy of single classifiers in online and incremental learning [1], [2], [3], [4], [5]. However, online environments are often nonstationary and the variables to be predicted by the learning machine may change with time (concept drift). For example, in an information filtering system, the users may change their subjects of interest with time. So, learning machines used to model these environ- ments should be able to adapt quickly and accurately to possible changes. We consider that the term concept refers to the whole distribution of the problem in a certain point in time [6], being characterized by the joint distribution pðx;wÞ, where x represents the input attributes and w represents the classes. So, a concept drift represents a change in the distribution of the problem [7], [8]. Even though there are some ensemble approaches designed to handle concept drift, only very recently a deeper study of why, when, and how ensembles can be helpful for dealing with drifts has been done [9]. The study reveals that different levels of ensemble diversity are required before and after a drift in order to obtain better generalization on the new concept. However, even though diversity by itself can help to improve generalization right after the beginning of a drift, it does not provide a faster recovery from drift in long term. So, additional strategies with different levels of diversity are necessary to better handle drifts. This paper provides an analysis of different strategies to be used with diversity, which are then combined into a new approach to deal with drifts. In all the experimental comparisons we have carried out, the proposed approach always performed at least as well as other drift handling approaches under various conditions, with very few exceptions. Section 2 explains the research questions answered by this paper in more detail. 2 RESEARCH QUESTIONS AND PAPER ORGANIZATION This paper aims at answering the following research questions, which are not answered by previous works: 1. Online learning often operates in the scenario explained in [10] and further adopted in many works, such as [2], [7], [11], and [12]: Learning proceeds in a sequence of trials. In each trial, the algorithm receives an instance from some fixed domain and is to produce a binary prediction. At IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, VOL. 24, NO. 4, APRIL 2012 619 . The authors are with the Centre of Excellence for Research in Computational Intelligence and Applications (CERCIA), School of Computer Science, The University of Birmingham, Edgbaston, Birmingham B15 2TT, United Kingdom. E-mail: {l.l.minku, x.yao}@cs.bham.ac.uk. Manuscript received 21 Nov. 2009; revised 17 May 2010; accepted 26 July 2010; published online 6 Feb. 2011. Recommended for acceptance by D. Cook. For information on obtaining reprints of this article, please send e-mail to: [email protected], and reference IEEECS Log Number TKDE-2009-11-0795. 1041-4347/12/$31.00 ß 2012 IEEE Published by the IEEE Computer Society http://ieeexploreprojects.blogspot.com
15

Ddd a new ensemble approach for dealing with concept drift.bak

Jan 27, 2015

Download

Documents

 
Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
Page 1: Ddd a new ensemble approach for dealing with concept drift.bak

DDD: A New Ensemble Approachfor Dealing with Concept DriftLeandro L. Minku, Member, IEEE, and Xin Yao, Fellow, IEEE

Abstract—Online learning algorithms often have to operate in the presence of concept drifts. A recent study revealed that different

diversity levels in an ensemble of learning machines are required in order to maintain high generalization on both old and new

concepts. Inspired by this study and based on a further study of diversity with different strategies to deal with drifts, we propose a new

online ensemble learning approach called Diversity for Dealing with Drifts (DDD). DDD maintains ensembles with different diversity

levels and is able to attain better accuracy than other approaches. Furthermore, it is very robust, outperforming other drift handling

approaches in terms of accuracy when there are false positive drift detections. In all the experimental comparisons we have carried

out, DDD always performed at least as well as other drift handling approaches under various conditions, with very few exceptions.

Index Terms—Concept drift, online learning, ensembles of learning machines, diversity.

Ç

1 INTRODUCTION

ONLINE learning has been showing to be very useful for agrowing number of applications in which training data

are available continuously in time (streams of data) and/orthere are time and space constraints. Examples of suchapplications are industrial process control, computersecurity, intelligent user interfaces, market-basket analysis,information filtering, and prediction of conditional branchoutcomes in microprocessors.

Several definitions of online learning can be found in theliterature. In this work, we adopt the definition that onlinelearning algorithms process each training example once “onarrival,” without the need for storage or reprocessing [1]. Inthis way, they take as input a single training example aswell as a hypothesis and output an updated hypothesis [2].We consider online learning as a particular case ofincremental learning. The latter term refers to learningmachines that are also used to model continuous processes,but process incoming data in chunks, instead of having toprocess each training example separately [3].

Ensembles of classifiers have been successfully used toimprove the accuracy of single classifiers in online andincremental learning [1], [2], [3], [4], [5]. However, onlineenvironments are often nonstationary and the variables tobe predicted by the learning machine may change with time(concept drift). For example, in an information filteringsystem, the users may change their subjects of interest withtime. So, learning machines used to model these environ-ments should be able to adapt quickly and accurately topossible changes.

We consider that the term concept refers to the wholedistribution of the problem in a certain point in time [6],being characterized by the joint distribution pðx; wÞ, where x

represents the input attributes and w represents the classes.So, a concept drift represents a change in the distribution ofthe problem [7], [8].

Even though there are some ensemble approachesdesigned to handle concept drift, only very recently adeeper study of why, when, and how ensembles can behelpful for dealing with drifts has been done [9]. The studyreveals that different levels of ensemble diversity arerequired before and after a drift in order to obtain bettergeneralization on the new concept. However, even thoughdiversity by itself can help to improve generalization rightafter the beginning of a drift, it does not provide a fasterrecovery from drift in long term. So, additional strategieswith different levels of diversity are necessary to betterhandle drifts.

This paper provides an analysis of different strategies tobe used with diversity, which are then combined into a newapproach to deal with drifts. In all the experimentalcomparisons we have carried out, the proposed approachalways performed at least as well as other drift handlingapproaches under various conditions, with very fewexceptions. Section 2 explains the research questionsanswered by this paper in more detail.

2 RESEARCH QUESTIONS AND PAPER

ORGANIZATION

This paper aims at answering the following researchquestions, which are not answered by previous works:

1. Online learning often operates in the scenarioexplained in [10] and further adopted in manyworks, such as [2], [7], [11], and [12]:Learning proceeds in a sequence of trials. In eachtrial, the algorithm receives an instance from somefixed domain and is to produce a binary prediction. At

IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, VOL. 24, NO. 4, APRIL 2012 619

. The authors are with the Centre of Excellence for Research in ComputationalIntelligence and Applications (CERCIA), School of Computer Science, TheUniversity of Birmingham, Edgbaston, Birmingham B15 2TT, UnitedKingdom. E-mail: {l.l.minku, x.yao}@cs.bham.ac.uk.

Manuscript received 21 Nov. 2009; revised 17 May 2010; accepted 26 July2010; published online 6 Feb. 2011.Recommended for acceptance by D. Cook.For information on obtaining reprints of this article, please send e-mail to:[email protected], and reference IEEECS Log NumberTKDE-2009-11-0795.

1041-4347/12/$31.00 � 2012 IEEE Published by the IEEE Computer Society

http://ieeexploreprojects.blogspot.com

Page 2: Ddd a new ensemble approach for dealing with concept drift.bak

the end of the trial, the algorithm receives a binarylabel, which can be viewed as the correct predictionfor the instance.

Several real-world applications operate in thissort of scenario, such as spam detection, predictionof conditional branches in microprocessors, informa-tion filtering, face recognition, etc. Besides, duringdrifts which take some time to be completed(gradual drifts), the system might be required tomake predictions on instances belonging to boththe old and new concepts. So, it is important toanalyze the prequential accuracy [13], which usesthe prediction done at each trial and considersexamples of both concepts at the same time when thedrift is gradual. The work presented in [9] considersonly the generalization calculated using a test setrepresenting either the old or the new concept. So,how would a low and a high diversity ensemblebehave considering the prequential accuracy in thepresence of different types of drift?

2. Even though high diversity ensembles may obtainbetter generalization than low diversity ensemblesafter the beginning of a drift, the study presented in[9] also reveals that high diversity ensembles presentalmost no convergence to the new concept, havingslow recovery from drifts. So, is it possible to make ahigh diversity ensemble trained on the old conceptconverge to the new concept? How? A more generalresearch question would be: is it possible to useinformation from the old concept to aid the learningof the new concept? How?

3. If it is possible, would that ensemble outperform anew ensemble created from scratch when the driftbegins? For which types of drift?

4. How can we use diversity to improve the prequentialaccuracy in the presence of drifts, at the same time aswe maintain good accuracy in the absence of drifts?

In order to answer the first three questions, we perform astudy of different strategies with low/high diversityensembles and analyze their effect on the prequentialaccuracy in the presence of drifts. The study analyzes theensembles/strategies presented in Table 1 using artificialdata sets in which we know when a drift begins and whattype of drift is present.

Before the drift, a new low diversity and a new highdiversity ensemble are created from scratch. After the driftbegins, the ensembles created before the drift are kept andreferred to as old ensembles. The old high diversityensemble starts learning with low diversity, in order tocheck if it is possible to converge to the new concept. Inaddition, a new low and a new high diversity ensemble arecreated from scratch.

The analysis identifies what ensembles are the mostaccurate for each type of drift and reveals that highdiversity ensembles trained on the old concept are able toconverge to the new concept if they start learning the newconcept with low diversity. In fact, these ensembles areusually the most accurate for most types of drift. The studyuses a technique presented in [9] (and explained here inSection 3.2) to explicitly encourage more or less diversity inan ensemble.

In order to answer the last question, we propose a newonline ensemble learning approach to handle concept driftscalled Diversity for Dealing with Drifts (DDD). The approachaims at better exploiting diversity to handle drifts, being morerobust to false alarms (false positive drift detections), andhaving faster recovery from drifts. In this way, it manages toachieve improved accuracy in the presence of drifts at thesame time as good accuracy in the absence of drifts ismaintained. Experiments with artificial and real-world datashow that DDD usually obtains similar or better accuracythan Early Drift Detection Method (EDDM) and betteraccuracy than Dynamic Weighted Majority (DWM).

This paper is further organized as follows: Section 3explains related work. Section 4 explains the data sets usedin the study. Section 5 presents the study of the effect of low/high diversity ensembles on the prequential accuracy in thepresence of drifts using the strategies presented in Table 1.Section 6 introduces and analyzes DDD. Section 7 concludesthe paper and gives directions for further research.

3 RELATED WORK

Several approaches proposed to handle concept drift can befound in the literature. Most of them are incrementallearning approaches [8], [14], [15], [16], [17], [18], [19], [20],[21]. However, most of the incremental learning approachestend to give little attention to the stability of the classifiers,giving more emphasis to the plasticity when they allowonly a new classifier to learn a new chunk of data. Whilethis could be desirable when drifts are very frequent, it isnot a good strategy when drifts are not so frequent. Besides,determining the chunk size in the presence of concept driftsis not a straightforward task. A too small chunk size will notprovide enough data for a new classifier to be accurate,whereas a too large chunk size may contain data belongingto different concepts, making the adaptation to newconcepts slow.

The online learning algorithms which handle conceptdrift can be divided into two groups: approaches which usea mechanism to detect drifts [7], [12], [22], [23], [24] andapproaches which do not explicitly detect drifts [25], [26],[27]. Both of them handle drifts based directly or indirectlyon the accuracy of the current classifiers. The formerapproaches use some measure related to the accuracy todetect drifts. They usually discard the current system andcreate a new one after a drift is detected and/or confirmed.In this way, they can have quick response to drifts, as long asthe drifts are detected early. However, these approaches cansuffer from nonaccurate drift detections. The latter ap-proaches usually associate weights to each ensemble’sclassifier based on its accuracy, possibly allowing pruning

620 IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, VOL. 24, NO. 4, APRIL 2012

TABLE 1Ensembles Analyzed to Address Research Questions

http://ieeexploreprojects.blogspot.com

Page 3: Ddd a new ensemble approach for dealing with concept drift.bak

and addition of new classifiers. These approaches need sometime for the weights to start reflecting the new concept.

Section 3.1 presents an example of drift detection methodand two approaches to handle drifts: an approach based ona drift detection method and an approach which does notdetect drifts explicitly. Section 3.2 briefly explains ensemblediversity in the context of online learning in the presence ofconcept drift.

3.1 Approaches to Detect and/or Handle Drifts

An example of drift detection method is the one used by theapproach presented in [12]. It is based on the idea that thedistance between two consecutive errors increases when astable concept is being learned. So, the distance ismonitored and, if it reduces considerably according to apredefined constant which we call � in this paper, it isconsidered that there is a concept drift.

The approach presented to handle concept drift in [12] iscalled Early Drift Detection Method. It uses the driftdetection method explained above with two values for �:� and �, � > �. When a concept drift is alarmed by the driftdetection method using �, but not �, a warning level istriggered, indicating that a concept drift might havehappened. From this moment, all the training examplespresented to the system are used for learning and thenstored. If a concept drift is alarmed using �, the conceptdrift is confirmed and the system is reset. If we considerthat a new online classifier system is created when thewarning level is triggered, instead of storing the traininginstances for posterior use, EDDM is considered a trueonline learning system.

An example of a well-cited approach which does not usea drift detection method is Dynamic Weighted Majority[26]. It maintains an ensemble of classifiers whose weightsare reduced by a multiplier constant �,1 � < 1, when theclassifier gives a wrong prediction, if the current time step isa multiple of p. The approach also allows the addition andremoval of classifiers at every p time step. The removal iscontrolled by a predefined weight threshold �. In this way,new classifiers can be created to learn new concepts andpoorly performing classifiers, which possibly learned oldconcepts, can be removed.

A summary of the parameters used by the drift detectionmethod, EDDM and DWM is given in Table 2.

3.2 Ensemble Diversity in the Presence of Drifts

Even though several studies of ensemble diversity can befound in the literature, e.g., [28], [29], [30], the studypresented in [9] is the first diversity analysis regardingonline learning in the presence of concept drift.

The study shows that different diversity levels arerequired before and after a drift in order to improvegeneralization on the old or new concepts and concludesthat diversity by itself can help to reduce the initial increasein error caused by a drift, but does not provide a fasterrecovery from drifts in long term. So, the potential ofensembles for dealing with concept drift may have not beenfully exploited by the existing ensemble approaches yet, asthey do not encourage different levels of diversity indifferent situations.

Intuitively speaking, the key to the success of anensemble of classifiers is that the base classifiers performdiversely. Despite the popularity of the term diversity, thereis no single definition or measure of it [30]. A popularmeasure is Yule’s Q statistic [31]. Based on an analysis of10 measures, Q statistic is recommended by Kuncheva andWhitaker [29] to be used for the purpose of minimizing theerror of ensembles. This measure is recommended especiallydue to its simplicity and ease of interpretation. Consideringtwo classifiers Di and Dk, the Q statistic can be calculated as

Qi;k ¼N11N00 �N01N10

N11N00 þN01N10;

where Na;b is the number of training examples for which theclassification given by Di is a and the classification given byDk is b, 1 represents a correct classification and 0 representsa misclassification.

Classifiers which tend to classify the same examplescorrectly will have positive values of Q, whereas classifierswhich tend to classify different examples incorrectly willhave negative values of Q. For an ensemble of classifiers,the averaged Q statistic over all pairs of classifiers can beused as a measure of diversity. Higher/lower averageindicates less/more diversity. In this paper, we willconsider that low/high diversity refers to high/low averageQ statistic.

In online learning, an example of how to explicitlyencourage more or less diversity in an ensemble is by usinga modified version [9] of online bagging [1]. The originalonline bagging (Algorithm 1) is based on the fact that, whenthe number of training examples tends to 1 in offlinebagging, each base learner hm contains K copies of each

MINKU AND YAO: DDD: A NEW ENSEMBLE APPROACH FOR DEALING WITH CONCEPT DRIFT 621

TABLE 2Parameters Description

1. The original greek letter to refer to this parameter was �. We changedit to � to avoid confusion with EDDM’s parameter �.

http://ieeexploreprojects.blogspot.com

Page 4: Ddd a new ensemble approach for dealing with concept drift.bak

original training example d, where the distribution of Ktends to a Poissonð1Þ distribution. So, in online bagging,whenever a training example is available, it is presented Ktimes for each base learner hm, where K is drawn from aPoissonð1Þ distribution. The classification is done byunweighted majority vote, as in offline bagging.

Algorithm 1. Online Bagging

Inputs: ensemble h; ensemble size M; training example d;

and online learning algorithm for the ensemble

members OnlineBaseLearningAlg;

1: for m 1 to M do

2: K Poissonð1Þ3: while K > 0 do

4: hm OnlineBaseLearningAlgðhm; dÞ5: K K � 1

6: end while

7: end for

Output: updated ensemble h

In order to encourage different levels of diversity,Algorithm 1 can be modified to include a parameter � forthe Poissonð�Þ distribution, instead of enforcing � ¼ 1. Inthis way, higher/lower � values are associated with higher/lower average Q statistic (lower/higher diversity), as shownin [9, Section 5.4.1]. This parameter is listed in Table 2.

4 DATA SETS

When working with real-world data sets, it is not possibleto know exactly when a drift starts to occur, which type ofdrift is present, or even if there really is a drift. So, it is notpossible to perform a detailed analysis of the behavior ofalgorithms in the presence of concept drift using only purereal-world data sets. In order to analyze the effect of low/high diversity ensembles in the presence of concept driftand to assist the analysis of DDD, we first used the artificialdata sets described in [9]. Then, in order to reaffirm theanalysis of DDD, we performed experiments using threereal-world problems. Section 4.1 describes the artificial datasets and Section 4.2 describes the real-world data sets.

4.1 Artificial Data Sets

The artificial data sets used in the experiments (Table 3)comprise the following problems [9]: circle, sine movingvertically, sine moving horizontally, line, plane, andBoolean. In the equations presented in the table, a, b, c, d,r, ai, eq, and op can assume different values to definedifferent concepts. The examples of all problems but Booleancontain x=xi, and y as the input attributes and the concept(which can assume value 0 or 1) as the output attribute.

The Boolean problem is inspired by the STAGGERproblem [32], but it allows the generation of different drifts,with different levels of severity and speed. In this problem,each training example has three input attributes: color (redR, green G, or blue B), shape (triangle T , rectangle R, orcircle C), and size (small S, medium M, or large L). Theconcept is then represented by the Boolean equation givenin Table 3, which indicates the color, shape, and size of theobjects which belong to class 1 (true). In that expression, arepresents a conjunction or disjunction of different possible

colors, b represents shapes, c represents sizes, eq represents¼ or 6¼ , and op represents the logical connective ^ or _. Forexample, the first concept of the Boolean data set whichpresents a drift with 11 percent of severity in Table 3 isrepresented by

ðcolor ¼ R ^ shape ¼ RÞ ^ size ¼ S _ M _ L:

Each data set contains one drift and different drifts weresimulated by varying among three amounts of severity (asshown in Table 3) and three speeds, thus generating ninedifferent drifts for each problem. Severity representsthe amount of changes caused by a new concept. Here,the measure of severity is the percentage of the input spacewhich has its target class changed after the drift iscomplete. Speed is the inverse of the time taken for anew concept to completely replace the old one. The speedwas measured by the inverse of the number of time stepstaken for a new concept to completely replace the old oneand was modeled by the following linear degree ofdominance functions:

vnðtÞ ¼t�N

drifting time; N < t � N þ drifting time

and

voðtÞ ¼ 1� vnðtÞ; N < t � N þ drifting time;

where vnðtÞ and voðtÞ are the degrees of dominance of thenew and old concepts, respectively; t is the current timestep; N is the number of time steps before the drift started tooccur; and drifting time varied among 1, 0:25N , and 0:50Ntime steps. During the drifting time, the degrees ofdominance represent the probability that an example ofthe old or the new concept will be presented to the system.For a detailed explanation of different types of drift, werecommend [9].

The data sets are composed of 2N examples and eachexample corresponds to one time step of the learning. The

622 IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, VOL. 24, NO. 4, APRIL 2012

TABLE 3Artificial Data Sets

http://ieeexploreprojects.blogspot.com

Page 5: Ddd a new ensemble approach for dealing with concept drift.bak

first N examples of the training sets were generatedaccording to the old concept (voðtÞ ¼ 1, 1 � t � N), whereN ¼ 1;000 for circle, sineV, sineH, and line and N ¼ 500 forplane and Boolean. The next drifting time training exam-ples (N < t � N þ drifting time) were generated accordingto the degree of dominance functions, vnðtÞ and voðtÞ. Theremaining examples were generated according to the newconcept (vnðtÞ ¼ 1, N þ drifting time < t � 2N).

The range of x or xi was ½0; 1� for circle, line, and plane;½0; 10� for sineV; and ½0; 4�� for sineH. The range of ywas ½0; 1�for circle and line, ½�10; 10� for sineV, ½0; 10� for sineH, and½0; 5� for plane. For plane and Boolean, the input attributesare normally distributed through the whole input space. Forthe other problems, the number of instances belonging toclass 1 and 0 is always the same, having the effect ofchanging the unconditional probability distribution functionwhen the drift occurs. Eight irrelevant attributes and10 percent class noise were introduced in the plane data sets.

4.2 Real-World Data Sets

The real-world data sets used in the experiments with DDDare electricity market [7], KDD Cup 1999 network intrusiondetection data [33], and PAKDD 2009 credit card data.

The electricity data set is a real-world data set from theAustralian New South Wales Electricity Market. This dataset was first described in [34]. In this electricity market, theprices are not fixed and may be affected by demand andsupply. Besides, during the time period described in thedata, the electricity market was expanded with the inclusionof adjacent areas, causing a dampening of the extremeprices. This data set contains 45,312 examples, dated fromMay 1996 to December 1998. Each example contains fourinput attributes (time stamp, day of the week, and twoelectricity demand values) and the target class, whichidentifies the change of the price related to a movingaverage of the last 24 hours.

The KDD Cup 1999 data set is a network intrusiondetection data set. The task of an intrusion detection systemis to distinguish between intrusions/attacks and normalconnections. The data set contains a wide variety ofintrusions simulated in a military network environment.During the simulation, the local area network was operatedas if it were a true Air Force environment, but pepperedwith multiple attacks, so that attack is not a minority class.The data set contains 494,020 examples. Each examplecorresponds to a connection and contains 41 inputattributes, such as the length of the connection, the typeof protocol, the network service on the destination, etc. Thetarget class identifies whether the connection is an attack ora normal connection.

The PAKDD 2009 data set comprises data collected fromthe private label credit card operation of a major Brazilianretail chain, along stable inflation condition. We used themodeling data, which contain 50,000 examples and corre-spond to a one year period. Each example corresponds to aclient and contains 27 input attributes, such as sex, age,marital status, profession, income, etc. The target classidentifies if the client is a “good” or “bad” client. The class“bad” is a minority class and composes around 19.5 percentof the data.

5 THE EFFECT OF LOW/HIGH DIVERSITY

ENSEMBLES ON THE PREQUENTIAL ACCURACY

This section presents an analysis of the prequential accuracy

of the ensembles/strategies presented in Table 1, aiming at

answering the first three research questions explained in

Section 2. The prequential [13] accuracy is the average

accuracy obtained by the prediction of each example to

be learned, before its learning, calculated in an online way.

The rule used to obtain the prequential accuracy on time

step t is presented in (1) [12]:

accðtÞ ¼accexðtÞ; if t ¼ f;accðt� 1Þ þ accexðtÞ � accðt� 1Þ

t� f þ 1; otherwise;

8<:

ð1Þ

where accex is 0 if the prediction of the current training

example ex before its learning is wrong and 1 if it is correct;

and f is the first time step used in the calculation.In order to analyze the behavior of the ensembles before

and after the beginning of a drift, the prequential accuracy

shown in the graphs is reset whenever a drift starts

(f 2 f1; N þ 1g). The learning of each ensemble is repeated

30 times for each data set.The online ensemble learning algorithm used in the

experiments is the modified version of online bagging

proposed in [9] and explained in Section 3.2. As commented

in that section, Minku et al. [9] show that higher/lower �s

produce ensembles with higher/lower average Q statistic

(lower/higher diversity).Section 5.1 presents the analysis itself.

5.1 Experimental Results and Analysis

The experiments used 25 lossless ITI online decision trees

[35] as the base learners for each ensemble. The parameter

�l for the Poissonð�Þ distribution of the low diversity

ensembles is 1, which is the value used for the original

online bagging [1]. The �h values for the high diversity

ensembles were chosen in the following way:

1. Perform five preliminary executions using �h ¼0:0005, 0.001, 0.005, 0.01, 0.05, 0.1, 0.5, giving a totalof 35 executions for each data set.

2. Determine the prequential accuracy obtained byeach execution at the time step 1:1N . This time steprepresents the moment in which high diversityensembles are more likely to outperform lowdiversity ensembles, according to [9].

3. Calculate the main effect of �h on the averageprequential accuracy, considering all the data setsfor each particular problem at the same time. Forexample, the main effect for the circle problem is theaverage among the online accuracies obtained by thefive executions using all the nine circle data sets.

4. For each problem, choose the �h which obtained thebest main effect. These values are 0.05 for circle,sineH, and plane, and 0.005 for line and sineV. ForBoolean, both �h ¼ 0:1 and �h ¼ 0:5 obtained thesame main effect. So, we chose �h ¼ 0:1, as itobtained the best main effect at 1:5N and 2N .

MINKU AND YAO: DDD: A NEW ENSEMBLE APPROACH FOR DEALING WITH CONCEPT DRIFT 623

http://ieeexploreprojects.blogspot.com

Page 6: Ddd a new ensemble approach for dealing with concept drift.bak

Fig. 1 shows the prequential accuracy obtained forthe circle problem for low and high severities and speeds.The figures for the other data sets were omitted due tospace limitations. In order to answer the first researchquestion, we analyze the prequential accuracy after thebeginning of the drift, for each data set. We can observe thatdifferent ensembles obtain the best prequential accuracydepending on the type of drift.

For drifts with low severity and high speed (e.g., Fig. 1a),the best accuracy after the beginning of the drift is usuallyobtained by the old high diversity ensemble. For theBoolean problem, the old low diversity gets similaraccuracy to the old high diversity ensemble. So, in general,it is a good strategy to use the old high diversity ensemblefor this type of drift.

An intuitive explanation for the reason why old highdiversity ensembles are helpful for low severity drifts isthat, even though the new concept is not the same as the oldconcept, it is similar to the old concept. When a high level ofdiversity is enforced, the base learners are forced to classifythe training examples very differently from each other. So,the ensemble learns a certain concept only partly, being ableto converge to the new concept by learning it with lowdiversity. At the same time, as the old concept was partlylearned, the old high diversity ensemble can use informa-tion learned from the old concept to aid the learning of thenew concept. An old low diversity ensemble would provideinformation from the old concept, but would have problemsto converge to the new concept [9].

For drifts with high severity and high speed (e.g., Fig. 1b),the new low diversity ensemble usually obtains the bestaccuracy, even though that accuracy is similar to the oldhigh diversity ensemble’s in half of the cases (Boolean,sineV, and line). For the Boolean problem, the old highdiversity, old low diversity, and new low diversity obtainsimilar accuracy. So, in general, it is a good strategy to usethe new low diversity ensemble for this type of drift.

The reason for that is that, when a drift has high severityand high speed, it causes big changes very suddenly. In thiscase, the new concept has almost no similarities to the oldconcept. So, an ensemble which learned the old concepteither partly or fully will not be so helpful (and could beeven harmful) for the accuracy on the new concept. A newensemble learning from scratch is thus the best option.

For drifts with medium severity and high speed, thebehavior of the ensembles is similar to when the severity ishigh for sineH, circle, plane, and line, although the difference

between the old high diversity and the new low diversityensemble tends to be smaller for sineH, circle, and plane. Thebehavior for Boolean and sineV tends to be more similar towhen severity is low. So, drifts with medium severitysometimes have similar behavior to low severity andsometimes to high severity drifts when the speed is high.

For drifts with low speed (e.g., Figs. 1c and 1d), either theold low or both the old ensembles present the best accuracyin the beginning of the drift, independent of the severity. So,considering only shortly after the beginning of a drift, thebest strategy is to use the old low diversity ensemble for slowspeed drifts. Longer after the drift, either the old high or boththe old high and the new low diversity ensembles usuallyobtain the best accuracies. So, considering only longer afterthe drift, the best strategy is to use the old high diversityensemble. If we consider both shortly and longer after thedrift, it is a good strategy to use the old high diversityensemble, as it is the most likely to have good accuracyduring the whole period after the beginning of the drift.

For drifts with medium speed, the behavior is similar tolow speed, although the period of time in which the oldensembles have the best accuracies is reduced and the oldlow diversity ensemble rarely has the best accuracy by itselfshortly after the beginning of the drift (it usually obtainssimilar accuracy to the old high diversity ensemble). Whenthe severity is high, there are two cases (sineH and plane) inwhich the best accuracy is obtained by the new lowdiversity ensemble longer after the drift. This behaviorapproximates the behavior obtained when the severity ishigh and the speed is high.

This analysis is a straight answer to the first researchquestion and also allows us to answer the second and thirdquestions as follows: 2) Ensembles which learned an oldconcept using high diversity can converge to a new conceptif they start learning the new concept with low diversity;and 3) when the drift has low severity and high speed orlonger after the drift when the speed is low or medium, thehigh diversity ensembles learning with low diversity are themost accurate in most of the cases. Besides, when the speedis low or medium, shortly after the beginning of the drift,they are more accurate than the new ensembles andfrequently have similar accuracy to the old low diversityensembles. Even when the drift has medium or highseverity and high speed, the old high diversity ensemblessometimes obtain similar accuracy to the new low diversityensembles. So, in fact, it is a good strategy to use the oldhigh diversity ensembles for most types of drift.

624 IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, VOL. 24, NO. 4, APRIL 2012

Fig. 1. Average prequential accuracy (1) of the four ensembles analyzed for the circle problem considering 30 runs using “perfect” drift detections.The accuracy is reset when the drift starts (f 2 f1; 1001g). The new ensembles are created from scratch at the time steps 1 and 1,001. The oldensembles correspond to the new ensembles before the beginning of the drift. (a) Low Sev, High Sp. (b) High Sev, High Sp. (c) Low Sev, Low Sp.(d) High Sev, Low Sp.

http://ieeexploreprojects.blogspot.com

Page 7: Ddd a new ensemble approach for dealing with concept drift.bak

The analysis shows that the strategy of resetting thelearning system as soon as a drift is detected, which isadopted by many approaches, such as [7], [12], [23], is notalways ideal, as an ensemble which learned the old conceptcan be helpful depending on the drift type.

6 DIVERSITY FOR DEALING WITH DRIFTS

This section proposes Diversity for Dealing with Drifts.2

Section 6.1 describes DDD. Sections 6.2 and 6.3 explain theexperiments done to validate DDD and provide an answerto the last research question from Section 2.

6.1 DDD’s Description

DDD (Algorithm 2) operates in two modes: prior to driftdetection and after drift detection. We chose to use a driftdetection method, instead of treating drifts implicitly,because it allows immediate treatment of drifts once theyare detected. So, if the parameters of the drift detectionmethod are tuned to detect drifts the earliest possible andthe approach is designed to be robust to false alarms, wecan obtain fast adaptation to new concepts. The parametersused by DDD are summarized in Table 2.

Algorithm 2. DDD

Inputs:

. multiplier constant W for the weight of the old low

diversity ensemble;. online ensemble learning algorithm

EnsembleLearning;

. parameters for ensemble learning with low

diversity pl and high diversity ph;

. drift detection method DetectDrift;

. parameters for drift detection method pd;

. data stream D;

1: mode before drift2: hnl new ensemble =� new low diversity �=

3: hnh new ensemble =� new high diversity �=

4: hol hoh null =� old low and high diversity �=

5: accol accoh accnl accnh 0 =� accuracies �=

6: stdol stdoh stdnl stdnh 0 =� standard

deviations �=

7: while true do

8: d next example from D

9: if mode == before drift then

10: prediction hnlðdÞ11: else

12: sumacc accnl þ accol �W þ accoh13: wnl ¼ accnl=sumacc

14: wol ¼ accol �W=sumacc

15: woh ¼ accoh=sumacc

16: prediction WeightedMajority(hnlðdÞ, holðdÞ,hohðdÞ, wnl, wol, woh)

17: Update(accnl, stdnl, hnl, d)

18: Update(accol, stdol, hol, d)

19: Update(accoh, stdoh, hoh, d)

20: end if

21: drift DetectDrift(hnl, d, pd)

22: if drift == true then

23: if mode == before drift OR(mode == after drift AND accnl > accoh) then

24: hol hnl25: else

26: hol hoh27: end if

28: hoh hnh29: hnl new ensemble

30: hnh new ensemble31: accol accoh accnl accnh 0

32: stdol stdoh stdnl stdnh 0

33: mode after drift

34: end if

35: if mode == after drift then

36: if accnl > accoh AND accnl > accol then

37: mode before drift

38: else

39: if accoh � stdoh > accnl þ stdnl AND accoh �stdoh > accol þ stdol then

40: hnl hoh41: accnl accoh42: mode before drift

43: end if

44: end if

45: end if

46: EnsembleLearning(hnl; d; pl)

47: EnsembleLearning(hnh; d; ph)

48: if mode == after drift then

49: EnsembleLearning(hol; d; pl)

50: EnsembleLearning(hoh; d; pl)

51: end if

52: if mode == before drift then

53: Output hnl, prediction54: else

55: Output hnl; hol; hoh; wnl; wol; woh, prediction

56: end if

57: end while

Before a drift is detected, the learning system iscomposed of two ensembles: an ensemble with lowerdiversity (hnl) and an ensemble with higher diversity(hnh). Both ensembles are trained with incoming examples(lines 46 and 47), but only the low diversity ensemble isused for system predictions (line 10). The reason for notusing the high diversity ensemble for predictions is that it islikely to be less accurate on the new concept being learnedthan the low diversity ensemble [9]. DDD assumes that, ifthere is no convergence of the underlying distributions to astable concept, new drift detections will occur, triggeringthe mode after drift detection. DDD then allows the use ofthe high diversity ensemble in the form of an old highdiversity ensemble, as explained later in this section.

A drift detection method based on monitoring the lowdiversity ensemble is used (line 21). This method can be anyof the methods existing in the literature, for instance, theone explained in Section 3.1. After a drift is detected, newlow diversity and high diversity ensembles are created(lines 29 and 30). The ensembles corresponding to the lowand high diversity ensembles before the drift detection arekept and denominated old low and old high diversity

MINKU AND YAO: DDD: A NEW ENSEMBLE APPROACH FOR DEALING WITH CONCEPT DRIFT 625

2. An initial version of DDD can be found in [36].

http://ieeexploreprojects.blogspot.com

Page 8: Ddd a new ensemble approach for dealing with concept drift.bak

ensembles (lines 24 and 28). The old high diversityensemble starts to learn with low diversity (line 50) inorder to improve its convergence to the new concept, asexplained in Section 5. Maintaining the old ensemblesallows not only a better exploitation of diversity and the useof information learned from the old concept to aid thelearning of the new concept, but also helps the approach tobe robust to false alarms.

Both the old and the new ensembles perform learning(lines 46-50) and the system predictions are determined bythe weighted majority vote of the output of 1) the old highdiversity, 2) the new low diversity, and 3) the old lowdiversity ensemble (lines 12-16). The new high diversityensemble is not considered because it is likely to have lowaccuracy on the new concept [9].

The weights are proportional to the prequential accuracysince the last drift detection until the previous time step(lines 13-15). The weight of the old low diversity ensemble ismultiplied by a constant W (line 15), which allows control-ling the trade-off between robustness to false alarms andaccuracy in the presence of concept drifts, and all the weightsare normalized. It is important to note that weighting theensembles based on the accuracy after the drift detection isdifferent from the weighting strategy adopted by theapproaches in the literature which do not use a driftdetection method. Those approaches use weights which arelikely to represent more than one concept at the same timeand need some time to start reflecting only the new concept.

During the mode after drift detection, the new lowdiversity ensemble is monitored by the drift detectionmethod (line 21). If two consecutive drift detections happenand there is no shift back to the mode prior to drift detectionbetween them, the old low diversity ensemble after thesecond drift detection can be either the same as the old highdiversity learning with low diversity after the first driftdetection or the ensemble corresponding to the new lowdiversity after the first drift detection, depending on whichof them is the most accurate (lines 24 and 26). The reason forthat is that, soon after the first drift detection, the new lowdiversity ensemble may be not accurate enough to becomethe old low diversity ensemble. This strategy also helps theapproach to be more robust to false alarms.

All the four ensembles are maintained in the system untileither the condition in line 36 or the condition in line 39 issatisfied. When one of these conditions is satisfied, thesystem returns to the mode prior to drift detection. Theaccuracies when considering whether the old high diversityensemble is better than the others are reduced/summed totheir standard deviations to avoid premature return to themode prior to drift, as this ensemble is more likely to havehigher accuracy than the new low diversity very soon afterthe drift, when this ensemble learned just a few examples.

When returning to the mode prior to drift, either the oldhigh diversity or the new low diversity ensemble becomesthe low diversity ensemble used in the mode prior to driftdetection, depending on which of them is the most accurate(lines 36-44). An additional parameter to control themaximum number of time steps in the mode after driftdetection can be used to avoid a too long time maintainingfour ensembles and is proposed as future work.

As we can see, DDD is designed to better exploit theadvantages of diversity to deal with concept drift than theother approaches in the literature, by maintaining ensem-bles with different diversity levels, according to theexperiments presented in Section 5.

Besides, the approaches which use old classifiers as partof an ensemble in the literature, such as [25], [26], [27], donot adopt any strategy to improve their learning on the newconcept. Nevertheless, as it was shown in [9], theseclassifiers are likely to have low accuracy on the newconcept if no additional strategy is adopted to improve theirlearning. DDD is designed to use information learned fromthe old concept in order to aid the learning of the newconcept, by training an old high diversity ensemble withlow diversity on the new concept. Such a strategy hasshown to be successful in Section 5.

Moreover, the approach is designed to be more robust tofalse alarms than approaches which reset the system when adrift is detected [7], [12], [23] and to have faster recoveryfrom drifts than approaches which do not use a driftdetection method [26], [27], as it maintains old ensemblesafter a drift detection, but takes immediate action to treatthe drift when it is detected, instead of having to wait forthe weights to start reflecting the new concept.

The approach is not yet prepared to take advantage ofrecurrent or predictable drifts. We propose the use ofmemories for dealing with these types of drift as future work.

A detailed analysis of time and memory occupied byDDD is not straightforward, as it depends on theimplementation, base learner, and source of diversity.However, it is easy to see that, if we have a sequentialimplementation, the complexity of each ensemble is OðfðnÞÞand the source of diversity does not influence this complex-ity, DDD would have, in the worst case, complexityOð4fðnÞÞ ¼ OðfðnÞÞ. So, DDD does not increase the com-plexity of the system in comparison to a single ensemble.

6.2 Experimental Objectives, Design, and MeasuresAnalyzed

The objective of the experiments with DDD is to assist itsanalysis and to validate it, showing that it is an answer tothe last research question presented in Section 2, i.e., itobtains good accuracy both in the presence and absence ofdrifts. We also aim at identifying for which types of driftDDD works better and why it behaves in that way.

In order to do so, we analyze measures such as weightsattributed by DDD to each ensemble, number of driftdetections, and prequential accuracy ((1), from Section 5). Insome cases, the false positive and negative rate is alsoanalyzed. DDD is compared to a low diversity ensemblewith no drift handling abilities, EDDM [12], and DWM [26].The diversity study using “perfect” drift detections pre-sented in Section 5.1 and an approach which would alwayschoose to use the same ensemble (i.e., always choose the oldhigh diversity ensemble, or always choose the old lowdiversity ensemble, etc.) are also used in the analysis.

The prequential accuracy is calculated based on thepredictions given to the current training example before theexample is used for updating any component of the system.It is important to observe that the term update here refers notonly to the learning of the current training example by thebase learners, but also to the changes in weights associated

626 IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, VOL. 24, NO. 4, APRIL 2012

http://ieeexploreprojects.blogspot.com

Page 9: Ddd a new ensemble approach for dealing with concept drift.bak

with the base learners (in the case of DWM) and to theensembles (in the case of DDD). The prequential accuracy iscompared both visually, considering the graphs of theaverage prequential accuracy and standard deviationthroughout the learning, and using T student statisticaltests [37] at particular time steps, for the artificial data sets.

The T tests were performed as follows: for each artificialproblem, T student statistical tests were done at the timesteps 0:99N , 1:1N , 1:5N , and 2N . These time steps werechosen in order to analyze the behavior soon after the drift,fairly longer after the drift and long after the drift, as in [9].Bonferroni corrections considering all the combinations ofseverity, speed, and time step were used. The overallsignificance level was 0.01, so that the significance levelused for each individual comparison was � ¼ 0:01=ð3 �3 � 4Þ ¼ 0:000278.

The drift detection method used by DDD was the sameas the one used by EDDM, in order to provide a faircomparison. There is the possibility that there are other driftdetection methods more accurate in the literature. However,the study presented in Section 5 shows that the oldensembles are particularly useful right after the beginningof the drift. So, a comparison to an approach using a driftdetection method which could detect drifts earlier wouldgive more advantages to DDD.

Thirty repetitions were done for each data set andapproach. The ensemble learning algorithm used by DDDand EDDM was the (modified) online bagging, as inSection 5. The drift detection method used by DDD in theexperiments was the one explained in Section 3.1.

The analysis of DDD and its conclusions are based onthe following:

1. EDDM always uses new learners created fromscratch. Nevertheless, resetting the system upon driftdetections is not always the best choice. DWM allowsus to use old classifiers, but does not use any strategyto help the convergence of these classifiers to the newconcept. So, it cannot use information from the oldconcept to learn the new concept faster, as DDD does.At least in the situations in which new learners createdfrom scratch or old learners which attempted to learnthe old concept well are not the best choice, DDD willperform better if it manages to identify these situa-tions and adjust the weights of the ensemblesproperly. This is independent of the base learner.

2. The diversity study presented in Section 5 showsthat such situations exist and that an old highdiversity ensemble, which is used by DDD, is benefic

for several different types of drift. In Section 6.3.1,we show using artificial data and decision trees thatDDD is able to identify such situations and adjustthe ensemble weights properly, being accurate in thepresence and absence of drifts. We also identify thedrifts for which DDD works better and explain why.Section 6.3.2 analyzes the number of time steps inwhich DDD maintains four ensembles. Section 6.3.3shows that DDD is robust to false alarms andexplains the influence of the parameter W . Addi-tional experiments using naive bayes (NB) andmultilayer perceptrons (MLPs) on real-world pro-blems (Section 6.3.4) further confirm the analysisdone with the artificial data.

6.3 Experimental Results and Analysis

6.3.1 Experiments with Artificial Data

The sets of parameters considered for the experiments are

shown in Table 4. The parameters �h were chosen to be the

same as in Section 5.1. The parameter � (� for EDDM) was

chosen so as to provide early drift detections. This

parameter was the same for DDD and EDDM, providing

a fair comparison.Preliminary experiments with five runs for each data set

show that � does not influence much EDDM’s accuracy.Even though an increase in � is associated with increases inthe total number of time steps in warning level, the system isusually in warning state for very few consecutive time stepsbefore the drift level is triggered, even when very large � isadopted. That number is not enough to significantly affectthe accuracy. So, we decided to use � ¼ 0:99.

Similarly to �h, the parameter p of DWM was chosen

considering the best average accuracy at the time step 1:1N ,

using the default values of � ¼ 0:5 and � ¼ 0:01. The average

was calculated using five runs for all the data sets of a

particular problem at the same time, as for the choice of �h.

After selecting p, a fine tuning was done for DWM, again

based on five preliminary runs, by selecting the parameter �

which provides the best main effect on the prequential

accuracy at the time step 1.1N when using the best p. The

execution time using � ¼ 0:7 and 0.9 became extremely high,

especially for Plane and Circle. The reason for that is that a

combination of a low pwith a high � causes DWM to include

new base learners very frequently, whereas the weights

associated with each base learner reduce slowly. So, we did

not consider these values for these problems.The base learners used in the experiments were lossless

ITI online decision trees [35] and both DDD and EDDM

MINKU AND YAO: DDD: A NEW ENSEMBLE APPROACH FOR DEALING WITH CONCEPT DRIFT 627

TABLE 4Parameters Choice for Artificial Data where W ¼ 1, �l ¼ 1, and � ¼ 0:01 Were Fixed

http://ieeexploreprojects.blogspot.com

Page 10: Ddd a new ensemble approach for dealing with concept drift.bak

used ensemble size of 25 ITIs. DWM automatically selectsthe ensemble size.

Fig. 2 shows the prequential accuracy and Fig. 3 showsthe weights attributed by DDD for some representativedata sets. Graphs for other data sets were omitted due tospace restrictions.

We first compare DDD’s prequential accuracy toEDDM’s. During the first concept, DDD is equivalent toEDDM if there are no false alarms. Otherwise, DDD hasbetter accuracy than EDDM. This behavior is expected, asEDDM resets the system when there is a false alarm, having

to relearn the current concept. DDD, on the other hand, can

make use of the old ensembles by increasing their weights.

Fig. 3 shows that indeed DDD increases the weights of the

old ensembles when there are false alarms (the average

number of drift detections is shown in brackets in Fig. 2).We concentrate now on comparing DDD to EDDM in

terms of accuracy after the average time step of the first drift

detection done during the second concept. The experiments

show that DDD presents better accuracy than EDDM

mainly for the drifts known from the diversity study

628 IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, VOL. 24, NO. 4, APRIL 2012

Fig. 3. Average weight attributed by DDD to each ensemble, considering 30 runs. (a) SineH Low Sev High Sp. (b) Plane Low Sev High Sp. (c) SineHHigh Sev High Sp. (d) Plane High Sev High Sp. (e) SineH Low Sev Low Sp. (f) Circle Low Sev Low Sp. (g) Circle High Sev Low Sp. (h) Plane HighSev Low Sp.

Fig. 2. Average prequential accuracy (1) of DDD, EDDM, DWM and an ensemble without drift handling, considering 30 runs. The accuracy is resetwhen the drift begins (f 2 f1; N þ 1g). The vertical black bars represent the average time step in which a drift is detected at each concept. Thenumbers in brackets are the average numbers of drift detections per concept. The results of the comparisons aided by T tests at the time steps0:99N, 1:1N, 1:5N, and 2N are also shown: “> ” means that DDD attained better accuracy than EDDM, “< ” means worse, and “¼ ” means similar.(a) SineH Low Sev High Sp ð1:13; 2:23Þ >¼�. (b) Plane Low Sev High Sp ð0:17; 1:47Þ ¼>¼>. (c) SineH High Sev High Sp ð1:13; 2:17Þ >¼�.(d) Plane High Sev High Sp ð0:10; 1:03Þ ¼�<. (e) SineH Low Sev Low Sp ð1:13; 2:43Þ >¼�. (f) Circle Low Sev Low Sp ðnone; 1:26Þ ¼¼� . (g) CircleHigh Sev Low Sp ðnone; 3:13Þ ¼�>. (h) Plane High Sev Low Sp ð0:10; 2:09Þ ¼¼¼¼.

http://ieeexploreprojects.blogspot.com

Page 11: Ddd a new ensemble approach for dealing with concept drift.bak

(Section 5) to benefit from the old ensembles (drifts withlow severity or low speed). Figs. 2a, 2b, 2e, 2f, and 2g showexamples of this behavior.

In the case of low severity and high speed drifts, the bestensemble to be used according to the study presented inSection 5 is the old high diversity, especially during the verybeginning of the learning of the new concept, when the newlow diversity is still inaccurate to the new concept. DDDgives considerable weight to the old high diversity ensem-ble, as shown in Figs. 3a and 3b. Even though it is not alwaysthe highest weight, it allows the approach to get betteraccuracy than EDDM. When there are false alarms, there aresuccessful sudden increases on the use of the old lowdiversity ensemble, as can be observed in the same figures.

However, the nonperfect (late) drift detections sometimesmake the use of the old high diversity ensemble lessbeneficial, making DDD get similar accuracy to EDDM,instead of better accuracy. Let’s observe, for example, Fig. 4a.Even though the study presented in Section 5 shows thatwhen there are perfect drift detections, the best ensemble forthis type of drift would be the old high diversity, theaccuracy of an approach which always chooses thisensemble becomes similar to the new low diversity ensem-ble’s during some moments when the drift detection methodis used. DDD obtains similar accuracy to EDDM exactlyduring these moments (Fig. 2b) and just becomes betteragain because of the false alarms.

In the case of low speed drifts, the best ensembles to beused according to the study presented in Section 5 are theold ensembles (especially the old low diversity) soon afterthe beginning of the drift. DDD manages to attain betteraccuracy than EDDM (e.g., Figs. 2e, 2f, and 2g) because itsuccessfully gives high weight for these ensembles rightafter the drift detections and keeps the weight of the old lowdiversity ensemble high when there are false alarms, asshown in Figs. 3e, 3f, and 3g.

The nonperfect (especially the late) drift detections alsoreduce the usefulness of the old ensembles for the lowspeed drifts, sometimes making DDD attain similaraccuracy to EDDM. An example of this situation is shownby Figs. 2h, 3h, and 4b. As we can see in Fig. 4b, theaccuracy of an approach which always chooses the old high

diversity is similar to an approach which always choosesthe new low diversity because of the nonperfect driftdetections. For only one problem, when the speed was lowand severity high, the late drift detections made DDD attainworse accuracy than EDDM.

When the drifts present high severity and high speed, theaccuracy of DDD was expected to be similar (not better) toEDDM, as the new low diversity is usually the most accurateand EDDM’s strategy is equivalent to always choosing thisensemble. However, the experiments show that DDD some-times presents similar, sometimes worse (Fig. 2d), andsometimes better (Fig. 2c) accuracy than EDDM.

The reason for the worse accuracy is the inaccuracy ofthe initial weights given to the new low diversity soon afterthe drift detection, as this ensemble learned too fewexamples. If the initial weights take some time to becomemore accurate, as shown in Fig. 3d, DDD needs some timefor its prequential accuracy to eventually recover andbecome similar to EDDM’s, as shown by the accuracy’sincreasing tendency in Fig. 2d. If the accuracy of the oldensembles decreases very fast in relation to the time takenby the new low diversity ensemble to improve its accuracy,DDD manages to attain similar accuracy to EDDM sincesoon after the drift detection. Besides, DDD can attain betteraccuracy than EDDM even for this type of drift due to thefalse alarms (Figs. 2c and 3c).

The accuracy of DDD for medium severity or speeddrifts was never worse than EDDM’s and is explainedsimilarly to the other drifts.

We shall now analyze the number of win/draw/loss ofDDD in comparison to EDDM at the time steps analyzedthrough T tests after the average time step of the first driftdetection during the second concept. It is important to notethat, when there are false alarms, DDD can get betteraccuracy than EDDM before this time step. Considering thetotal number of win/draw/loss independent of the drifttype, DDD obtains better accuracy in 45 percent of the cases,similar in 48 percent of the cases, and worse in only 7 percentof the cases. So, it is a good strategy to use DDD when thedrift type is unknown, as it obtains either better or similaraccuracy and only rarely obtains worse accuracy.

Considering the totals per severity, DDD has more wins(67 percent) in comparison to draws (33 percent) or losses

MINKU AND YAO: DDD: A NEW ENSEMBLE APPROACH FOR DEALING WITH CONCEPT DRIFT 629

Fig. 4. Average prequential accuracy (1) obtained by an approach which always chooses the same ensemble for prediction, considering 30 runs. Theaccuracy is reset when the drift begins (f 2 f1; 501g). The vertical black bars represent the average time step in which a drift was detected at eachconcept. The numbers in brackets are the average numbers of drift detections per concept. (a) Plane Low Sev High Sp (0.17, 1.47). (b) Plane HighSev Low Sp (0.10, 2.90).

http://ieeexploreprojects.blogspot.com

Page 12: Ddd a new ensemble approach for dealing with concept drift.bak

(0 percent) when severity is low. When severity is medium,DDD is similar in most of the cases (68 percent), beingsometimes better (32 percent) and never worse (0 percent).When severity is high, DDD is usually similar (42 percent)or better (38 percent) than EDDM, but in some cases, it isworse (20 percent). If we consider the totals per speed, theapproach has more wins (61 percent) in comparison todraws (34 percent) or losses (5 percent) when speed is low.When the speed is medium, the number of draws is higher(64 percent, against 36 percent for wins and 0 percent forlosses). When speed is high, the number of draws andwins is more similar (47 and 39 percent, respectively), butthere are some more losses (14 percent) than for the otherspeeds. This behavior is understandable, as, according toSection 5, the old ensembles are more helpful when theseverity or the speed is low.

DDD has usually higher accuracy than DWM, both in thepresence and absence of drifts (e.g., Fig. 2a to Fig. 2h).Before the drift, DDD is almost always better than DWM.Considering the total number of win/draw/loss indepen-dent of the drift type for the time steps 1:1N , 1:5N , and 2N ,DDD obtains better accuracy than DWM in 59 percent of thecases, similar in 25 percent of the cases, and worse in15 percent of the cases. As we can see in the figures, DDDusually presents faster recovery from drifts.

A probable reason for the lower accuracy of DWMduring stable concepts is that it adds new classifiers when itgives misclassifications, independent of how accurate theensemble is to the current concept. The new classifiers areinitially inaccurate and, even though the old classifierscompensate somewhat their misclassifications, the accuracyof the ensemble as a whole is reduced. A probable reasonfor the lower accuracy of DWM in the presence of drifts isthat its weights take some time steps to start reflecting thenew concept, causing slow recovery from drifts. So, DWMis usually better than an ensemble without drift handling,but worse than DDD. In a few cases, when drifts that do notaffect much the accuracy of old ensembles are detected,DWM obtained better accuracy than DDD. A detailedanalysis of DWM is outside the scope of this paper.

In a very few occasions, not only DDD, but also EDDM andDWM get worse accuracy than an ensemble without drifthandling during a few moments soon after the beginning ofthe drift when the drift is not fast (e.g., Figs. 2e and 2h). Thathappens because, in the beginning of the drift, ensembleswhich learned the old concept are expected to be among thehighest accuracies while the old concept is still dominant overthe new concept. Nevertheless, as the number of time stepsincreases and the old concept becomes less dominant, theaccuracy of an ensemble without drift handling is highlyaffected and reduced.

In summary, the experiments in this section show thatDDD gets usually similar or better accuracy than EDDMand usually better accuracy than DWM both in the presenceand absence of drifts. DDD also usually gets better accuracythan an ensemble without drift handling in the presence ofdrifts and similar accuracy in the absence of drifts.

6.3.2 Time Steps Maintaining Four Ensembles

In this section, we compare the time and memory occupiedby DDD to EDDM and DWM indirectly, by considering the

number of ensembles maintained in a sequential imple-mentation using � as the source of diversity and decisiontrees as the base learners. The high diversity ensembleshave faster training and occupy less memory, as they aretrained with much less examples (on average, � times thetotal number of examples). So, we will compare the numberof time steps in which DDD requires four ensembles to thenumber of time steps in which EDDM requires twoensembles to be maintained.

The experiments presented in Section 6.3.1 show thatDDD required the maintenance of four ensembles during,on average, 4.11 times more time steps than EDDMrequired two ensembles. Considering the total number oftime steps of the learning, DDD is likely to use, on average,1.22 times more time and memory than EDDM. DWMalways maintains one ensemble with variable size and thissize was, on average, 0.45 times the size of the ensemblesmaintained by DDD and EDDM. However, DWM is likelyto create/delete ensembles with a high rate when theaccuracy is not very high, increasing its execution time.

6.3.3 Robustness to False Alarms and the Impact of W

We performed additional experiments by forcing falsealarms on particular time steps during the first concept ofthe artificial data sets corresponding to low severity drifts,instead of using a drift detection method.

When varying W , the experiments show that thisparameter allows tuning the trade-off between robustnessto false alarms and accuracy in the presence of real drifts.The graphs are omitted due to space limitations.

Higher W (W ¼ 3) makes DDD more robust to falsealarms, achieving very similar accuracy to an approach withno drift handling, which is considered the best one in thiscase. W ¼ 3 makes DDD less accurate in the presence of realdrifts, but still more accurate than an ensemble withoutdrift handling in the presence of drifts.

Lower W (W ¼ 1) makes DDD less robust to falsealarms, but still with considerably good robustness andmore robust than EDDM, besides being more accurate inthe presence of real drifts. So, unless we are expecting tohave many false alarms and few real drifts, it is a goodstrategy to use W ¼ 1.

6.3.4 Experiments with Real-World Data

The experiments using real-world data were repeated usingtwo different types of base learners: MLPs and NB. TheMLPs contained 10 hidden nodes each and were trainedusing backpropagation with 1 epoch (online backpropaga-tion [38], [39]), learning rate 0.01, and momentum 0.01.These base learners were chosen because they are fasterthan ITIs when the data set is very large. Both DDD andEDDM used ensemble size of 100.

The parameters used in the experiments are shown inTable 5. All the parameters were chosen so as to generatethe most accurate accuracy curves, based on five runs. Thefirst parameters chosen were � and p, which usually havebigger influence in the accuracy. The five runs to choosethem used the default values of �h ¼ 0:005 and � ¼ 0:5.After that, a fine tuning was done by selecting theparameters �h and �. For electricity, preliminary experi-ments show that the drift detection method does notprovide enough detections. So, instead of using the drift

630 IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, VOL. 24, NO. 4, APRIL 2012

http://ieeexploreprojects.blogspot.com

Page 13: Ddd a new ensemble approach for dealing with concept drift.bak

detection method, we forced drift detections at every FA ¼f5; 25; 45g time steps. The only exception was EDDM usingNB. In this case, � ¼ 1:15 provided better results.

Each real-world data set used in the experiments hasdifferent features. So, in this section, we analyze thebehavior of DDD according to each real-world data setseparately, in combination with a more detailed analysis ofthe features of the data. For each data set, the priorprobability of class 1 at the time step t estimated accordingto the sample, P ð1ÞðtÞ, is given by

P ð1ÞðtÞ ¼Ptþwsize�1

i¼t yðiÞwsize

;

where yðiÞ is the target class (1 or 0) of the training examplepresented at the time step i, and wsize is the size of thewindow of examples considered for calculating the average.

The first data set analyzed is electricity. In this data set,the prior probability of an increase in price calculatedconsidering the previous 1,440 examples (one month ofobservations) varies smoothly during the whole learningperiod. These variations possibly represent several contin-uous drifts, to which DDD is expected to have a goodbehavior. Fig. 5a shows the accuracy obtained for this dataset using MLPs. As we can see, DDD is able to outperformDWM and EDDM in terms of accuracy. DDD was able toattain even higher accuracy in comparison to DWM andEDDM when using NB. The graph was omitted due tospace limitations.

The second data set analyzed is PAKDD. In this dataset, the probability of a fraudulent customer consideringthe previous 2,000 examples has almost no variationduring the whole learning period. So, this data set islikely to contain no drifts. In this case, DDD is expected toobtain at least similar accuracy to EDDM and DWM. Ifthere are false alarms, DDD is expected to outperformEDDM. The experiments show that all the approaches

manage to attain similar accuracy for this problem whenusing MLPs (Fig. 5b). A similar behavior happens whenusing NB. In particular, the drift detection methodperformed almost no drift detections—average of 0.37 driftdetections during the whole learning when using MLPsand of 3.60 when using NB. So, EDDM did not haveproblems with false alarms. Experiments using a para-meters setup which causes more false alarms show thatDDD is able to maintain the same accuracy as DWM inthat condition, whereas EDDM has reduced accuracy.

Nevertheless, the class representing a fraudulent custo-mer (class 1) is a minority class. So, it is important toobserve the rates of false positives fpr and negatives fnr,which are calculated as follows:

fprðtÞ ¼ numfpðtÞ=numnðtÞ and

fnrðtÞ ¼ numfnðtÞ=numpðtÞ;

where numfpðtÞ and numfnðtÞ are the total number of false

positives and negatives until the time step t; and numnðtÞand numpðtÞ are the total number of examples with true

class zero and one until the time step t.In PAKDD, it is important to obtain a low false negative

rate, in order to avoid fraud. When attempting to maximize

accuracy, the false positive rate becomes very low, but

the false negative rate becomes very high for all the

approaches analyzed. So, the parameters which obtain the

best false negative rate are different from the parameters

which obtain the best accuracy.Besides, DDD can be easily adapted for dealing with

minority classes by getting inspiration from the skewed(imbalanced) data sets literature [40]. Increasing anddecreasing diversity based on the parameter � of thePoisson distribution is directly related to sampling techni-ques. A � < 1 can cause similar effect to undersampling,whereas a � > 1 can cause similar effect to oversampling.

MINKU AND YAO: DDD: A NEW ENSEMBLE APPROACH FOR DEALING WITH CONCEPT DRIFT 631

TABLE 5Parameters Choice for Real-World Data where W ¼ 1, �l ¼ 1, and � ¼ 0:01 Were Fixed

Fig. 5. Average prequential accuracy (1) reset at every third of the learning, using MLPs/NB as base learners. (a) Electricity using MLPs. (b) PAKDDusing MLPs. (c) KDD using MLPs. (d) KDD using NB.

http://ieeexploreprojects.blogspot.com

Page 14: Ddd a new ensemble approach for dealing with concept drift.bak

So, experiments were done using �l ¼ �h ¼ 2 for theminority class,�h ¼ 0:005 for the majority class, � ¼ � ¼ 1:15,� ¼ 1:20, � ¼ 0:3, p ¼ 1 both when using MLPs and NB; �l ¼0:4 for the majority class when using MLPs; and �l ¼ 0:1 forthe majority class when using NB. As we can see in Fig. 6,DDD obtains the best false negative rate when using MLPs. Italso obtains the best rate when using NB. Additional workwith minority classes is proposed as future work.

The last data set analyzed is KDD. In this problem, theprobability of an intrusion considering the previous2,000 examples has several jumps from 1 to 0 and viceversa during the first and last third of the learning. So, thereare probably several severe and fast drifts which reoccurwith a high rate during these periods. Even though DDD(and EDDM) is prepared for dealing with severe and fastdrifts, it is not prepared for dealing with recurrent conceptsyet. DWM does not have specific features to deal withrecurrent drifts either, but it can obtain good accuracy ifthese drifts are close enough to each other so that theweights of the base learners do not decay enough for themto be eliminated.

Figs. 5c and 5d show that DDD obtains worse accuracythan DWM during the first and last thirds of the learningwhen using MLPs, but similar (or slightly better) whenusing NB. The experiments show that, if the drifts are veryclose to each other and there are no false alarms, DDD canmake use of the learning of the old concept through the oldensembles when the concept reoccurs. This is what happenswhen using NB, as the weight given to the old low diversityensemble presents peaks during the learning and thenumber of drift detections is consistent with the changesin the estimated prior probability of attack. However, falsealarms cause DDD to lose the ensembles which learnedthe old concept (the old ensembles are replaced), beingunable to use them when this concept reoccurs. This is whathappens when using MLPs, as the number of driftdetections was more than twice the number of detectionswhen using NB, probably representing false alarms.

In summary, the experiments in this section reaffirm theanalyses done in the previous sections: for a database likelyto contain several continuous drifts, DDD attained betteraccuracy than EDDM and DWM. For a database likely tocontain no drifts, DDD performed similarly to otherapproaches. EDDM would perform worse if there werefalse alarms. For a database which may contain very severeand fast drifts which reoccur with a high rate, DDDperformed similarly to DWM when it could make use ofthe ensembles which learned the old concept, but per-formed worse when these ensembles were lost.

7 CONCLUSIONS

This paper presents an analysis of low and high diversityensembles combined with different strategies to deal withconcept drift and proposes a new approach (DDD) tohandle drifts.

The analysis shows that different diversity levels obtainthe best prequential accuracy depending on the type of drift.It also shows that it is possible to use information learnedfrom the old concept in order to aid the learning of the newconcept, by training ensembles which learned the old conceptwith high diversity, using low diversity on the new concept.Such ensembles are able to outperform new ensemblescreated from scratch after the beginning of the drift,especially when the drift has low severity and high speed,and soon after the beginning of medium or low speed drifts.

DDD maintains ensembles with different diversity levels,exploiting the advantages of diversity to handle drifts andusing information from the old concept to aid the learning ofthe new concept. It has better accuracy than EDDM mainlywhen the drifts have low severity or low speed, due to theuse of ensembles with different diversity levels. DDD hasalso considerably good robustness to false alarms. Whenthey occur, its accuracy is better than EDDM’s also duringstable concepts due to the use of old ensembles. Besides,DDD’s accuracy is almost always higher than DWM’s, bothduring stable concept and after drifts. So, DDD is accurateboth in the presence and in the absence of drifts.

Future work includes experiments using a parameter tocontrol the maximum number of time steps maintainingfour ensembles, further investigation of the performance onskewed data sets, and extension of DDD to better deal withrecurrent and predictable drifts.

ACKNOWLEDGMENTS

The authors are grateful to Dr. Nikunj C. Oza for sharing hisimplementation of Online Bagging and to Garcia et al. formaking EDDM available as open source. This work waspartially funded by an Overseas Research Students Awardand a School Research Scholarship.

REFERENCES

[1] N.C. Oza and S. Russell, “Experimental Comparisons of Onlineand Batch Versions of Bagging and Boosting,” Proc. ACM SIGKDDInt’l Conf. Knowledge Discovery and Data Mining, pp. 359-364, 2001.

[2] A. Fern and R. Givan, “Online Ensemble Learning: An EmpiricalStudy,” Machine Learning, vol. 53, pp. 71-109, 2003.

[3] R. Polikar, L. Udpa, S.S. Udpa, and V. Honavar, “Learn++: AnIncremental Learning Algorithm for Supervised Neural Net-works,” IEEE Trans. Systems, Man, and Cybernetics - Part C,vol. 31, no. 4, pp. 497-508, Nov. 2001.

[4] F.L. Minku, H. Inoue, and X. Yao, “Negative Correlation inIncremental Learning,” Natural Computing J., Special Issue onnature-Inspired Learning and Adaptive Systems, vol. 8, no. 2,pp. 289-320, 2009.

[5] H. Abdulsalam, D.B. Skillicorn, and P. Martin, “StreamingRandom Forests,” Proc. Int’l Database Eng. and Applications Symp.(IDEAS), pp. 225-232, 2007.

[6] A. Narasimhamurthy and L.I. Kuncheva, “A Framework forGenerating Data to Simulate Changing Environments,” Proc. 25thIASTED Int’l Multi-Conf.: Artificial Intelligence and Applications,pp. 384-389, 2007.

[7] J. Gama, P. Medas, G. Castillo, and P. Rodrigues, “Learning withDrift Detection,” Proc. Seventh Brazilian Symp. Artificial Intelligence(SBIA ’04), pp. 286-295, 2004.

632 IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, VOL. 24, NO. 4, APRIL 2012

Fig. 6. Average false positive and negative error rates for PAKDD, resetat every third of the learning using MLPs as base learners andconsidering 30 runs.

http://ieeexploreprojects.blogspot.com

Page 15: Ddd a new ensemble approach for dealing with concept drift.bak

[8] J. Gao, W. Fan, and J. Han, “On Appropriate Assumptions to MineData Streams: Analysis and Practice,” Proc. IEEE Int’l Conf. DataMining (ICDM), pp. 143-152, 2007.

[9] F.L. Minku, A. White, and X. Yao, “The Impact of Diversity on On-Line Ensemble Learning in the Presence of Concept Drift,” IEEETrans. Knowledge and Data Eng., vol. 22, no. 5, pp. 730-742, http://dx.doi.org/10.1109/TKDE.2009.156, May 2010.

[10] N. Littlestone and M.K. Warmuth, “The Weighted MajorityAlgorithm,” Information and Computation, vol. 108, pp. 212-261,1994.

[11] N. Kasabov, Evolving Connectionist Systems. Springer, 2003.[12] M. Baena-Garcıa, J. Del Campo- �Avila, R. Fidalgo, and A. Bifet,

“Early Drift Detection Method,” Proc. Fourth ECML PKDD Int’lWorkshop Knowledge Discovery from Data Streams (IWKDDS ’06),pp. 77-86, 2006.

[13] A. Dawid and V. Vovk, “Prequential Probability: Principles andProperties,” Bernoulli, vol. 5, no. 1, pp. 125-162, 1999.

[14] W. Street and Y. Kim, “A Streaming Ensemble Algorithm (SEA)for Large-Scale Classification,” Proc. ACM SIGKDD Int’l Conf.Knowledge Discovery and Data Mining, pp. 377-382, 2001.

[15] H. Wang, W. Fan, P.S. Yu, and J. Han, “Mining Concept-DriftingData Streams Using Ensemble Classifiers,” Proc. ACM SIGKDDInt’l Conf. Knowledge Discovery and Data Mining, pp. 226-235, 2003.

[16] F. Chu and C. Zaniolo, “Fast and Light Boosting for AdaptiveMining of Data Streams,” Proc. Pacific-Asia Conf. KnowledgeDiscovery and Data Mining (PAKDD ’04), pp. 282-292, 2004.

[17] M. Scholz and R. Klinkenberg, “An Ensemble Classifier forDrifting Concepts,” Proc. Second Int’l Workshop Knowledge Dis-covery from Data Streams, pp. 53-64, 2005.

[18] M. Scholz and R. Klinkenberg, “Boosting Classifiers for DriftingConcepts,” Intelligent Data Analysis, Special Issue on KnowledgeDiscovery from Data Streams, vol. 11, no. 1, pp. 3-28, 2007.

[19] S. Ramamurthy and R. Bhatnagar, “Tracking Recurrent ConceptDrift in Streaming Data Using Ensemble Classifiers,” Proc. Int’lConf. Machine Learning and Applications (ICMLA ’07), pp. 404-409,2007.

[20] J. Gao, W. Fan, J. Han, and P. Yu, “A General Framework forMining Concept-Drifting Data Streams with Skewed Distribu-tions,” Proc. SIAM Int’l Conf. Data Mining (ICDM), 2007.

[21] H. He and S. Chen, “IMORL: Incremental Multiple-ObjectRecognition and Localization,” IEEE Trans. Neural Networks,vol. 19, no. 10, pp. 1727-1738, Oct. 2008.

[22] K. Nishida and K. Yamauchi, “Adaptive Classifiers-EnsembleSystem for Tracking Concept Drift,” Proc. Sixth Int’l Conf. MachineLearning and Cybernetics (ICMLC ’07), pp. 3607-3612, 2007.

[23] K. Nishida and K. Yamauchi, “Detecting Concept Drift UsingStatistical Testing,” Proc. 10th Int’l Conf. Discovery Science (DS ’07),pp. 264-269, 2007.

[24] K. Nishida, “Learning and Detecting Concept Drift,” PhDdissertation, Hokkaido Univ., http://lis2.huie.hokudai.ac.jp/knishida/paper/nishida2008-dissertation .pdf, 2008.

[25] K.O. Stanley, “Learning Concept Drift with a Commitee ofDecision Trees,” Technical Report AI-TR-03-302, Dept. of Com-puter Sciences, Univ. of Texas, Austin, 2003.

[26] J.Z. Kolter and M.A. Maloof, “Dynamic Weighted Majority: AnEnsemble Method for Drifting Concepts,” J. Machine LearningResearch, vol. 8, pp. 2755-2790, 2007.

[27] J.Z. Kolter and M.A. Maloof, “Using Additive Expert Ensemblesto Cope with Concept Drift,” Proc. Int’l Conf. Machine Learning(ICML ’05), pp. 449-456, 2005.

[28] K. Tumer and J. Ghosh, “Error Correlation and Error Reduction inEnsemble Classifiers,” Connection Science, vol. 8, no. 3, pp. 385-404,1996.

[29] L.I. Kuncheva and C.J. Whitaker, “Measures of Diversity inClassifier Ensembles and Their Relationship with the EnsembleAccuracy,” Machine Learning, vol. 51, pp. 181-207, 2003.

[30] E.K. Tang, P.N. Sunganthan, and X. Yao, “An Analysis ofDiversity Measures,” Machine Learning, vol. 65, pp. 247-271, 2006.

[31] G. Yule, “On the Association of Attributes in Statistics,”Philosophical Trans. Royal Soc. of London, Series A, vol. 194,pp. 257-319, 1900.

[32] J. Schlimmer and R. Granger, “Beyond Incremental Processing:Tracking Concept Drift,” Proc. Fifth Nat’l Conf. Artificial Intelligence(AAAI), pp. 502-507, 1986.

[33] “The UCI KDD Archive,” http://mlr.cs.umass.edu/ml/databases/kddcup99/kddcup99.html, 1999.

[34] M. Harries, “Splice-2 Comparative Evaluation: Electricity Pri-cing,” Technical Report UNSW-CSE-TR-9905, Artificial Intelli-gence Group, School of Computer Science and Eng., The Univ. ofNew South Wales, Sydney, 1999.

[35] P. Utgoff, N. Berkman, and J. Clouse, “Decision Tree InductionBased on Efficient Tree Restructuring,” Machine Learning, vol. 29,no. 1, pp. 5-44, 1997.

[36] F.L. Minku and X. Yao, “Using Diversity to Handle Concept Driftin On-Line Learning,” Proc. Int’l Joint Conf. Neural Networks(IJCNN ’09), pp. 2125-2132, 2009.

[37] I.H. Witten and E. Frank, Data Mining - Practical Machine LearningTools and Techniques with Java Implementations. Morgan KaufmannPublishers, 2000.

[38] N.C. Oza and S. Russell, “Online Bagging and Boosting,” Proc.IEEE Int’l Conf. Systems, Man and Cybernetics, vol. 3, pp. 2340-2345,2005.

[39] F.L. Minku and X. Yao, “On-Line Bagging Negative CorrelationLearning,” Proc. Int’l Joint Conf. Neural Networks (IJCNN ’08),pp. 1375-1382, 2008.

[40] S. Kotsiantis, D. Kanellopoulos, and P. Pintelas, “HandlingImbalanced Datasets: A Review,” GESTS Int’l Trans. ComputerScience and Eng., vol. 30, no. 1, pp. 25-36, 2006.

Leandro L. Minku received the BSc, MSc, andPhD degrees in computer science from theFederal University of Parana, Brazil, in 2003,from the Federal University of Pernambuco,Brazil, in 2006, and from the University ofBirmingham, United Kingdom, in 2011, respec-tively. His research interests include onlinelearning, concept drift, neural network ensem-bles, and evolutionary computation. He was therecipient of the Overseas Research Students

Award (ORSAS) from the British Government in 2006 for three yearsand of several Brazilian Council for Scientific and TechnologicalDevelopment (CNPq) scholarships in 2006, 2004, 2002, and 2001. Heis a member of the IEEE. (Minku 63341860)

Xin Yao (M’91-SM’96-F’03) received the BScdegree from the University of Science andTechnology of China (USTC), Hefei, Anhui, in1982, the MSc degree from the North ChinaInstitute of Computing Technology, Beijing, in1985, and the PhD degree from USTC in 1990.He worked as an associate lecturer, lecturer,senior lecturer, and associate professor in Chinaand later on in Australia. Currently, he is aprofessor at the University of Birmingham, United

Kingdom, a visiting chair professor at the USTC, and the director of theCentre of Excellence for Research in Computational Intelligence andApplications (CERCIA). He was the editor-in-chief of the IEEE Transac-tions on Evolutionary Computation from 2003 to 2008, an associateeditor or editorial board member of 12 other journals, and the editor of theWorld Scientific Book Series on Advances in Natural Computation. Hismajor research interests include several topics under machine learningand data mining. He was awarded the President’s Award for OutstandingThesis by the Chinese Academy of Sciences for his PhD work onsimulated annealing and evolutionary algorithms in 1989. He also wonthe 2001 IEEE Donald G. Fink Prize Paper Award for his work onevolutionary artificial neural networks. He is a fellow of the IEEE.

. For more information on this or any other computing topic,please visit our Digital Library at www.computer.org/publications/dlib.

MINKU AND YAO: DDD: A NEW ENSEMBLE APPROACH FOR DEALING WITH CONCEPT DRIFT 633

http://ieeexploreprojects.blogspot.com