Efficient handling of concept drift and concept evolution over Stream Data. Haque, A., Khan, L., Baron, M., Thuraisingham, B., & Aggarwal, C. In 2016 IEEE 32nd International Conference on Data Engineering (ICDE), pages 481–492, May, 2016.
doi  abstract   bibtex   
To decide if an update to a data stream classifier is necessary, existing sliding window based techniques monitor classifier performance on recent instances. If there is a significant change in classifier performance, these approaches determine a chunk boundary, and update the classifier. However, monitoring classifier performance is costly due to scarcity of labeled data. In our previous work, we presented a semi-supervised framework SAND, which uses change detection on classifier confidence to detect a concept drift. Unlike most approaches, it requires only a limited amount of labeled data to detect chunk boundaries and to update the classifier. However, SAND is expensive in terms of execution time due to exhaustive invocation of the change detection module. In this paper, we present an efficient framework, which is based on the same principle as SAND, but exploits dynamic programming and executes the change detection module selectively. Moreover, we provide theoretical justification of the confidence calculation, and show effect of a concept drift on subsequent confidence scores. Experiment results show efficiency of the proposed framework in terms of both accuracy and execution time.
@inproceedings{haque_efficient_2016,
	title = {Efficient handling of concept drift and concept evolution over {Stream} {Data}},
	doi = {10.1109/ICDE.2016.7498264},
	abstract = {To decide if an update to a data stream classifier is necessary, existing sliding window based techniques monitor classifier performance on recent instances. If there is a significant change in classifier performance, these approaches determine a chunk boundary, and update the classifier. However, monitoring classifier performance is costly due to scarcity of labeled data. In our previous work, we presented a semi-supervised framework SAND, which uses change detection on classifier confidence to detect a concept drift. Unlike most approaches, it requires only a limited amount of labeled data to detect chunk boundaries and to update the classifier. However, SAND is expensive in terms of execution time due to exhaustive invocation of the change detection module. In this paper, we present an efficient framework, which is based on the same principle as SAND, but exploits dynamic programming and executes the change detection module selectively. Moreover, we provide theoretical justification of the confidence calculation, and show effect of a concept drift on subsequent confidence scores. Experiment results show efficiency of the proposed framework in terms of both accuracy and execution time.},
	booktitle = {2016 {IEEE} 32nd {International} {Conference} on {Data} {Engineering} ({ICDE})},
	author = {Haque, Ahsanul and Khan, Latifur and Baron, Michael and Thuraisingham, Bhavani and Aggarwal, Charu},
	month = may,
	year = {2016},
	keywords = {Classifier Confidence, Concept Drift, Data mining, Data models, Dynamic Chunk, Dynamic programming, Electronic mail, Error analysis, Labeling, Training data},
	pages = {481--492},
}

Downloads: 0