Open Access Open Access  Restricted Access Subscription or Fee Access

Learning using Heterogeneous Classifier in Data Mining

Amit Thakkar, Reshma Idresh Lakhani, Amit Ganatra

Abstract


Data Mining can be considered an analytic process designed to explore business or market data to search for consistent patterns and/or systematic relationships between variables, and then to validate the findings by applying the detected patterns to new subsets of data. Data mining is useful for prediction. We can improve accuracy of different classifiers by combining various classifiers and taking their predictions. One such method is Stacking, an ensemble method in which a number of base classifiers are combined using one meta-classifier which learns their outputs. This enhances the benefits obtained by individual classifiers. This paper is a review work of different approaches proposed by various authors in their paper.

Keywords


Ensemble of Classifiers, Bagging, Boosting, Staking, Troika.

Full Text:

PDF

References


L. Breiman. “Bagging predictors”. Machine learning, Kluwer Academic publishers, Boston. vol. 24, pp123-240. 1996.

Y. Freund and R. Schapire. “Experiments with a new boosting algorithm”. Machine learning: Proceedings of the Thirteenth International Conference on Machine Learning. Morgan Kaufmann. 1996.

3] D. Wolpert, “Stacked Generalization”. Neural Networks, vol. 5, No2,pp241-260. 1992.

A. Asuncion, D. J. Newman. “UCI Machine Learning Repository”. Irvine, CA: University of California, Department of Information and Computer Sciences. 2007.

P.K. Chan. “An Extensible Meta-learning Approach for Scalable and Accurate inductive learning”, Ph..D thesis, Columbia University, 1996.

A.L. Prodromidis, P.K. Chan, S.J. Stolfo, “Meta-learning in Distributed Data Mining Systems: Issues and Approaches”. In H. Kagupta & P. Chan(Eds.), Advances in distributed and parallel knowledge discovery. MIT/AAAI Press, 2000.

K. M. Ting, Ian H. Witten, “Issues in Stacked Generalization”, Journal of Artificial Intelligence Research, USA, vol 10, pp271- 289,1999.

A.K. Seewald, J. Furnkranz. “An Evalutation of Grading Classifiers”, In Hoffmann F.et al.(eds.), Advances in Intelligent Data Analysis, Proc. 4th International Conference, Springer, pp.115-124, 2001.

S Džeroski , B Ženko, “Is Combining Classifiers with Stacking Better than Selecting the Best One?”, Machine Learning, USA, vol.54, no.3, pp.255-273, 2004

C.J. Merz, “Using Correspondence analysis to Combine Classifiers”. Machine Learning, USA, Vol. 36, pp.33-58. 1999

G. Tsoumakas, I. Vlahavas. “Distributed Data Mining of Large Classifier Ensembles”. Proceedings of 2nd Hellenic Conference on AI, SETN-2002, Thessaloniki, Greece, Companion Volume, pp249- 256. 2002.

Nima Hatami, Reza Ebrahimpour. “Combining Multiple Classifiers: Diversify with Boosting and Combining by Stacking”. International Journal of Computer Science and Network Security, vol.7, no.1, January, 2007.

Niall Rooney, David Patterson, Chris Nugent. “Pruning extensions to stacking”. Intelligent Data Analysis, USA, vol.10, no.1, pp.47-66, IOS Press. 2006 .

G. Hulten, P. Domingos. “Vfml-A toolkit for mining high-speed time-changing data streams”, [http://www.cs.washington.edu/dm/ vfml/main.html]. 2003.

Ian H. Witten, Eibe Frank. “Data Mining: practical machine learning tools and techniques”, 2nd Edition, Morgan Kaufmann, San Francisco. 2005.

Combining Distributed Classifies by Stacking - WEI Yanyan, LI Taoshen, GE Zhihui

Troika-An Improved Stacking Schema for Classification Tasks- Eitan Menahem, Lior Rokach, Yuval Elovici

Data Mining, Concepts and Techniques, Jiawei Han and Micheline Kamber

http://www.anderson.ucla.edu/faculty/jason.frand/teacher/technologies/palace/datamining.htm

http://archive.ics.uci.edu/ml/datasets.html


Refbacks

  • There are currently no refbacks.


Creative Commons License
This work is licensed under a Creative Commons Attribution 3.0 License.