Open Access Open Access  Restricted Access Subscription or Fee Access

An Efficient Active Training Algorithm for SVM-based Binary Classifiers

Sandeep Gutta, Qi Cheng

Abstract


Pattern classification has been a key task in many scientific and engineering applications. In this paper, the problem of support vector machine (SVM) based binary classification is considered. Though the SVM is an effective supervised learning algorithm, its high computational complexity often limits its real-time implementation. The support vector machines require enormous amount of training time for problems of large magnitude. In this paper, a new efficient active training approach is proposed, which makes use of the fact that the SVM decision boundary is completely characterized by a small subset of the training data. The proposed method is evaluated both theoretically and experimentally using the benchmark KDD 1999 intrusion detection datasets. The proposed method is also compared with some of the existing training methods. The experimental results show that the proposed method can successfully reduce the training time without significantly degrading the classification performance.

Keywords


Classification, Support Vector Machine, Active Training, Clustering, Convex Optimization.

Full Text:

PDF

References


S. Gutta and Q. Cheng, “An efficient training approach for SVM-based binary classifiers,” Proceedings of the International Conference on Smart Technologies (ICST), Chennai, India, Jan. 2011.

A. Smola and B. Schölkopf, “Sparse greedy matrix approximation for machine learning,” Proceedings of the Seventeenth International Conference on Machine Learning, Stanford, USA, June 2000, pp 911-918.

S. Fine and K. Scheinberg, “Efficient SVM training using low-rank kernel representations,” Journal of Machine Learning Research, 2001, pp 243-264.

B. Li, Q. Wang and J. Hu, “A fast SVM training method for very large datasets,” Proceedings of International Joint Conference on Neural Networks, Atlanta, Georgia, USA, June 2009.

S. Tong and D. Koller, “Support vector machine active learning with applications to text classification,” Journal of Machine Learning Research, 2001.

H. Shin and S. Cho, “Fast pattern selection for support vector classifiers,” Lecture Notes in Computer Science, Springer, 2003.

R. Koggalage and S. Halgamuge, “Reducing the number of training samples for fast support vector machine classification,” Neural Information Processing - Letters and Reviews, Vol. 2, No. 3, 57–65, 2004.

S. Abe and T. Inoue, “Fast training of support vector machines by extracting boundary data,” Proceedings of the International Conference on Artificial Neural Networks, 2001.

Y. J. Lee and O. L. Mangasarian, “RSVM: Reduced support vector machines,” Proceedings of the First SIAM International Conference on Data Mining, 2001.

J. Cervantes, X. Li and W. Yu, “Support vector machine classification based on fuzzy clustering for large data sets,” Lecture Notes in Computer Science, Vol. 4293, Springer, 2006.

S. Boyd and L. Vandenberghe, Convex Optimization. Cambridge University Press, 2004.

V. Vapnik, Estimation of Dependeces Based on Empirical Data. Springer-Verlag, 1982.

J. C. Platt, “Sequential Minimal Optimization: A Fast Algorithm for Training Support Vector Machines,” Advances in Kernel Methods - Support Vector Learning, 1998.

Y. Ye, M. J. Todd and S. Mizuno, “An O(√nL)-iteration homogeneous and self-dual linear programming algorithm,” Mathematics of Operations Research, 1994.

S. Hettich and S. D. Bay, The UCI KDD Archive [http://kdd.ics.uci.edu], Irvine, CA: University of California, Department of Information and Computer Science, 1999. http://kdd.ics.uci.edu/databases/kddcup99/kddcup99.html


Refbacks

  • There are currently no refbacks.