Comparative Analysis of Different Classifiers for the Wisconsin Breast Cancer Dataset

DOI: 10.4236/oalib.1100660   PDF        1,346 Downloads   3,204 Views   Citations


The Wisconsin Breast Cancer Dataset has been heavily cited as a benchmark dataset for classification. Neural Network techniques such as Neural Networks, Probabilistic Neural Networks, and Regression Neural Networks have been shown to perform very well on this dataset. However, despite its obvious practical importance and implications for cancer research, a thorough investigation of all modern classification techniques on this dataset remains to be done. In this paper we examine the efficacy of classifiers such as Random Forests with varying number of trees, Support Vector Machines with different kernels, Naive Bayes model and neural networks on the accuracy of classifying the masses in the dataset as benign/malignant. Results indicate that Support Vector machines with a Radial Basis function kernel give the best accuracy of all the models attempted. This indicates that there are non-linearities present in the dataset and that the Support vector machine does a good job of mapping the data into a higher dimensional space in which the non-linearities fade away and the data becomes linearly separable by large margin classifier like the support vector machine. These methods show that modern machine learning methods could provide for improved accuracy for early prediction of cancerous tumors.

Share and Cite:

Vig, L. (2014) Comparative Analysis of Different Classifiers for the Wisconsin Breast Cancer Dataset. Open Access Library Journal, 1, 1-7. doi: 10.4236/oalib.1100660.

Conflicts of Interest

The authors declare no conflicts of interest.


[1] Estimated New Cancer Cases and Deaths for 2004.
[2] Wang, T.C. and Karayiannis, N.B. (1998) Detection of Microcalcifications in Digital Mammograms Using Wavelets. IEEE Transactions on Medical Imaging, 17, 49-509.
[3] Huo, Z., Giger, M., Vyborny, C., Wolverton, D., Schmidt, R. and Doi, K. (1998) Automated Computerized Classification of Malignant and Benign Mass Lesions on Digital Mammograms. Academic Radiology, 5, 155-168.
[4] Cheng, H.-D., Lui Y.M. and Freimanis, R.I. (1998) A Novel Approach to Microcalcification Detection Using Fuzzy Logic Technique. IEEE Transactions on Medical Imaging, 17, 442-450.
[5] Pendharkar, P.C., Rodger, J.A., Yaverbaum, G.J., Herman, N. and Benner, M. (1999) Association, Statistical, Mathematical and Neural Approaches for Mining Breast Cancer Patterns, Expert Systems with Applications, 17, 223-232. DRAFT VERSION of paper to Appear at the Oncology Reports, Special Issue Computational Analysis and Decision Support Systems in Oncology, Last Quarter 2005.
[6] Setiono R. (2000) Generating Concise and Accurate Classification Rules for Breast Cancer Diagnosis. Artificial Intelligence in Medicine, 18, 205-219.
[7] Chen, D., Chang, R.F. and Huang, Y.L. (2000) Breast Cancer Diagnosis Using Self-Organizing Map for Sonography. Ultrasound in Medical Biology, 26, 405-411.
[8] Giger, M., Huo, Z., Kupinski, M. and Vyborny, C. (2000) Computer-Aided Diagnosis in Mammography. In: Sonka, M. adn Fitzpatrick, J., Eds., Handbook of Medical Imaging, Medical Image Processing and Analysis, Vol. 2, SPIE Press, 386-408.
[9] Wisconsin Diagnostic Breast Cancer (WDBC) Dataset and Wisconsin Prognostic Breast Cancer (WPBC) Dataset.
[10] Tourassi, G.D., Markey, M.K., Lo, J.Y. and Floyd Jr., C.E. (2001) A Neural Network Approach to Breast Cancer Diagnosis as a Constraint Satisfaction Problem. Medical Physics, 28, 804-811.
[11] Wolberg, W.H., Street, W.N., Heisey, D.M. and Mangasarian, O.L. (1995) Computer-Derived Nuclear Features Distinguish Malignant from Benign Breast Cytology. Human Pathology, 26, 792-796.
[12] Wolberg, W.H., Street, W.N. and Mangasarian, O.L. (1994) Machine Learning Techniques to Diagnose Breast Cancer from Image-Processed Nuclear Features of Fine-Needle Aspirates. Cancer Letters, 77, 163-171.
[13] Wolberg, W.H., Street, W.N. and Mangasarian, O.L. (1995) Image Analysis and Machine Learning Applied to Breast Cancer Diagnosis and Prognosis. Analytical and Quantitative Cytology and Histology, 17, 77-87.
[14] Jiang, Y., Nishikawa, R., Wolverton, D., Metz, C., Giger, M.L., Schmidt, R. and Doi, K. (1996) Malignant and Benign Clustered Microcalcifications: Automated Feature Analysis and Classification. Radiology, 198, 671-678.
[15] Taylor, P., Fox, J. and Todd-Pokropek, A. (1998) Evaluation of a Decision Aid for the Classification of Microcalcifications. In: Digital Mammography, Kluwer Academic Publishers, Nijmegen, 237-244.
[16] Hoya, T. and Chambers, J.A. (2001) Heuristic Pattern Correction Scheme Using Adaptively Trained Generalized Regression Neural Networks. IEEE Transactions on Neural Networks, 12, 91-100.
[17] Kaban, A. and Girolami, M. (2000) Initialized and Guided EM-Clustering of Sparse Binary Data with Application to Text Based Documents. 15th International Conference on Pattern Recognition, 2, 744-747.
[18] Rosenblatt, F. (1958) The Perceptron: A Probabilistic Model for Information Storage and Organization in the Brain. Cornell Aeronautical Laboratory, Psychological Review, 65, 386-408.
[19] Minsky, M.L. and Papert, S.A. (1969) Perceptrons. MIT Press, Cambridge.
[20] Rumelhart, D.E., Hinton, G.E. and Williams, R.J. (1986) Learning Representations by Back-Propagating Errors. Nature, 323, 533-536.
[21] Kolmogorov, A.N. (1957) On the Representation of Continuous Functions of Many Variables by Superposition of Continuous Functions of One Variable and Addition. Doklady Akademii Nauk SSSR, 144, 679-681. American Mathematical Society Translation, 28, 55-59 [1963].
[22] Breiman, L. (2001) Random Forests. Machine Learning, 45, 5-32.
[23] Ho, T.K. (1995) Random Decision Forest. Proceedings of the 3rd International Conference on Document Analysis and Recognition, Montreal, 14-16 August 1995, 278-282.
[24] Chipman, H.A., George, E.I. and McCulloch, R.E. (1998) Bayesian CART Model Search. Journal of the American Statistical Association, 93, 935-948.
[25] Breiman, L. (1996) Bagging Predictors. Machine Learning, 24, 123-140.
[26] Platt, J. (1998) Sequential Minimal Optimization: A Fast Algorithm for Training Support Vector Machines, Advances in Kernel Methods. Support Vector Learning, MIT Press, Boston.
[27] Altman, N.S. (1992) An Introduction to Kernel and Nearest Neighbor Nonparametric Regression. The American Statistician, 46, 175-185.

comments powered by Disqus

Copyright © 2020 by authors and Scientific Research Publishing Inc.

Creative Commons License

This work and the related PDF file are licensed under a Creative Commons Attribution 4.0 International License.