Abstract
In this paper, we propose a new optimization framework for improving feature selection in medical data classification. We call this framework Support Feature Machine (SFM). The use of SFM in feature selection is to find the optimal group of features that show strong separability between two classes. The separability is measured in terms of inter-class and intra-class distances. The objective of SFM optimization model is to maximize the correctly classified data samples in the training set, whose intra-class distances are smaller than inter-class distances. This concept can be incorporated with the modified nearest neighbor rule for unbalanced data. In addition, a variation of SFM that provides the feature weights (prioritization) is also presented. The proposed SFM framework and its extensions were tested on 5 real medical datasets that are related to the diagnosis of epilepsy, breast cancer, heart disease, diabetes, and liver disorders. The classification performance of SFM is compared with those of support vector machine (SVM) classification and Logical Data Analysis (LAD), which is also an optimization-based feature selection technique. SFM gives very good classification results, yet uses far fewer features to make the decision than SVM and LAD. This result provides a very significant implication in diagnostic practice. The outcome of this study suggests that the SFM framework can be used as a quick decision-making tool in real clinical settings.
Similar content being viewed by others
References
Asuncion, A., & Newman, D. (2007). UCI machine learning repository.
Bradley, P. S., & Mangasarian, O. L. (1998). Feature selection via concave minimization and support vector machines. In Proc. 15th international conf. on machine learning (pp. 82–90). San Francisco: Kaufmann.
Chaovalitwongse, W., Fan, Y., & Sachdeo, R. (2007). On the time series k-nearest neighbor for abnormal brain activity classification. IEEE Transactions on Systems, Man, and Cybernetics, Part A: Systems and Humans, 37(6), 1005–1016.
Chaovalitwongse, W., Pardalos, P., & Prokoyev, O. (2006). Electroencephalogram (EEG) time series classification: applications in epilepsy. Annals of Operations Research, 148, 227–250.
Chaovalitwongse, W. A., Fan, Y. J., & Sachdeo, R. C. (2008, to appear). Novel optimization models for abnormal brain activity classification. Operations Research.
Fung, G., & Mangasarian, O. L. (2002). A feature selection Newton method for support vector machine classification. Computational Optimization and Applications, 28, 185–202.
Fung, G., & Mangasarian, O. L. (2003). Finite Newton method for Lagrangian support vector machine classification. Neurocomputing, 1–2, 39–55.
Glen, J. J. (2003). An iterative mixed integer programming method for classification accuracy maximizing discriminant analysis. Computers and Operations Research, 30, 181–198.
Glover, F. (1990). Improved linear programming models for discriminant analysis. Decision Sciences, 21(4), 771–785.
Glover, F. (1993). Improved linear and integer programming models for discriminant analysis. Creative and innovative approaches to the science of management (pp. 187–215). USA: RGK Foundation Press.
Hammer, P. L., & Bonates, T. O. (2006). Logical analysis of data—An overview: From combinatorial optimization to medical applications. Annals of Operations Research, 148, 203–225.
Zhang, J., & Mani, I. (2003). knn approach to unbalaned data distributions: A case study involving information extration. Workshop on Learning from Imbalanced Datasets II, ICML.
Kim, S. B., Chen, V. C. P., Park, Y., Ziegler, T. R., & Jones, D. P. (2008). Controlling the false discovery rate for feature selection in high-resolution nmr spectra. Statistical Analysis and Data Mining, 1(2), 57–66.
Mangasarian, O. (1965). Linear and nonlinear separation of pattern by linear programming. Operations Research, 31, 445–453.
Mangasarian, O. (1968). Multisurface method for pattern separation. IEEE Transactions on Information Theory, IT-14, 801–807.
Pardalos, P. M., Boginski, V. L., & Vazacopoulos, A. (2007). Data mining in biomedicine. New York: Springer.
Saastamoinen, K., & Ketola, J. (2006). Medical data classification using logical similarity based measures. In Proceedings of 2006 IEEE conference on cybernetics and intelligent systems (pp. 1–5).
Tan, S. (2005). Neighbor-weighted k-nearest neighbor for unbalanced text corpus. Expert Systems with Applications, 28(4), 667–671.
Tsirogiannis, G., Frossyniotis, D., Stoitsis, J., Golemati, S., Stafylopatis, A., & Nikita, K. (2004). Classification of medical data with a robust multi-level combination scheme. In Proceedings of 2004 IEEE international joint conference on neural networks (Vol. 3, pp. 25–29).
Vapnik, V. N. (1998). Statistical learning theory. New York: Wiley.
Zhang, H. H., Wahba, G., Lin, Y., Voelker, M., Ferris, M., Klein, R., & Klein, B. (2004). Variable selection and model building via likelihood basis pursuit. Journal of the American Statistical Association, 99(467), 659–672.
Author information
Authors and Affiliations
Corresponding author
Additional information
This work is supported by the National Science Foundation under CAREER Grant No. 0546574.
Rights and permissions
About this article
Cite this article
Fan, YJ., Chaovalitwongse, W.A. Optimizing feature selection to improve medical diagnosis. Ann Oper Res 174, 169–183 (2010). https://doi.org/10.1007/s10479-008-0506-z
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10479-008-0506-z