Acessibilidade / Reportar erro

Wrappers Feature Selection in Alzheimer’s Biomarkers Using kNN and SMOTE Oversampling Work presented at Congress of Applied and ComputationalMathematics - CMAC SE 2015.

ABSTRACT

Biomarkers are characteristics that are objectively measured and evaluated as an indicator of normal biological processes, pathogenic processes or pharmacological responses to a therapeutic intervention. The combination of different biomarker modalities often allows an accurate diagnosis classification. In Alzheimer’s disease (AD), biomarkers are indispensable to identify cognitively normal individuals destined to develop dementia symptoms. However, using the combination of canonical AD biomarkers, studies have repeatedly shown poor classification rates to differentiate between AD, mild cognitive impairment and control individuals. Furthermore, the design of classifiers to access multiple biomarker combinations includes issues such as imbalance classes and missing data. Due to the number of biomarkers combinations wrappers are used to avoid multiple comparisons. Here, we compare the ability of three wrappers feature selection methods to obtain biomarker combinations which maximize classification rates. Also, as the criterion to the wrappers feature selection we use the k-nearest neighbor classifier with balance aids, random undersampling and SMOTE oversampling. Overall, our analyses showed how biomarkers combinations affect the classifier precision and how imbalance strategy improve it. We show that non-defining and non-cognitive biomarkers have less precision than cognitive measures when classifying AD. Our approach surpasses in average the support vector machine and the weighted k-nearest neighbor classifiers and reaches 94.34 ± 3.91% of precision reproducing class definitions.

Keywords:
k-nearest neighbor; SMOTE; feature selection; Alzheimer’s biomarkers; Alzheimer’s disease classification

RESUMO

Biomarcadores são medidas biológicas que ajudam a rastrear e compreender a progressão fisiopatológica de várias doenças. A combinação de diferentes modalidades de biomarcadores muitas vezes permite uma classificação de diagnóstico preciso. Na doença de Alzheimer (DA), os biomarcadores são indispensáveis para identificar indivíduos cognitivamente normais destinados a desenvolver sintomas de demência. No entanto, usando combinações de biomarcadores canônicas DA estudos têm mostrado repetidamente que as taxas de classificação são baixas quando diferenciando entre indivíduos controle, comprometimento cognitivo leve e DA. Além disso, na avaliação de múltiplas combinações de biomarcadores os classificadores enfrentam dificuldade tais como falta de dados e dados desbalanceados. Uma vez que o número de combinações biomarcadores é fatorial então usamos wrappers para evitar as múltiplas comparações. Neste trabalho comparamos a capacidade de três técnicas wrapper de selecção de caracteristicas na obtenção de combinações de biomarcadores ao maximizar taxas de classificação. Além disso, como critério para os wrappers usamos o classificador k-vizinhos mais próximos com preprosessamento de balanço de dados, sobreamostragem aleatória e sobamostragem (SMOTE). Em geral nossa análise mostra como as combinações de biomarcadores são afetadas pela estratégia de equilíbrio de dados. Mostramos que os biomarcadores não-definidores de classe e não-cognitivos têm menos precisão do que as medidas cognitivas para classificar AD. A nossa abordagem supera em média os classificadores máquina de vetores de suporte e k-vizinhos mais próximos ponderado com 94,34±3,91% de precisão para biomarcadores que definem a classe.

Palavras-chave:
k-vizinhos mais próximos; SMOTE; seleção de características; biomarcadores de Alzheimer; problema de classificação

- For the Alzheimer’s Disease Neuroimaging Initiative (ADNI)** ** Data used in the preparation of this article were obtained from the ADNI database. As such, the investigators within the ADNI contributed to the provided data but did not participate in analysis or writing of this report.

1 INTRODUCTION

Alzheimer’s disease (AD) is the most common cause of dementia worldwide, posing enormous economic and social costs for the society11 M.S. Fiandaca, M.E. Mapstone, A.K. Cheema & H.J. Federoff. The critical need for defining preclinical biomarkers in Alzheimer’s disease. Alzheimer’s & Dementia, 10(3) (2014), S196-S212.. AD22 C.R. Jack, D.S. Knopman, W.J. Jagust, L.M. Shaw, P.S. Aisen, M.W. Weiner, R.C. Petersen & J.Q. Trojanowski. Hypothetical model of dynamic biomarkers of the Alzheimer’s pathological cascade. The Lancet Neurology, 9(1) (2010), 119-128. is pathophysiologically characterized by the gradual brain deposition of amyloid plaques, neurofibrillary tangles, and eventual neuronal depletion22 C.R. Jack, D.S. Knopman, W.J. Jagust, L.M. Shaw, P.S. Aisen, M.W. Weiner, R.C. Petersen & J.Q. Trojanowski. Hypothetical model of dynamic biomarkers of the Alzheimer’s pathological cascade. The Lancet Neurology, 9(1) (2010), 119-128.. The AD spectrum is composed by preclinical (CN), mild cognitive impairment (MCI) and AD dementia phases22 C.R. Jack, D.S. Knopman, W.J. Jagust, L.M. Shaw, P.S. Aisen, M.W. Weiner, R.C. Petersen & J.Q. Trojanowski. Hypothetical model of dynamic biomarkers of the Alzheimer’s pathological cascade. The Lancet Neurology, 9(1) (2010), 119-128.. Preclinical AD individuals are those cognitively normal with amyloid plaques and tangles, individualswith MCI have cognitive symptoms withoutmeeting clinical criteria for dementia, and AD dementia individuals present severely compromised cognitive faculties33 R.A. Sperling, P.S. Aisen, L.A. Beckett, D.A. Bennett, S. Craft, A.M. Fagan, T. Iwatsubo, C.R. Jack, J. Kaye & T.J. Montine et al. Toward defining the preclinical stages of Alzheimer’s disease: Recommendations from the national institute on aging-Alzheimer’s association workgroups on diagnostic guidelines for Alzheimer’s disease. Alzheimer’s & dementia, 7(3) (2011), 280-292.. In recent years, a plethora of biomarkers has been developed in order to track AD progression, such as biomarkers for beta peptide 1-42 (Aβ 1−42) and tau proteins that indicate the presence of the hallmark pathological features of AD, amyloid plaques, and neurofibrillary tangles, respectively11 M.S. Fiandaca, M.E. Mapstone, A.K. Cheema & H.J. Federoff. The critical need for defining preclinical biomarkers in Alzheimer’s disease. Alzheimer’s & Dementia, 10(3) (2014), S196-S212.), (22 C.R. Jack, D.S. Knopman, W.J. Jagust, L.M. Shaw, P.S. Aisen, M.W. Weiner, R.C. Petersen & J.Q. Trojanowski. Hypothetical model of dynamic biomarkers of the Alzheimer’s pathological cascade. The Lancet Neurology, 9(1) (2010), 119-128..

It is a well-established fact that combined biomarkers provide higher classification rates than single biomarkers44 A.A. Motsinger-Reif, H. Zhu, M.A. Kling, W. Matson, S. Sharma, O. Fiehn, D.M. Reif, D.H. Appleby, P.M. Doraiswamy & J.Q. Trojanowski et al. Comparing metabolomic and pathologic biomarkers alone and in combination for discriminating Alzheimer’s disease from normal cognitive aging. Acta neuropathologica communications, 1(1) (2013), pp. 28.. In this regard, neuropsychological tests associated with different biomarker modalities have been used to classify AD55 S.J. Teipel, O. Sabri, M. Grothe, H. Barthel, D. Prvulovic, K. Buerger, A.L. Bokde, M. Ewers, W. Hoffmann & H. Hampel. Perspectives for multimodal neurochemical and imaging biomarkers in Alzheimer’s disease. Journal of Alzheimer’s Disease, 33(s1) (2013), S329-S347.. These studies have combined positron emission tomography (PET), magnetic resonance imaging (MRI), functionalmagnetic resonance imaging (fMRI) as well as cerebrospinal fluid (CSF) and blood biomarkers to perform binary classifications of AD11 M.S. Fiandaca, M.E. Mapstone, A.K. Cheema & H.J. Federoff. The critical need for defining preclinical biomarkers in Alzheimer’s disease. Alzheimer’s & Dementia, 10(3) (2014), S196-S212., e.g. healthy versus unhealthy individuals. Despite AD’s classification problembeing an inherentlymulticlass binary classificationmulticlass driven by binary strategies are the rule11 M.S. Fiandaca, M.E. Mapstone, A.K. Cheema & H.J. Federoff. The critical need for defining preclinical biomarkers in Alzheimer’s disease. Alzheimer’s & Dementia, 10(3) (2014), S196-S212.. This happens since some classifiers are naturally binary and must be adapted to be multiclass by means of one-versus-all and one-versus-one strategies, e.g. support vector machine (SVM)66 C.M. Bishop. Pattern Recognition and Machine Learning (Information Science and Statistics). Springer, (2007).. Thus, to solve an n-class problem using binary classifiers, n(n-1)/2 rules are required to build a multiclass classifier. As benefit, binary classifiers are well suited for the receiver operator characteristic (ROC) analysis77 T. Fawcett. An introduction to roc analysis. Pattern recognition letters, 27(8) (2006), 861-874. which has been largely applied in AD comparative studies, biomarkers model selection and conversion diagnosis prediction.

Recently, various approaches used for AD’s identification have achieved successful results and satisfactory classification rates. For instance, Khedher et al.88 L. Khedher, J. Ramírez, J.M. Górriz, A. Brahim, F. Segovia & A.D.N. Initiative et al. Early diagnosis of Alzheimer’s disease based on partial least squares, principal component analysis and support vector machine using segmented mri images. Neurocomputing, 151(2015), 139-150. were able to accurately differentiating the three clinical classes of the AD spectrum reaching the maximum sensitivity (85.11%), specificity (91.27%), and accuracy (88.49%) values by implementing binary strategy and reduction of input space with SVM and principal component analysis (PCA) techniques66 C.M. Bishop. Pattern Recognition and Machine Learning (Information Science and Statistics). Springer, (2007).. Khazaee et al.99 A. Khazaee, A. Ebrahimzadeh & A. Babajani-Feremi. Identifying patients with Alzheimer’s disease using resting-state fmri and graph theory. Clinical Neurophysiology, 126(11) (2015), 2132-2141. were able to perfectly differentiate between cognitively healthy and AD classes in a small dataset of 40 individuals, using graph theory applied to brain connectivity assessed with fMRI they reach an accuracy of 100% for linear SVM and 87.5% for k-nearest neighbor (kNN). Although the separation between extreme cases is straightforward, difficulties are expected when considering the overlapped intermediary classes. Classifiers performance can be potentially affected by data issues, such as class overlapping, feature space with high dimensionality, missingdata, class imbalance, etc1010 Q. Yang & X. Wu. 10 challenging problems in data mining research. International Journal of Information Technology & Decision Making, 5(04) (2006), 597-604.. Particularly, imbalanced datasets1111 H. He & E.A. Garcia. Learning from imbalanced data. IEEE Transactions on knowledge and data engineering, 21(9) (2009), 1263-1284. are considered one of the 10 most challenging problems in machine learning 1010 Q. Yang & X. Wu. 10 challenging problems in data mining research. International Journal of Information Technology & Decision Making, 5(04) (2006), 597-604.. When imbalanced data issues are disregarded, it could lead to a decreased classification rate for the minor represented class and in globally averaged scores1212 B. Krawczyk. Learning from imbalanced data: open challenges and future directions. Progress in Artificial Intelligence, (2016), 1-12.. Solutions that address this problem are based on re-sampling methods when the distribution mechanisms are known. Alternatively, the methods are mainly based on the creation of synthetic data for minor classes or/and pruning data for major classes1111 H. He & E.A. Garcia. Learning from imbalanced data. IEEE Transactions on knowledge and data engineering, 21(9) (2009), 1263-1284..

There is a need to identify biomarker combinations that maximize the classification and understand how much they contribute to differentiate between AD classes1313 C. Humpel. Identifying and validating biomarkers for Alzheimer’s disease. Trends in biotechnology, 29(1) (2011), 26-32.), (11 M.S. Fiandaca, M.E. Mapstone, A.K. Cheema & H.J. Federoff. The critical need for defining preclinical biomarkers in Alzheimer’s disease. Alzheimer’s & Dementia, 10(3) (2014), S196-S212.. However, this goal faces multiple classifier’s comparisons when assessing biomarker combinations. In order to avoid the excessive number of comparisons, feature selection techniques are able to find a set of biomarkers that meet defined criteria1414 I. Guyon & A. Elisseeff. An introduction to variable and feature selection. Journal of machine learning research, 3(Mar) (2003), 1157-1182.. For a given task (e.g. classification) examples of criteria are: to identify the most cost-effective biomarkers, with higher precision and low false-positive, find a subspace of reduced dimensionalitywith the same or enhanced discriminant properties; extract/build relevant features from raw data1515 Y. Saeys, I. Inza & P. Larrañaga. A review of feature selection techniques in bioinformatics. Bioinformatics, 23(19) (2007), 2507-2517..

Techniques of feature selection have been largely applied to AD-related problems, intending to provide a better understanding of biomarkers relationship1313 C. Humpel. Identifying and validating biomarkers for Alzheimer’s disease. Trends in biotechnology, 29(1) (2011), 26-32. and achieve defined criteria of usefulness1414 I. Guyon & A. Elisseeff. An introduction to variable and feature selection. Journal of machine learning research, 3(Mar) (2003), 1157-1182.. Interesting applications of feature selection techniques contributed to the understanding of AD, like the construction of potential biomarkers for enhanced classification. For instance, Lopez-de-Ipiña et al. owing to determine preclinical biomarkers for AD apply feature selection techniques on spontaneous speech to extract discriminant features1616 K. Lopez-de Ipiña, J.B. Alonso, J. Solé-Casals, N. Barroso, P. Henriquez, M. Faundez-Zanuy, C.M. Travieso, M. Ecay-Torres, P. Martinez-Lage & H. Eguiraun. On automatic diagnosis of Alzheimer’s disease based on spontaneous speech analysis and emotional temperature. Cognitive Computation, 7(1) (2015), 44-55.. They also were able to correctly classify AD subjects using kNN and multilayer perceptron (MLP) classifiers obtaining accuracy of 87.30% and 90.90%, respectively to each classifier. Feature selection AD-related also is found in the gene microarray analysis1515 Y. Saeys, I. Inza & P. Larrañaga. A review of feature selection techniques in bioinformatics. Bioinformatics, 23(19) (2007), 2507-2517. and neuroimaging both with highdimensional feature spaces and its own big data challenges. These fields have been provoking adaptation of feature selection techniques to deal with high dimensionality (tens of thousands features) and small sample size in the case of microarray datasets1515 Y. Saeys, I. Inza & P. Larrañaga. A review of feature selection techniques in bioinformatics. Bioinformatics, 23(19) (2007), 2507-2517.. In neuroimaging, the feature selection methods in 3D matrices are able to mitigate performances issues and improve the classification accuracy1717 A. Sarica, G. Di Fatta, G. Smith, M. Cannataro & J.D. Saddy et al. Advanced feature selection in multinominal dementia classification from structural mri data, in Proc MICCAI Workshop Challenge on Computer-Aided Diagnosis of Dementia Based on Structural MRI Data, pp. 82-91 (2014)..

Here, we propose to find subsets of features among several feature combinationswhich maximize classification rates between three AD classes. Specifically, we solve a multiclass classification problem in which test patterns are assigned to one of following classes: control normal (CN), mild cognitive impairment (MCI) or AD. To do that, we compare three feature selection techniques that depend on the classifier’s outcome as a measure of usefulness1414 I. Guyon & A. Elisseeff. An introduction to variable and feature selection. Journal of machine learning research, 3(Mar) (2003), 1157-1182.. This requirement characterizes the feature selection techniques called wrappers which select features based on the classifier’s performance. However, instead of widely applied binary strategies, here we will use the all-versus-all strategy naturally achieved by the kNN classifier. The misclassification and the comparison between the biomarker combinations will be done by scalar measures of confusion matrices1818 J.S. Marques. Reconhecimento de Padrões: métodos estatistícos e neuronais. IST press (2005).. In order to observe the effect of training set size, we compare two validation processes, 10-fold cross-validation (10-fold CV) and leave-one-out cross-validation (LOOCV)66 C.M. Bishop. Pattern Recognition and Machine Learning (Information Science and Statistics). Springer, (2007).. Our analysis shows how the imbalanced dataset affects classification rates and shows a comparison of the feature’s probability to reach higher precision. Two techniques to aid the class balances are compared with the imbalanced dataset: an oversampling technique called synthetic minority oversampling technique (SMOTE) and the random undersampling. All algorithms and plot codes are available on-line https://github.com/yurier/TEMA-R-CODES.

2 METHODS

2.1 Dataset

Data used in the preparation of this article were obtained from the ADNI database (adni.loni.usc.edu). The ADNI was launched in 2003 as a public-private partnership, led by Principal Investigator Michael W. Weiner, MD. The goal of ADNI has been to test whether serial MRI, PET, other biological markers, and clinical and neuropsychological assessment can be combined to measure the progression of MCI and early AD. For up-to-date information, see www.adni-info.org. Features in this work consist of two neuroimaging biomarkers (labels 1,2), four neuropsychological tests (labels 3,4,5,6) and two proteomic biomarkers (labels 7,8)1919 T. Tapiola, I. Alafuzoff, S.-K. Herukka, L. Parkkinen, P. Hartikainen, H. Soininen & T. Pirttilä. Cerebrospinal fluid β-amyloid 42 and tau proteins as biomarkers of Alzheimer-type pathologic changes in the brain. Archives of neurology, 66(3) (2009), 382-389., respectively: 2-[18F]fluoro-2-Deoxy-D-glucose (FDG) PET, florbetapir-fluorine-18 (18F-AV-45) PET, clinical dementia rating sum of boxes (CDRSB), Alzheimer’s disease assessment scale-cognitive with 11 items (ADAS11), mini mental state examination (MMSE), Ray auditory verbal learning test percent forgetting (RAVLT), Aβ 1−42 CSF, phosphorylated tau protein (p-tau181) CSF. Table 1 depicts dataset demographics.

Table1:
Dataset demographics described by mean and standard deviation.

Due to different probability in pathologies stages, clinical datasets are subject to imbalanced classes. Also, the data imbalance is a critical issue that generates an unfair separation between classes, when the imbalance is extreme1212 B. Krawczyk. Learning from imbalanced data: open challenges and future directions. Progress in Artificial Intelligence, (2016), 1-12. it will lead to decreased prediction rates for validation stages. The present study evaluates two strategies to adjust the class imbalance by assuming any class posterior distribution. Since we are using the kNN that is a distance based algorithm features were scaled by min-max normalization in all experiments2121 C.-M. Ma, W.-S. Yang & B.-W. Cheng. How the parameters of k-nearest neighbor algorithm impact on the best classification accuracy: In case of parkinson dataset. Journal of Applied Sciences, 14(2) (2014), 171-176. except for figures.Importantly, since features CDRSB and MMSE are employed to define the diagnosis or are similar to categorization protocol they will be used as contrast. The feature-wise comparison will be performed only for non-defining and non-cognitive features since it is well known thatcognitive measures have more discriminative power11 M.S. Fiandaca, M.E. Mapstone, A.K. Cheema & H.J. Federoff. The critical need for defining preclinical biomarkers in Alzheimer’s disease. Alzheimer’s & Dementia, 10(3) (2014), S196-S212.. Moreover, feature selection will beapplied for non-defining features which include cognitive measures.

2.2 k-nearest neighbors (kNN) algorithm

Classifiers can be defined by discriminant functions1818 J.S. Marques. Reconhecimento de Padrões: métodos estatistícos e neuronais. IST press (2005)., which are a set of functions to predict categorical dependent patterns. Here, we define a classifier C as a function that assigns a pattern x ∈ℝn to a class into the class space ω ∈ Ω := {ω 1, . . . ,ωc },

C(x): ℝn →Ω

The maximum a posteriori (MAP) classifier uses a set of discriminant functions to assign the most probable class. Let {fωi (x)}ci =1 be a set of discriminant functions, a classifier C is said to be well-defined if, for all patterns, is possible to assign a class. Let be the calculated class, the MAP classifier is given by,

(2.1)

The discriminant functions set is monotonous, since for a given x associated to a class ωj we have that fωi (x) ≥ fωj (x) ∀ij. An example of a set of discriminant functions is the class conditional probability {p(x|ωi)}ci =11818 J.S. Marques. Reconhecimento de Padrões: métodos estatistícos e neuronais. IST press (2005).. The classifiers taxonomy is based on how they approximate its discriminant functions66 C.M. Bishop. Pattern Recognition and Machine Learning (Information Science and Statistics). Springer, (2007).. Generative models parametrically approximate the posterior class probabilities, p(x|ωi ), through the class conditional probability p(ωi|x ) and the class prior probability, p(ωi ) (e.g. gaussian, gamma, etc). Alternatively, discriminative models directly approximate posterior class probabilities, p(ωi|x ), without assuming a distribution for it (e.g. kNN, MLP, etc). Aside from generative and discriminative models, there are the non-probabilistic models in which the discriminant functions are not required to be a distribution (e.g. SVM). Generally, the goal of discriminant functions is to divide the pattern space into decision regions {R 1 , . . . , Rc }. Class densities are depicted in Figure 1 and the classification mapping example is depicted in Figure 2. Also, Figure 2 shows a binary classification problem for feature combination labels 7 and 8 in which overlapped regions are subject to higher misclassification.

Figure 1:
On the left and middle, class distributions for CN and AD classes, respectively. On the right, AD and CN classes distribution overlapped in some regions.

Figure 2:
Decision boundaries created by the 3NN classifier.

In order to observe how the distributions of CN and AD are overlapped, one can use the Bhattacharyya distance2222 A. Bhattacharyya. On a measure of divergence between two multinomial populations. Sankhyā: the indian journal of statistics, 7(4) (1946), 401-406. which is a metric to measure how two distributions differ and also provides a bound for the probability of classification error using the Bayes optimum classifier. By assuming Gaussian shapes the Bhattacharyya coefficient coincides with Mahalanobis measure. The Bhattacharyya coefficient2222 A. Bhattacharyya. On a measure of divergence between two multinomial populations. Sankhyā: the indian journal of statistics, 7(4) (1946), 401-406. derived from Bhattacharyya distance ranges between 0 (distribution non-overlapped) and 1 (distribution overlapped). Assuming normal distributions between CN and AD distribution the Bhattacharyya coefficient is 0.6398824. Further this measure will be used to show how oversampling modifies the original distribution.

The kNN is a non-parametric classifier that estimates the class posterior probability assuming that nearest patterns (here assumed to be vectors in a metric space) have similar properties (classes)2323 L. Devroye, L. Györfi & G. Lugosi. A probabilistic theory of pattern recognition, 31. Springer Science & Business Media, 31(2013), 638 pages.. This assumption was proved for infinite sample case as shown in2323 L. Devroye, L. Györfi & G. Lugosi. A probabilistic theory of pattern recognition, 31. Springer Science & Business Media, 31(2013), 638 pages.. There are preprocessing techniques, which depend on kNN, for instance: in data imbalance, completing missing values, dimensionality reduction, and metric learning. Here we focus specifically on data imbalance challenge and a feasible solution; SMOTE algorithm.

The kNN classifier deals with multiclass problems straightforwardly by means of all-versus-all strategy. To achieve this, it uses a neighborhood defined by k training instances nearest to a query instance to be classified. Let T = {(xi, ωi )}mi =1 be a training set, with tuples xi ∈ℝn and ωi ∈Ω as labeled patterns. Also, let x * be a test instance with an unknown label to be assigned into a class ω* ∈Ω. Using the MAP framework given in equation (2.1) the kNN classifier can be written as,

(2.2)

where N(x *,k) is a neighborhood with k training instances around x * for a given metric and δ(.,.) is the Kronecker delta function. Rewriting equation (2.2) we have,

(2.3)

The monotonicity of discriminant functions is applied to equation (2.3) and dividing it by k,

(2.4)

Thus, each term in equation (2.4) is the posterior class probability,

(2.5)

From equation (2.5) we have that the most probable class for the training instance x * is given by,

(2.6)

The parameter k that adjusts the k-neighborhood N(x *,k) is searched empirically since there are no guidelines for its optimality. However, Bhattacharyya2424 G. Bhattacharya, K. Ghosh & A.S. Chowdhury. An affinity-based new local distance function and similarity measure for knn algorithm. Pattern Recognition Letters, 33(3) (2012), 356-363. proposes a bound to the optimal k, that is k< k by using only odd values in order to avoid ties in equation (2.1). The class posterior distribution, p(ωj|x ), is an alternative to the optimum Bayesian classification approach, which requires the complete knowledge ofdata generation underlying mechanisms. As showed by Cover & Hart2525 T.M. Cover & P.E. Hart. Nearest neighbor pattern classification. Information Theory, IEEE Transactions on, 13(1) (1967), 21-27., the error of the nearest neighbor is limited between the optimal error and twice the optimal error for the infinite sample case. This means that the more data available the closer the optimum error will be. This was supported by Stone on the existence of a universally consistent classifier2323 L. Devroye, L. Györfi & G. Lugosi. A probabilistic theory of pattern recognition, 31. Springer Science & Business Media, 31(2013), 638 pages.. Let’s show how the parameters affect the classifier properties. Figure 3 depicts the role of parameter k in the decision regions as well as the probability of class assignment described by the equation (2.6). Also, in Figure 3, note that there are only two probability values for k=1 (1NN classifier). This happens since there are no misclassification errors or ties for training set in 1NN. A perfect score in training set would be an indicator of overfitting that generally leads to poor performance issues, that is, the classifier is unable to generalize the training results. Contrasting to 1NN, when parameter k is increased, more data is needed in order to evaluate the pattern assignment leading to more stratified probability values.. In binary classification, one can restrain the range of

Figure 3:
Left to right, the influence of parameter k in the decision boundaries, for k=1, k=5 and k=15, respectively.

Non-parametric methods as kNN do not rely on distribution assumptions1818 J.S. Marques. Reconhecimento de Padrões: métodos estatistícos e neuronais. IST press (2005). (e.g. Gaussian shape). As an example see that the kNN classifier given in equation (2.1) depends only on the training data T and k to approximate the class posterior distribution. That is to say, data modification or removals imply in different classifiers since T is modified. By that, the biomarker combinations are validated using 10-fold CV and LOOCV in order to observe the variation between different training data sizes. Although, variations using the same dataset also can occur due to ties inkNN. For instance, with 5NN some query pattern class could be evaluated as “AD+AD+CN+CN+MCI”. The random tie-breaking is adopted in this work because is computationally more efficient than tie-breaking strategies described in2626 T. Bailey & A.K. Jain. A Note on Distance-Weighted k-Nearest Neighbor Rules. IEEE Transactions on Systems, Man, and Cybernetics, SMC-8(4) (1978), 311-312.. Figure 4 shows the effect of random tie breaking in the decision regions for the three class problem proposed.

Figure 4:
Decision region for a three class problem solved with 5NN using the same number of samples for each class. From left to right, decision region of CN, MCI and AD, respectively.

One can observe in Figure 4, apart from noisy region created by the random tie breaking, the decision regions are complementary. Furthermore, unlike the binary case that requires only one class posterior probability to describe the classification mapping, the multiclass problem needs all the posterior probabilities to describe the classification mapping. For instance, binary mapping needs only to evaluate p(ω 2|x) = 1 − p(ω 1|x) and the decision border can be described with p(ω 1|x) = p(ω 2|x) = 0.5. In turn, the multiclass problems need all posteriors {p(ωi |x)}ci =1 to describe the classification mapping and the decision borders are drawn between the class transitions. In equation (2.2) we suppose that every nearest neighbor contributes equally to calculate the class, independently from the distance to the query point. The wkNN is a kNN’s extension to handle this issue. The wkNN attributes weights for each voting pattern reducing the high dimensionality effects. In high-dimensional feature spaces, the training patterns become sparse requiring more data to fill out the decision region. The wkNN overcomes the kNN issues inhigh-dimensional spaces using a weighted scheme W(.,.) as an argument of a kernel function K(.). This is to adjust relative distances between patterns avoiding the sparsity2727 H. Dubey & V. Pudi. Class based weighted k-nearest neighbor over imbalance dataset, in Pacific-Asia Conference on Knowledge Discovery and Data Mining, Springer, 7819(2013), 305-316.. Furthermore, fractional metric modifications in kNN (or metric based classifiers) also can grant some level of reliability in high-dimensional feature spaces2828 C.C. Aggarwal, A. Hinneburg & D.A. Keim. On the surprising behavior of distance metrics in high dimensional space, in International Conference on Database Theory, Springer, (2001), 420-434.. The wkNN can be written as,

(2.7)

where the weighted scheme2727 H. Dubey & V. Pudi. Class based weighted k-nearest neighbor over imbalance dataset, in Pacific-Asia Conference on Knowledge Discovery and Data Mining, Springer, 7819(2013), 305-316. can be defined as follows,

2.3 SMOTE

Data imbalance shows to be an adverse setup to achieve high classification rates. This happens due to the rare or less frequent instances of minor represented classes (e.g. bank fraud, cancer malignancy grading)1212 B. Krawczyk. Learning from imbalanced data: open challenges and future directions. Progress in Artificial Intelligence, (2016), 1-12.. When the minority class has few training patterns it turns out to be misrepresented leading to shrunken decision regions. For instance, in Figure 5 the minor class was generated inside the circle with uniform distribution with 8 samples while the major class follows the distribution N([2,2],4I) with 100 samples generated from it.

Figure 5:
Decision region for an imbalanced dataset. The irregular solid line is the decision border between class A and B.

In order to avoid problems in classification mapping, one can randomly remove instances from the major classes until achieving the same prior probability (proportion). This technique is named the random undersampling. Instead to prune data one can rise the low priors of minor classes using an oversampling technique, in this case SMOTE, which creates synthetic patterns based on the existing ones2929 N.V. Chawla, K.W. Bowyer, L.O. Hall & W.P. Kegelmeyer. Smote: synthetic minority over-sampling technique. Journal of artificial intelligence research, 16(2002), 321-357.. Balance aid techniques developed for high-dimensional applications are also useful in situations with moderate imbalance as well1212 B. Krawczyk. Learning from imbalanced data: open challenges and future directions. Progress in Artificial Intelligence, (2016), 1-12.. In supervised learning, strategies to prevent imbalance are mainly focused in class reorganization or resampling owing to achieve the same number of training instances1111 H. He & E.A. Garcia. Learning from imbalanced data. IEEE Transactions on knowledge and data engineering, 21(9) (2009), 1263-1284.. Methods to aid the imbalanced problems are organized into the following categories1212 B. Krawczyk. Learning from imbalanced data: open challenges and future directions. Progress in Artificial Intelligence, (2016), 1-12.: data-level, algorithm-level or hybrid. We compare the effect of undersampling and oversampling techniques that are data-level methods since data is balanced disregarding the subsequent classifier.

SMOTE is similar to kNN and can be implemented as follows2929 N.V. Chawla, K.W. Bowyer, L.O. Hall & W.P. Kegelmeyer. Smote: synthetic minority over-sampling technique. Journal of artificial intelligence research, 16(2002), 321-357.: first, a pattern in the minor class is randomly selected, then, a synthetic pattern is included between a randomly chosen pattern in its k-neighborhood, repeated the procedure until achieving desired prior. SMOTE increases the parameter space that must be searched to obtain the model with highest prediction rate. In this work, we set SMOTE’s k-neighborhood parameter as k=5. Figure 6 depicts how the amount of oversampling changes the distribution for 50% and 100% of synthetic data added in AD class relative to MCI class, using k=5. The Bhattacharyya coefficient for 50% and 100% oversampling relative to the original distribution are 0.994579 and 0.9975564 respectively. Figure 7 depicts the effect of parameter k in SMOTE algorithm. For k=7, k=15 and k=30 the Bhattacharyya coefficient (assuming Gaussian distributions), from original sample is respectively, 0.9924127, 0.9972362 and 0.9935776, give that the value 1 means a complete distribution overlap.

Figure 6:
From left to right, the original data, synthetic oversampled 50% with original data, synthetic oversampled 100% with original data, respectively and their distributions.

Figure 7:
Original data with SMOTE oversampled 100% for values of parameter k, from left to right, k=7, k=15 and k=30.

2.4 Wrapper feature selection

The goal of feature selection methods is to select a subset of features that is useful to enhance a given classifier’s measure, e.g. precision. Since classifiers are induced by data with unknown underlying distributions the feature selection methods allow sub-optimal answers. There are comprehensive definitions of usefulness that would be criteria to select relevant features, e.g. correlation and information theoretical criterion1414 I. Guyon & A. Elisseeff. An introduction to variable and feature selection. Journal of machine learning research, 3(Mar) (2003), 1157-1182.. As shown in3030 R. Kohavi & G.H. John. Wrappers for feature subset selection. Artificial intelligence, 97(1) (1997), 273-324. the optimal choice of features does not imply the choice of relevant features. Conversely, optimality does not imply in relevancy. For instance, features that are presumably redundant may enhance the precision when combined with useful features1414 I. Guyon & A. Elisseeff. An introduction to variable and feature selection. Journal of machine learning research, 3(Mar) (2003), 1157-1182..

Despite the lack of guarantees presented, feature selection methods were invaluable to deal with high-dimensional real-world problems. Feature selection methods were initially designed to deal with classification problems with no more than 40 features3030 R. Kohavi & G.H. John. Wrappers for feature subset selection. Artificial intelligence, 97(1) (1997), 273-324., now they are able to deal with thousands of features. For instance, in classification problems related with genetics feature space dimensionality ranges from 6000 to 600001414 I. Guyon & A. Elisseeff. An introduction to variable and feature selection. Journal of machine learning research, 3(Mar) (2003), 1157-1182. one can expect a strong effect of the curse of dimensionality. Such extreme problems have received attention uncovering the molecular mechanisms related to AD3131 L. Scheubert, M. Luštrek, R. Schmidt, D. Repsilber & G. Fuellen. Tissue-based Alzheimer gene expression markers-comparison of multiple machine learning approaches and investigation of redundancy in small biomarker sets. BMC bioinformatics, 13(1) (2012), 266, pp 17. and have motivated initiatives like AlzGene focused on providing data resources for AD genetic research. Despite that, in this work, the feature selection techniques will be applied to at least 9 features in order to observe the group-wise probability of a feature being more relevant than other. Examples of feature selection techniques are feature extraction, feature construction, feature selection techniques for non-supervised learning, etc.

Feature selection methods are divided into three categories due to the relation with the classifier: filter, wrapper and embedded methods. Filters select a subset of features independently of the chosen classifier and the procedure mainly focus on ranking the features given defined criteria. Conversely, wrappers use classifiers’ measures as criteria to select subsets. Lately, embedded methods use a structured model to get the set of relevant features subject to a classifier1414 I. Guyon & A. Elisseeff. An introduction to variable and feature selection. Journal of machine learning research, 3(Mar) (2003), 1157-1182.. For a complete discussion on the feature selection strategies and benefits see1414 I. Guyon & A. Elisseeff. An introduction to variable and feature selection. Journal of machine learning research, 3(Mar) (2003), 1157-1182..

Here we compare three wrapper methods using the kNN classifier combined with SMOTE to select the most useful subset of features. The three wrapper methods compared are defined on the following search strategies: backward elimination, forward selection and hill-climbing selection1414 I. Guyon & A. Elisseeff. An introduction to variable and feature selection. Journal of machine learning research, 3(Mar) (2003), 1157-1182.. The subset obtained using the three methods will be compared to all combinations of features in order to observe if they are able to reach the optimum subset drawn from the rank with all feature combinations. Additionally, a noise feature will be included in the feature selection procedure in order to compare the features to a non-significant case. In Figure 8 we depict an example of search graph with all possibilities for three combinations. Regarding Figure 8 we have that in the far left stage, no features chosen and the far right all features chosen. Backward elimination moves right to left; forward selection left to right; hill climbing moves to any direction.

Figure 8:
Scheme showing how to choose a subset of 3 features, using forward selection, backward elimination, and hill-climbing.

The wrappers feature selection will search in a 9 (8 + noise) features graph scheme for the more useful subset. Forward selection initializes with any feature and steps up towards completing the feature subset. Iteratively it adds features to the chosen subset using the usefulness criteria that is to rise kNN classification rate. The usefulness criteria for kNN is obtained with the LOOCV, that means to classify one pattern using all remaining patterns as a training set. The backward elimination goes in the opposite direction in the searching graph. It initializes with all features and iteratively prunes features according to with the highest usefulness defined by the kNN classification rate criteria. The hill-climbing selection can go in any direction in the searching graph combining two previous approaches. Here we set up hill-climbing starting from the empty set feature. All wrappers in this work are greedy algorithms and are subject to be trapped in a local maximum3030 R. Kohavi & G.H. John. Wrappers for feature subset selection. Artificial intelligence, 97(1) (1997), 273-324. (relative to the classification rate). A greedy algorithm can only optimize in a short distance and do not prevent that a good choice for a given iteration can lead to losing better options. There are search strategies designed to avoid this greedy drawback, for instance, the simulated annealing and the genetic algorithms1414 I. Guyon & A. Elisseeff. An introduction to variable and feature selection. Journal of machine learning research, 3(Mar) (2003), 1157-1182..

2.5 Validation

Overfitting happens when the classifier’s prediction to the training phase are far better than the test phase66 C.M. Bishop. Pattern Recognition and Machine Learning (Information Science and Statistics). Springer, (2007).. The validation is appropriate to observe if the classifier is overfitted. An overfitted classifier cannot generalize results achieved in training phase for unseen patterns. Confusion matrix P is a tool to assess the classifier outcomes and to interpret classification precision. The confusion matrix shows the class wise probability of classifying x * in the class ωj given that it was generated by class ωi , for short Pi, j = p(x* ∈ωi |x* ∈ωj ) = p(ωi |ωj ). Furthermore, P is a stochastic matrix,

the P trace average is the probability of correct classifications for all classes or precision1818 J.S. Marques. Reconhecimento de Padrões: métodos estatistícos e neuronais. IST press (2005).. This measure defines a scalar magnitude that enables us to compare the many classifiers in the feature selection process. Furthermore, the 10-fold CV66 C.M. Bishop. Pattern Recognition and Machine Learning (Information Science and Statistics). Springer, (2007). is applied to the confusion matrices to obtain deviations of the feature’s subsets. The scalar for classifier’s comparisons can be written as,

3 RESULTS AND DISCUSSION

Features for specific and general cohort studies owing to identify AD spectrum come from various sources: cognitive, genetic, neuroimaging, proteomic and others1616 K. Lopez-de Ipiña, J.B. Alonso, J. Solé-Casals, N. Barroso, P. Henriquez, M. Faundez-Zanuy, C.M. Travieso, M. Ecay-Torres, P. Martinez-Lage & H. Eguiraun. On automatic diagnosis of Alzheimer’s disease based on spontaneous speech analysis and emotional temperature. Cognitive Computation, 7(1) (2015), 44-55.. These features intend to provide insights on biological factors AD which is critical to understanding the disease progression and to early prevention strategies11 M.S. Fiandaca, M.E. Mapstone, A.K. Cheema & H.J. Federoff. The critical need for defining preclinical biomarkers in Alzheimer’s disease. Alzheimer’s & Dementia, 10(3) (2014), S196-S212.. Feature combinations provide higher classification rates than features by itself, also there are combinations more precise than others. For instance, a feature may confer a poor classification rate when many classes are included, even being highly valuable to understand AD biological processes, as depicted in Figure 4. This is, the combination Aβ 1−42 and p-tau181, which sustain the main hypothesis for neurodegeneration11 M.S. Fiandaca, M.E. Mapstone, A.K. Cheema & H.J. Federoff. The critical need for defining preclinical biomarkers in Alzheimer’s disease. Alzheimer’s & Dementia, 10(3) (2014), S196-S212. achieves a precision of 42.29±1.32% and 43.40±4.61%, respectively for LOOCV and 10-fold CV for kNN with k optimized and imbalanced data. Thus, it is necessary to find additional features or combinations to better identify AD. However, for n features the number of combinations is given by 11 M.S. Fiandaca, M.E. Mapstone, A.K. Cheema & H.J. Federoff. The critical need for defining preclinical biomarkers in Alzheimer’s disease. Alzheimer’s & Dementia, 10(3) (2014), S196-S212.. We compare the three techniques of wrappers feature selection to the global rank of features for each sampling strategies using confusion matrices. Moreover, Gaussian noise (mean =0, sd =1) was added to feature space in order to compare an irrelevant feature to the features displayed in Table 1, with label N (noise). The Table 2 shows the test performance of sorted combinations by higher classification rate among the non-defining features and by the number of features., thus requiring strategies to avoid computational effort to uncover such combinations. Wrapper feature selection techniques are suitable to avoid the comparison of all features combinations while maximizing a chosen classifier’s precision. In case, the kNN that allows the all-versus-all strategy to observe how the classes affect each other all at once. Also, the all-versus-all strategy contrasts to the binary adapted strategies that are widely used in AD research along with ROC analysis

Table2:
Rank of combinations for the three imbalanced strategies. Note the highest classification rate was bolded for each validation method.

Excluding the defining features (3,5) most of the wrappers in this work were able to identify sub-optimal combinations given the precision of combinations available. For imbalanced dataset the combinations found are: 4,6 for hill climbing (position 2); 4,1,8 for backward elimination (position 28); 4,6,7,8 for forward selection (position 41). With random undersampling the combinations found are: 4,6 for hill climbing (position 8); 1,4,6,8 for backward elimination (position 2); 4,1 for forward selection (position 5). With oversampling (SMOTE) the combinations found are: 4,6,7,8,N for hill climbing (position 38); 1,2,4,6,7,8,N for backward elimination (position 9); 4,7,6,8 for forward selection (position 21). All strategies of sampling and wrapper feature selection found sub-optimal combinations relative to the rank position. For the complete ranking list see on-line contents. One can notice that even with hill-climbing which combines the forward and backward strategies it can be trapped in local maximum and be affect by the cross-validation components. For instance, using oversampling with backward elimination was found the position 9 and combined with hill-climbing the position 38 in the full ranking list.

The combination label 3 and 5 for the definition is depicted in Figure 9. From the selected features in training phase, the combination that provides the higher classification rate in validation phase is the (3,5). For more 2D plots see: https://github.com/yurier/TEMA-R-CODES/tree/master/PLOTS2D. The wrappers can be affected by the random nature of the cross-validation process and results may vary when the random generator number is unfixed. This variation ranges between the very first combinations to the middle-rank combinations. The list of 502 combinations for each technique to aid imbalance is available on-line, also the 120 combinations rank for non-defining features and the 26 combinations rank for non-cognitive features.

Figure 9:
Decision region for balanced (undersampled) three class problem solved with 5NN. From left to right, decision regions of CN, MCI and AD, respectively. Note that even for the defining features there is some errors.

Comparing the three techniques of sampling (imbalanced, undersampled and oversampled) we are able to say that there is a significant improvement using the imbalance aid techniques. The subtle dominance of kNN with SMOTE for the LOOCV happens due to the randomness effect of cross-validation and tie breaking. Furthermore, oversampled data is not pruned leaving more available data for training. In order to avoid overfitting, the synthetic data was used only to generate the classifier and to validate results. In Figure 10 is depicted the confusion matrices averaged in 10-fold cross-validation for the combinations ranked between all 502 combinations (on-line contents), respectively for each method and from 8th to 1st combinations. For instance, the first matrix in the first row at left in Figure 10 shows the probability of the MCI pattern be classified as CN is 35.6%. Also, in Figure 10 the confusion matrices for balanced problems (2nd and 3rd rows) show that the major class (MCI) have more classification errors than the minor classes (CN, AD). Conversely, in the imbalanced problem, the confusion matrices (1st row) for the major class are more precise and the minor class has more misclassification errors. This result is expected and can be used to control the compromise between class resampling in order to achieve a balance that minimizes the classification error.

Figure 10:
By row is depicted the confusion matrices for imbalanced, undersampled and oversampled combinations, respectively. Each column depicts from 8th to 1st ranked combinations in the 502 rank (on-line contents).

With the rank of all training results for all combinations of eight features plus noise producing 502 different combinations, one can ask the probability that a feature belonging to a given combination have higher classification rate than other. This is done by measuring how many times combinations that have feature A and not B are more precise than combinations that have B and not A divided by the number of comparisons between A and B. Figure 11 depicts this evaluation for each balance method (imbalanced, undersampling, oversampling) using only one significant digit by resolution reasons and just for the non-cognitive features (26 features). For instance, one can be interested in the probability of combinations that contain the feature 2 and not 8 to provide higher precision than combinations containing 8 and not 2, the left matrix in Figure 11 shows that is 38.7%. However, the matrices are not symmetrical due to some combinations have the same classification rate and due to round-off error. As argued by11 M.S. Fiandaca, M.E. Mapstone, A.K. Cheema & H.J. Federoff. The critical need for defining preclinical biomarkers in Alzheimer’s disease. Alzheimer’s & Dementia, 10(3) (2014), S196-S212. the neuropsychological tests are more precise and standardized measures to detect AD. Figure 11 shows only non-cognitive features (1,2,7,8). Despite neuropsychological tests being cost-effective biomarkers and its combinations provide a high precision, they do not provide information on the biological mechanism of AD. The suggestion of how they provide a higher precision is due to the limited possibilities of outcomes that define the neuropsychological scores, leading to overlaying patterns, as represented in Figure 9. This is, biomarkers that have fractional values are less subject to becoming an integer value grid in comparison to neuropsychological tests. In the right matrix of Figure 11 the noise has a higher probability in average to increase the classification rate than proteomic biomarker 7 however it does not mean irrelevancy1414 I. Guyon & A. Elisseeff. An introduction to variable and feature selection. Journal of machine learning research, 3(Mar) (2003), 1157-1182.. As argued by1414 I. Guyon & A. Elisseeff. An introduction to variable and feature selection. Journal of machine learning research, 3(Mar) (2003), 1157-1182., a feature that is supposed to be irrelevant could contribute to enhancing the classifier performance.

Figure 11:
From right to left, probability matrix of a feature to provide higher classification rate than others for the imbalance, undersampled and oversampled, ranks respectively. Label “A” stands for column-wise and row-wise average.

The SVM classifier that uses binary adapted strategy when applied to the same multiclass classification problem achieves 92.39±4.96% in 10 fold cross-validation for features 3,5, while kNN with SMOTE achieves 94.34±3.91%. Since SVM is not a distance based classifier as kNN there is no need for feature normalization. Using all features plus noise SVM achieves a higher precision of 89.56±6.01% when compared to kNN with SMOTE that achieves 80.84±6.96%. This precision difference between kNN and SVM for the larger feature combination is due to the metric behavior when the dimensionality is increased2828 C.C. Aggarwal, A. Hinneburg & D.A. Keim. On the surprising behavior of distance metrics in high dimensional space, in International Conference on Database Theory, Springer, (2001), 420-434.. The wKNN in equation (2.7) was idealized to prevent the dimensionality effect by means of assuming that not all k nearest neighbors would contribute equally to define a class assignment. The wkNN with triangular kernel function, optimized k and L 1 distance metric produces a precision of 81.03±2.12% for 10-fold cross-validation using all features combination plus noise and imbalanced classes.Using same validation process the precision for the combination 3,5 reaches 92.20±1.72% for wkNN. Despite the curse of dimensionality that affects kNN more than SVM and wkNN, the kNN with SMOTE achieves higher precision in average compared to these methods without balance aids. For the canonical biomarkers combination, Aβ 1−42 and p-tau181, the SVM reaches 33.34±0.68% and the wkNN reaches 37.36±3.02%.

4 CONCLUSION

Wrappers techniques for feature selection have shown to be efficient to find the suboptimalcombinations given by the rank for all imbalance aid strategies for the proposed problem. However, adding more features to test limits of greedy search could be less successful. Including features is challenging because it increases the number of patients who did not undergo to all examinations. This can be seen in the complete dataset available in ADNI. Fortunately, kNN inspired data missing techniques are available and would be useful to identify more precise combinations that include interpretation benefits for AD mechanisms. However, to deal with improvements to kNN which imply in non-convex optimizations will require more studies. There are options to increase kNN performance such as metric learning2828 C.C. Aggarwal, A. Hinneburg & D.A. Keim. On the surprising behavior of distance metrics in high dimensional space, in International Conference on Database Theory, Springer, (2001), 420-434. and normalization strategies2121 C.-M. Ma, W.-S. Yang & B.-W. Cheng. How the parameters of k-nearest neighbor algorithm impact on the best classification accuracy: In case of parkinson dataset. Journal of Applied Sciences, 14(2) (2014), 171-176.. However, this will increase the parameter space to optimize and improvements in the computation performance. Even with the curse of dimensionality, the kNN with SMOTE overcome in average SVM and wKNN in reproducing the definition. However, the benefits of data balance would increment any classifier precision. The performed all-versus-all strategy requires less classifiers to be built than binary strategies. The class defining features (labels 3,5) increase artificially the class combinations for rank with 502 possibilities, however, it was useful to obtain a comparative. Comparisons using non-cognitive features reveal that FDG contributes more to increase the classification rate. However, more non-cognitive features are needed to observe if dominance for FDG holds, this is a challenge given mentioned data issues for missing values and data imbalance.

ACKNOWLEDGMENTS

Data collection and sharing for this project was funded by the ADNI (National Institutes of Health Grant U01 AG024904) and DOD ADNI (Department of Defense award number W81XWH-12-2-0012). The author thanks CAPES for the master’s scholarship provided.

REFERENCES

  • 1
    M.S. Fiandaca, M.E. Mapstone, A.K. Cheema & H.J. Federoff. The critical need for defining preclinical biomarkers in Alzheimer’s disease. Alzheimer’s & Dementia, 10(3) (2014), S196-S212.
  • 2
    C.R. Jack, D.S. Knopman, W.J. Jagust, L.M. Shaw, P.S. Aisen, M.W. Weiner, R.C. Petersen & J.Q. Trojanowski. Hypothetical model of dynamic biomarkers of the Alzheimer’s pathological cascade. The Lancet Neurology, 9(1) (2010), 119-128.
  • 3
    R.A. Sperling, P.S. Aisen, L.A. Beckett, D.A. Bennett, S. Craft, A.M. Fagan, T. Iwatsubo, C.R. Jack, J. Kaye & T.J. Montine et al. Toward defining the preclinical stages of Alzheimer’s disease: Recommendations from the national institute on aging-Alzheimer’s association workgroups on diagnostic guidelines for Alzheimer’s disease. Alzheimer’s & dementia, 7(3) (2011), 280-292.
  • 4
    A.A. Motsinger-Reif, H. Zhu, M.A. Kling, W. Matson, S. Sharma, O. Fiehn, D.M. Reif, D.H. Appleby, P.M. Doraiswamy & J.Q. Trojanowski et al. Comparing metabolomic and pathologic biomarkers alone and in combination for discriminating Alzheimer’s disease from normal cognitive aging. Acta neuropathologica communications, 1(1) (2013), pp. 28.
  • 5
    S.J. Teipel, O. Sabri, M. Grothe, H. Barthel, D. Prvulovic, K. Buerger, A.L. Bokde, M. Ewers, W. Hoffmann & H. Hampel. Perspectives for multimodal neurochemical and imaging biomarkers in Alzheimer’s disease. Journal of Alzheimer’s Disease, 33(s1) (2013), S329-S347.
  • 6
    C.M. Bishop. Pattern Recognition and Machine Learning (Information Science and Statistics). Springer, (2007).
  • 7
    T. Fawcett. An introduction to roc analysis. Pattern recognition letters, 27(8) (2006), 861-874.
  • 8
    L. Khedher, J. Ramírez, J.M. Górriz, A. Brahim, F. Segovia & A.D.N. Initiative et al. Early diagnosis of Alzheimer’s disease based on partial least squares, principal component analysis and support vector machine using segmented mri images. Neurocomputing, 151(2015), 139-150.
  • 9
    A. Khazaee, A. Ebrahimzadeh & A. Babajani-Feremi. Identifying patients with Alzheimer’s disease using resting-state fmri and graph theory. Clinical Neurophysiology, 126(11) (2015), 2132-2141.
  • 10
    Q. Yang & X. Wu. 10 challenging problems in data mining research. International Journal of Information Technology & Decision Making, 5(04) (2006), 597-604.
  • 11
    H. He & E.A. Garcia. Learning from imbalanced data. IEEE Transactions on knowledge and data engineering, 21(9) (2009), 1263-1284.
  • 12
    B. Krawczyk. Learning from imbalanced data: open challenges and future directions. Progress in Artificial Intelligence, (2016), 1-12.
  • 13
    C. Humpel. Identifying and validating biomarkers for Alzheimer’s disease. Trends in biotechnology, 29(1) (2011), 26-32.
  • 14
    I. Guyon & A. Elisseeff. An introduction to variable and feature selection. Journal of machine learning research, 3(Mar) (2003), 1157-1182.
  • 15
    Y. Saeys, I. Inza & P. Larrañaga. A review of feature selection techniques in bioinformatics. Bioinformatics, 23(19) (2007), 2507-2517.
  • 16
    K. Lopez-de Ipiña, J.B. Alonso, J. Solé-Casals, N. Barroso, P. Henriquez, M. Faundez-Zanuy, C.M. Travieso, M. Ecay-Torres, P. Martinez-Lage & H. Eguiraun. On automatic diagnosis of Alzheimer’s disease based on spontaneous speech analysis and emotional temperature. Cognitive Computation, 7(1) (2015), 44-55.
  • 17
    A. Sarica, G. Di Fatta, G. Smith, M. Cannataro & J.D. Saddy et al. Advanced feature selection in multinominal dementia classification from structural mri data, in Proc MICCAI Workshop Challenge on Computer-Aided Diagnosis of Dementia Based on Structural MRI Data, pp. 82-91 (2014).
  • 18
    J.S. Marques. Reconhecimento de Padrões: métodos estatistícos e neuronais. IST press (2005).
  • 19
    T. Tapiola, I. Alafuzoff, S.-K. Herukka, L. Parkkinen, P. Hartikainen, H. Soininen & T. Pirttilä. Cerebrospinal fluid β-amyloid 42 and tau proteins as biomarkers of Alzheimer-type pathologic changes in the brain. Archives of neurology, 66(3) (2009), 382-389.
  • 20
    A.W. Toga & K.L. Crawford. The Alzheimer’s disease neuroimaging initiative informatics core: A decade in review. Alzheimer’s & Dementia, 11(7) (2015), 832-839.
  • 21
    C.-M. Ma, W.-S. Yang & B.-W. Cheng. How the parameters of k-nearest neighbor algorithm impact on the best classification accuracy: In case of parkinson dataset. Journal of Applied Sciences, 14(2) (2014), 171-176.
  • 22
    A. Bhattacharyya. On a measure of divergence between two multinomial populations. Sankhyā: the indian journal of statistics, 7(4) (1946), 401-406.
  • 23
    L. Devroye, L. Györfi & G. Lugosi. A probabilistic theory of pattern recognition, 31. Springer Science & Business Media, 31(2013), 638 pages.
  • 24
    G. Bhattacharya, K. Ghosh & A.S. Chowdhury. An affinity-based new local distance function and similarity measure for knn algorithm. Pattern Recognition Letters, 33(3) (2012), 356-363.
  • 25
    T.M. Cover & P.E. Hart. Nearest neighbor pattern classification. Information Theory, IEEE Transactions on, 13(1) (1967), 21-27.
  • 26
    T. Bailey & A.K. Jain. A Note on Distance-Weighted k-Nearest Neighbor Rules. IEEE Transactions on Systems, Man, and Cybernetics, SMC-8(4) (1978), 311-312.
  • 27
    H. Dubey & V. Pudi. Class based weighted k-nearest neighbor over imbalance dataset, in Pacific-Asia Conference on Knowledge Discovery and Data Mining, Springer, 7819(2013), 305-316.
  • 28
    C.C. Aggarwal, A. Hinneburg & D.A. Keim. On the surprising behavior of distance metrics in high dimensional space, in International Conference on Database Theory, Springer, (2001), 420-434.
  • 29
    N.V. Chawla, K.W. Bowyer, L.O. Hall & W.P. Kegelmeyer. Smote: synthetic minority over-sampling technique. Journal of artificial intelligence research, 16(2002), 321-357.
  • 30
    R. Kohavi & G.H. John. Wrappers for feature subset selection. Artificial intelligence, 97(1) (1997), 273-324.
  • 31
    L. Scheubert, M. Luštrek, R. Schmidt, D. Repsilber & G. Fuellen. Tissue-based Alzheimer gene expression markers-comparison of multiple machine learning approaches and investigation of redundancy in small biomarker sets. BMC bioinformatics, 13(1) (2012), 266, pp 17.
  • Work presented at Congress of Applied and ComputationalMathematics - CMAC SE 2015.
  • **
    Data used in the preparation of this article were obtained from the ADNI database. As such, the investigators within the ADNI contributed to the provided data but did not participate in analysis or writing of this report.

Publication Dates

  • Publication in this collection
    Jan-Apr 2017

History

  • Received
    20 Dec 2015
  • Accepted
    07 Feb 2017
Sociedade Brasileira de Matemática Aplicada e Computacional Rua Maestro João Seppe, nº. 900, 16º. andar - Sala 163 , 13561-120 São Carlos - SP, Tel. / Fax: (55 16) 3412-9752 - São Carlos - SP - Brazil
E-mail: sbmac@sbmac.org.br