CONSIDERATIONS REGARDING THE CHOICE OF RANKING MULTIPLE CRITERIA DECISION MAKING METHODS

Various methods, known as Multiple Criteria Decision Making Methods (MCDM), have been proposed to assist decision makers in the process of ranking alternatives. Given the variability of available methods, choosing an MCDM ranking method is a difficult task. There are key factors in the process of choosing an MCDM method such as: (i) available time; (ii) effort required by a given approach; (iii) importance of accuracy; (iv) transparency necessity; (v) conflict minimization necessity; and (vi) facilitator’s skill with the method. However, the problem is further increased by the knowledge that the solution of MCDM ranking methods may be sensitive to slight variations in entrance data and, in some cases, might replace the best alternative for the worst when the weightings for the criteria are changed. Some researchers have addressed this problem through different approaches, including the evaluation of MCDM ranking methods in the sense of predicting the initial rankings given by the decision maker. The objective of this study is to propose an empirical experiment to evaluate the propensity for initial ranking prediction of the principal MCDM ranking methods, namely: SAW, TOPSIS, ELECTRE III, PROMETHEE II and TODIM. The study also aimed to assess possible common ranking problems in MCDM methods, such as reversibility, found in the literature. It was found that just up to 20% of initial ranking order was predicted entirely correct by some of the methods. It was also found that just a few methods did not present internal ranking inconsistency. The results of this study and those found in the literature give a warning regarding the choice of MCDM ranking methods. It is suggested that special care must be taken in the choice of methods and, besides axiomatic comparisons, ranking comparisons could be a useful way to enhance the decision making process, since MCDM methods are tools for learning about the problem and do not prescribe solutions.

However, given the variability of available methods, choosing an MCDM ranking method is a difficult task.Almeida (2013) asserts that there are key factors in the process of choosing an MCDM method, such as: (i) available time; (ii) effort required by a given approach; (iii) importance of accuracy; (iv) transparency necessity; (v) conflict minimization necessity; and (vi) facilitator's skill with the method.Almeida (2013) complements by saying that the choice of an MCDM method should be aligned with the preference structure and the rationality assumption of the decision makers.The problem is further increased by the knowledge that solution of MCDM ranking methods may be sensitive to slight variations in entrance data, e.g.small changes in the weighting vector based on the decision maker's preferences, or to the computational algorithm employed (Yoon & Hwang, 1995;Buede & Maxwell, 1995;Zanakis et al., 1998;Yeh, 2002;Maxwell, 1995).There is also the fact that MCDM ranking methods, in some cases, might replace the best alternative for the worst when the weightings for the criteria are changed (Tallarico, 1990; Wang & Triantaphyllou, 2008;Brunner & Starkl, 2004).Some researchers addressed this problem through different approaches.For example, Mareschal (1988) defines stability intervals for the weightings of the different criteria in order to study the stability of the results generated by the PROMETHEE method.Buede & Maxwell (1995) used the Monte Carlo approach to study MCDM inconsistency in the Analytic Hierarchy Process (AHP) and TOPSIS methods.Zanakis et al. (1998) verified the ranking inconsistency of ELEC-TRE, TOPSIS, SAW, and four versions of AHP, also using the Monte Carlo approach based on twelve consistency measures.Yeh (2002) performed sensitivity analysis and proposed a measure for the degree of consistency, based on Shannon's entropy concepts, for evaluating the SAW and TOPSIS methods.Moshkovich et al. (2012) evaluated the stability of the results obtained through the SAW and TODIM methods.Gomes & Costa (2015) used a set of methods, including ELECTRE I, II and PROMETHEE II, in order to evaluate the differences between the rankings generated by these different methods in the choice of an electronic payment system.In turn, Yoon & Hwang (1995, p.68) suggested that MCDM ranking methods should be evaluated in the sense of predicting the initial rankings given by the decision maker, that is "how well a method predicts unaided decisions made independently of the judgments used to fit the model".
In this sense, the aim of this study was to propose an empirical experiment to evaluate the propensity for initial ranking predicting of the principal MCDM ranking methods, namely: SAW, TOPSIS, ELECTRE III, PROMETHEE II and TODIM.The study also aimed to assess possible common ranking problems in MCDM methods, such as reversibility, found in the literature.For this purpose, a multicriteria decision problem regarding the choice of a travel destination was applied to an experimental group using SANNA1 software (Jablonský, 2009) and with the help of a spreadsheet for the calculations of the TODIM method.In total, 20 undergraduate students from the Ribeirão Preto School of Economics, Business Administration and Accounting, participated in the experiment in order to verify the adherence of the rankings proposed by the methods with the participants' initial rankings suggestion.
As with other studies in the literature, the present study does not intend to make an axiomatic, numerical or deterministic comparison between the methods, but to make considerations regarding the choice of MCDM ranking methods by applying and evaluating different methods.The study is justified because none of the related studies in the literature carried out comparisons among all of the most notable MCDM ranking methods at the same time.

THEORETICAL FRAMEWORK
In the context of solving multiple criteria problems, the possibility of different rankings by the application of different MCDM ranking methods occurring is a factor that should be taken into consideration by the users who opt for this kind of approach.Buede & Maxwell (1995) have already identified problems related to final ranking inconsistency in some well-known multiple criteria methods.According to the authors, although these methods have been developed based on a different number of theories and algorithms, the decision is always made considering the preferences on a set of weighting criteria (Buede & Maxwell, 1995).
In their research, Buede & Maxwell (1995) chose the AHP and TOPSIS ranking methods, among others, to verify whether there were differences in the results when using Monte Carlo experiments.According to the authors, the selected methods have two common characteristics: (i) they require the decision maker to give a weighting to a set of data; and (ii) they produce alternative rankings indicating the best among them (Buede & Maxwell, 1995).The authors hypothesized that there would be a risk in the misapplication of these algorithms, given that the literature mentions ranking problems in methods such as AHP.The authors conducted a series of simulation experiments that allowed them to compare the best alternative indicated by each of the algorithms.Experiments have shown that the AHP method often does not present ranking disagreement with other compared methods, which was not the case of TOPSIS.The situations in which the most significant differences occurred were often associated with this last method (Buede & Maxwell, 1995).Mareschal (1998) states that the problem of assessing the relative importance of different criteria is commonly performed by sensitivity analysis.Alternatively, the author proposed stability intervals for the weights of the different criteria in additive methods, such as PROMETHEE II, for improving the technique of sensitivity analysis, reducing the time of this procedure.According to Marechal (1998, p. 54), such stability intervals are "values that the weighting of one criterion can take without altering the results given by the initial set of weightings, all other weightings being kept constant".Marechal (1998) studied the sensitivity of the results inducing variations of the strictly positive weightings normalized to one.The author proposed changes to the total weighting while the relative importance of other criteria were kept constant.Finally, the use of the method in a didactic example using the PROMETHEE II outranking method where the criteria intervals were calculated (Mareschal, 1998) was presented.Mareschal (1998) affirmed that the intervals provide a deeper knowledge of the decision problem to the users, which can prevent them from ranking changes.
In turn, Zanakis et al. (1998) noticed that ranking methods might produce different rankings even when applied to the same problem, apparently under the same conditions.According to the authors, this inconsistency occurs because: (i) each method uses different weighting calculations; (ii) the algorithms differ in their approach to selecting the best solution; and (iii) some algorithms introduce additional parameters that affect the chosen solution.Moreover, this situation can be intensified by differences in weighting extraction among different decision makers, even with similar preferences (Zanakis et al., 1998).
Therefore, according to Zanakis et al. (1998), the wide variety of available methods with complexity and varying solutions might confuse potential users.Thus, according to the authors, the decision maker must first face the task of selecting the most suitable method among the many possible alternatives.According to Zanakis et al. (1998), users could compare these methods considering different dimensions, such as simplicity, reliability, robustness and quality.An extensive literature review carried out by Zanakis et al. (1998) revealed that only a limited number of studies were devoted to comparing different methods.Zanakis et al. (1998) subsequently concluded that it is very difficult to answer questions such as "which method is the most appropriate for a specific type of problem and what are the advantages and disadvantages of using one method rather than another?"Based on a decision matrix with "n" weighted criteria and "m" alternatives, Zanakis et al. (1998) proposed a method to compare by means of simulations, ranking methods.In their study eight ranking methods were compared using twelve similarity measures of performance via parametric ANOVA and nonparametric Kruskal-Wallis tests.The methods chosen were: ELECTRE, TOPSIS, SAW, the Multiplicative Exponential Weighting (MEW), and four versions of AHP.Similarities and differences in the solution of the methods were investigated.The simulation parameters were the number of alternatives and criteria (Zanakis et al. 1998).Zanakis et al. (1998) found that ranking differences derive from the process of weighting the criteria, and become even more pronounced in problems with few alternatives, even though most importantly, the final ranking of the alternatives varies more in problems with many alternatives.In general, all AHP versions behave similarly and closer to SAW than the other methods.ELECTRE is the least similar to SAW, followed by MEW.TOPSIS behaves more like AHP, and more differently to ELECTRE and MEW, except for problems with few criteria.The number of criteria had little effect on AHP, ELECTRE, or MEW while TOPSIS ranking becomes different when the number of criteria increases (Zanakis et al., 1998).Based on the results obtained, the authors argue that these methods should help the user learn more about the problem and possible solutions for reaching a final decision, and thus do not advocate the use of MCDM for a prescriptive solution (Zanakis et al., 1998).
As reported by Yeh (2002), there is no best method for multiple criteria decision problems, and the validity of the ranking results remains an open question.In some particular cases, the solutions produced by different MCDMs are the same.However, in situations where the decision ranking of all alternatives is necessary, the author states that it is important to take into account that different methods produce different results for the same problem.In other words, for the same weighting vector the ranking order may vary depending on the method used, and this mismatch increases as the number of alternatives increases (Yeh, 2002).Consequently, choosing a method from a variety of MCDM methods has also become a multicriterial problem (Yeh, 2002).
Based on his postulation, Yeh (2002) proposed a new approach to the selection of MCDM methods via sensitivity analysis of attribute weightings, seeking to determine to what degree the ranking of the alternatives provided by the evaluated methods could vary when changes occur in the weightings of the criteria.Yeh (2002) also used the concept of Shannon's entropy to propose a measure for the degree of consistency of the rankings.Three methods were selected to be applied to a case study in which a college needed to select students to be awarded a scholarship.The methods chosen were SAW, MEW and TOPSIS.In the case study developed, the most suitable method identified was TOPSIS.According to the author, the proposed approach is particularly useful for large-scale problems where the ranking produced by different methods differ significantly (Yeh, 2002).Wang & Triantaphyllou (2008) also stated that different multiple criteria methods suffer from the disadvantage of providing different answers to exactly the same problem.According to Wang & Triantaphyllou (2008), some of the MCDM methods use the sum of the priorities of the alternatives, such as the AHP method (American school), while others use classification relationships, such as the ELECTRE method and its derivatives (French school).Irregularities in AHP ranking have been reported by many researchers and for the first time the authors also point out these irregularities for ELECTRE (Wang & Triantaphyllou, 2008).
According to Wang & Triantaphyllou (2008), irregularities in the ranking of alternatives occur when the MCDM method does not meet the following requirements: (i) maintaining the indication of the best alternative even when one of the alternatives is replaced by another worse alternative and the weightings determined for the criteria remain the same; (ii) obeying the property of transitivity for the final ranking of alternatives; (iii) providing the same ranking as for the original problem when the decision problem is divided into parts (Wang & Triantaphyllou, 2008).Wang & Triantaphyllou (2008) sought to identify why the above contradictions occur in the ELECTRE method, and essentially to explain why, when an alternative that is one of the worst is replaced by another worse alternative, the indication of the best alternative can be changed.In order to verify ranking irregularities within the ELECTRE II and III methods, computer programs written in MATLAB (computer language for high-level developments of algorithms and data visualization) were developed, to generate simulated decision problems and test the performance of ELECTRE II and III evaluating the three requirements listed.As a result, it was found that the best alternative remained the same for both methods, but there was a significant difference in the ranking of other alternatives (Wang & Triantaphyllou, 2008) 2 .Moshkovich et al. (2012, p. 523) affirmed that "multiple criteria decision aiding techniques are used to construct an aggregation model on the basis of preference information provided by the decision maker".The authors analyzed the differences in the implementation and the stability of the results obtained by TODIM and SAW through direct preferential information.Moshkovich et al. ( 2012) compared a set of 15 residential properties available for rent in the city of Volta Redonda in Brazil, against eight criteria using both aggregation methods.The authors concluded that it is difficult to select an appropriate multiple criteria ranking method because "criterion weightings and scale transformations for criterion values produced significant differences in the ranking of alternatives when two different methods are used for the aggregation of the preferential information" (Moshkovich et al., 2012, p. 538).The SAW method produced a significant different ranking when compared to the ranking produced by the TODIM method.
Gomes & Costa (2015) also studied why there is difference between the results when different MCDM ranking methods are applied to the same problem.In this sense, the objective of their research was to map the possible differences among the rankings provided by the application of THOR, ELECTRE I and II, and PROMETH ÉE II to the problem of choosing three different kinds of electronic payment by credit card (fourteen criteria were considered).Based on the results obtained by the application of the four different multicriteria methods and the use of sensitivity analysis, the authors claim that a decision maker could enhance his decision process with greater knowledge of the problem by considering a different method solution.

RESEARCH METHOD
In order to propose an empirical experiment to evaluate the propensity for initial ranking predicting of the principal MCDM ranking methods, namely: SAW, TOPSIS, ELECTRE III, PROMETHEE II and TODIM, twenty students from different years and courses from Ribeirão Preto School of Economic, Management and Accounting, were invited to evaluate a decision matrix containing five travel destinations (alternatives A to E) performed with eight different criteria, namely: (i) Hotel rating, which is the rating for a hotel (ranging from 1, worst to 5, the best); (ii) Time distance -distance in hours from the destination; (iii) Day length -the amount of days in the travel; (iv) Cost -the price of accommodation and flight ticket (in US$); (v) Shoppingthe index for the amount and diversity of shopping places (ranging from 1, worst to 10, the best); is not to discover absolute truth or, in this case, a pre-existing 'real' ranking".Therefore, in the perspective of the authors, ranking stability in the application of MCDM ranking methods "is not necessarily the evidence of an adequate processing of data".Here, the numerical approach of the Wang & Triantaphyllou (2008)  (vi) Cultural attractions -the index for the amount and diversity of cultural attractions (ranging from 1, worst to 10, the best); (vii) Natural landscape -the index for the presence of natural landscape (ranging from 1, worst to 10, the best); and (viii) Safety -whether it is safe in terms of health conditions, violence or terrorism (ranging from 1, most unsafe to 10, the safest).Table 1 presents the decision matrix whose performance criteria was derived from Brazilian travel agency packages (Leoneti et al., 2015).The participants were required to evaluate the matrix and to rank all alternatives, which was stored as their respective initial ranking.The participants were also required to rank all criteria from which the respective weighting vectors were calculated using the eliciting Rank Ordered Centroid (ROC) method.The decision matrix was processed by SANNA software (Jablonský, 2009) and for each participant the respective weighting vector was inserted manually in order to calculate the ranking generated by the SAW, TOPSIS, ELECTRE III and PROMETHEE II methods.The TODIM method was modeled using the Visual Basic Application in an Excel spreadsheet and was used to calculate the ranking related to this approach.
Subsequently, the output data from SANNA (SAW, TOPSIS, ELECTRE III and PROMETHEE II methods) and from the spreadsheet (TODIM method) were analyzed through descriptive statistics.Firstly, the rank-order correlation coefficient r s of Spearman, which is considered a measure for the association between the rankings of N objects generated by two observers (Siegel & Castellan JR., 1988), was applied to measure the correspondence between the initial ranking (defined by the participants) and the ranking calculated by each of the five MCDM ranking methods.The value of r s ranges from -1 to 1 and is compared with tabulated values for a two-sided or one-sided test.The decision criterion is to reject H 0 when r s is greater than the critical value, meaning that the ranking varies similarly.Equation 1 presents the calculation for the r s Spearman coefficient.
where, N = number of objects, Here the N objects are the five alternatives under evaluation.According to Siegel & Castellan Jr.
(1988), a high r s value indicates that the two rankings are associated (proportionately when the value is close to 1 and inversely, when the value is close to -1).
Secondly, the output rankings from all MCDM methods were compared with each other for all participants.In order to verify the degree of similarity between rankings of N objects generated by k observers or judges (for k greater than 2), the W Kendall coefficient concordance can be used as a measure of dependence between the rankings (Siegel & Castellan JR., 1988).A high value of W can be interpreted as the degree to which k observers or judges ranked N objects similarly.The value of W ranges from 0 to 1 and is compared with tabulated values for a onesided test.The decision criterion is to reject H 0 when W is closer to one, meaning the rankings are dependent.Equation 2 presents the calculation of the W Kendall coefficient.
where, k = number of ranking sets, N = number of objects, R i = average position of the criterion i from each ranking.
Here the k "judges" are the five MCDM methods, each one generating a ranking over the N objects, which are the five alternatives being evaluated.The W Kendall significance can be determined based on the probability associated with each occurrence.According to Siegel & Castellan Jr. (1988), for N and k equal to five, H 0 is rejected when W is greater than 0.571 for the significance level of 1%.
Thirdly, the participants were questioned regarding the comparison between their initial ranking and that provided by the methods.A question, namely "Do you regret your initial ranking of alternatives?" was responded by the participants at the end of the application using a Likerttype scale with seven levels (one, strongly disagree, and seven, strongly agree).The output data were then evaluated using descriptive statistics in graphic form and then compared to ranking problems identified in the literature.Additionally, the participant with the highest score for this question among those with the highest value for W Kendall significance was selected for sensitivity analysis using their initial weighting vector.The weighting vector of the chosen participant suffered variations of 20% in the value of a single criterion, selected randomly obeying a uniform distribution.There were 100 iterations and in each iteration, due to the distortions caused by the criterion that suffered the variation, the relationships with the other criteria were recalculated using Equation 3 according Ensslin et al. (2001). where, Pesquisa Operacional, Vol.36( 2), 2016 w i = original weighting for criterion i, w * i = ten percent changed weighting for criterion i, w n = original weighting for criterion n, w * n = recalculated weighting for criterion n.Finally, with the new weighting vector generated in each of the 100 iterations, the macro called all classes of SAW, TOPSIS, ELECTRE III and PROMETHEE II algorithms from SANNA and stored the output in a separate spreadsheet.An additional step was performed simultaneously in order to calculate the results from the TODIM method, the output also being stored.The flowchart depicted in Figure 1 summarizes all steps included in the Excel macro for sensitivity analysis.

Decision matrix set up and results generation
SANNA contains a unique decision matrix as input for the various methods available.It is possible to define the kind of criteria of each alternative that will be evaluated (detriment or benefit criteria).Considering that, before inserting the criteria performance for each alternative into the decision matrix, the criteria "Cost" and "Time distance" (originally detriments criteria) had to be transformed into benefits criteria by the calculation of "Cost" −1 and "Time distance" −1 .
In other words, it was decided that for all criteria the alternative should be classified as the larger the criteria values the higher the ranking order in the ranking.Figure 2 shows the decision matrix settled into the SANNA considering all criteria as benefit criteria.Subsequently, each weighting vector calculated by the ROC method for each participant (Table 2) was manually inserted into the SANNA and for each one all classes of SAW, TOPSIS, ELECTRE III and PROMETHEE II algorithms were run.In order to calculate the results from ELECTRE III the criteria were all considered as true-criteria (without pseudo-criteria), hence the parameters veto, indifference, and preference were set to zero.In order to calculate the results from PROMETHEE II the value for preference (p) to the function of linear preference ( ), as shown in Equation 4and Figure 3, was settled at 0.6 for each criterion k.At the end, the outputs were stored in a separate spreadsheet.The same procedure was carried out with the Excel spreadsheet containing the TODIM method, in this case using the parameter θ (the attenuation factor of the losses) settled at 10.

Accuracy and ranking similarity evaluation
In order to measure the correspondence (accuracy) between the initial ranking defined by the participants and the ranking calculated by SAW, TOPSIS, ELECTRE III, PROMETHEE II and  TODIM, the value of r s Spearman coefficient was calculated.Figure 4 shows the average of the r s Spearman coefficient calculated for the twenty participants.
In terms of accuracy (correlation with the initial ranking), ELECTRE III had on average the best performance in this application since the closer r s Spearman coefficient is to one, the more similar the rankings are.In fact, ELECTRE III predicted the ranking correctly at least for 20% of the cases.On the other hand, when considering the number of correct rank-order matches the 0 0,1 best scores were achieved by the TOPSIS method 3 .TOPSIS accuracy reached 50% of at least three matches in rank-orders for all twenty participants, followed by PROMETHEE II (40%), ELECTRE II (30%), and TODIM and SAW, both with 25%.It should be noted however that the number for ranking prediction (five correct matches) for TOPSIS was the same as ELECTRE III, both with 20%, followed by SAW and PROMETHEE II with 10% each and TODIM with 0%.This result can be explained considering that TOPSIS, according to Buede and Maxwell (1995), is subject to larger alteration of the rank-order including ranking reversal problems.In this sense, when a modification of rank-order occurs it tends to be very distinct.Finally, the prediction regarding the first alternative (best alternative) in the ranking was also checked.In this criterion, TOPSIS again had good performance.Together with PROMETHEE II, TOPSIS correctly predicted 15 times the first alternative ranking, which means 79% of accuracy.Both methods are followed by SAW (74%), ELECTRE III (57%) and TODIM (37%).Likewise to the case presented by Yeh (2002), TOPSIS performance attracted attention in terms of accuracy.
In terms of ranking similarity, which is how much ranking order varies according to the method used, the output rankings were tested using the W Kendall coefficient concordance.The hypothesis tested whether the ranking varies depending on the method used (independence condition).The calculated W Kendall coefficient is compared with the critical value of 0.571 and the null hypothesis of independency is rejected if the coefficient is larger than the critical value.In this application, just for eight cases the null hypothesis was rejected, meaning that the decision maker obtained significantly different rankings with the application of the five methods for 55% of the cases.This result is in accordance with Wang & Triantaphyllou (2008), Yeh (2002), and Zanakis et al. (1998) that for the same weighting vector, the ranking order may vary depending on the method used.Finally, the correlation between each method using the value of r s Spearman coefficient was evaluated.The results are shown in Table 3. that ELECTRE is one of the methods least similar to SAW, and that TOPSIS behaves more differently to ELECTRE.Considering the approach that belongs to each of the evaluated methods (SAW and TOPSIS from the utility based approach -American school, and ELECTRE III and PROMETHEE II from the outranking methods approach -French school), this result would be indeed expected.However, it is worth noting that the highest correlation value occurred between SAW and PROMETHEE II, followed by the correlation between TOPSIS and PROMETHEE II.It is also noteworthy that TODIM has no significant correlation with the other methods.This result had been previously reported by Moshkovich et al. (2012) for the comparison between the TODIM and SAW methods.It is possible that this difference is because the TODIM method differs significantly from the other ranking methods, for its structure is based on the principles of boundary rationality proposed by the Prospect Theory of Kahneman & Tversky (1979).In fact, risk aspects are not present in the decision matrix, making the evaluation of the matrix more adherent with the realm of certainty.
Some participants make comments regarding the use of the methods.One participant who had the independency condition rejected, stated that "all the rankings were similar and the first choice was the same".For another participant in a similar condition of ranking dependency the commentary was: "I still prefer my initial ranking, but if I had the option to evaluate the alternatives for longer and more deeply, perhaps the methods made me change my mind".It was revealed that four of the participants from the eight, who had the conditional of independency rejected, had had previous contact with MCDM methods, which might explain more coherency on the initial ranking proposal (evaluation of the matrix) with the MCDM ranking methods.In fact, Almeida (2013) states that knowledge of the MCDM methods is one of the key factors for success implementation.
Conversely, the median for the question "Do you regret your initial ranking of alternatives?" was equal to 2 meaning that the participants strongly disagree with the assertion of rejecting their initial ranking in favor of the solution provided by one of the MCDM ranking methods.Table 4 presents the summary of the data used in the descriptive analysis.

Ranking disagreement through sensitivity analysis
The participant chosen for ranking disagreement evaluation through sensitivity analysis was Participant 2, since he had the highest score for W Kendall (similar rankings from different MCDM ranking methods) and the stronger propensity for assuming the solutions provided by the MCDM ranking methods (higher value for initial ranking regret).The parameters for the sensitivity analysis were the five criteria and alternatives from the decision matrix and the weighting vector was [0.079; 0.033; 0.054; 0.215; 0.152; 0.111; 0.016; 0.340], according to the weights calculated by the ROC method for Participant 2 in Table 2. Table 5 presents the results for Participant 2 for each MCDM ranking method using his weigh vector.
Table 5 -Initial ranking, and MCDM rankings using the weigh vector of Participant 2.
Ranking Initial ranking ELECTRE III SAW TOPSIS PROMETHEE II TODIM As noted in Table 5, even using the same decision matrix and weighting vector, the five methods showed different rankings and indicated different best alternatives.As expected, this variability is common to many MCDM ranking methods (Buede & Maxwell, 1995;Zanakis et al., 1998;Yeh, 2002).However, the value of W Kendall to Participant 2 was 0.760 (one of the highest found), which indicates high concordance among the MCDM ranking methods.Therefore, in order to test the stability of the rankings proposed by the methods, sensitivity analysis was carried out.
The first step of the sensitivity analysis was the random choice of a criterion.The second step was a change in the selected criterion by 20 % (more or less) and the recalculation of all relations with other criteria using Equation 3. The third step was the call of all methods in SANNA and the calculation of the TODIM method in its particular spreadsheet.Finally, the outputs were stored for all iterations, 100 in total.The sensitivity analysis lasted 14 minutes on an Intel Core 2 Duo of 2.93 Ghz with 3 GB of RAM memory.Figure 5 presents the average of correlation for each method among the hundred iterations with the rankings presented in Table 5 (MCDM rankings for Participant 2).
Given that the extraction of the weighting vector is a process that depends on the objective transcription of the subjective preferences of the decision makers, slight variations can be expected.It has been found in the literature that even slight variations might affect the performance of the alternatives (Zanakis et al., 1998).According to Figure 5, the TOPSIS and SAW methods presented ranking disagreement among the hundred iterations considering random changes of up to 20% that occurred (positive or negative), in some criterion randomly chosen from the weighting vector of Paticipant  internal ranking inconsistency.Regarding the choice of the best alternative, only TOPSIS did not maintain the indication of the best alternative for all iterations, counting five changes of the best alternative for this method.This result is in accordance with those of Buede & Maxwell (1995).Therefore, although TOPSIS had good performance in descriptive analysis, it is subject to ranking disagreement.No method suffered from ranking reversal in this application experiment.
As with the conclusions of Gomes & Costa (2015) it is here pointed out that the decision maker should consider a different method solution to enhance his decision process for greater knowledge of the problem.

CONCLUSIONS
Based on the idea that MCDM ranking methods can be evaluated in the sense of predicting the initial rankings given by the decision maker, an empirical experiment to evaluate the propensity for ranking predicting was carried out.The principal MCDM ranking methods, namely: SAW, TOPSIS, ELECTRE III, PROMETHEE II and TODIM were evaluated in terms of ranking accuracy (correct prediction of the ranking given by the user) and ranking similarity.It was found that just up to 20% of the initial ranking order was predicted entirely correctly by some of the methods, attracting attention to the performance of TOPSIS and ELECTRE III for this evaluation.Regarding the similarity of the methods it has been found that rankings were significantly different for 55% of the cases.The TOPSIS, PROMETHEE II and SAW methods had the highest similarity among the rankings while TODIM had no significant correlation with any other method, probably because its structure differs significantly to the other ranking methods.
The study also aimed to assess common ranking problems in the use of MCDM ranking methods.Considering that a significant part of these ranking problems are associated with the process of extracting decision makers' preferences, a participant with the highest similar rankings was chosen for testing whether slight changes in his weighting vector would cause deviation in the solution of the different MCDM ranking methods.By means of sensitivity analysis it was found that TOPSIS and SAW presented internal ranking inconsistency.Therefore, although TOPSIS had good performance for predicting the initial ranking, it suffered considerable ranking disagreement, also presenting the problem of replacing the best alternative in some iterations.The reversibility problem did not occur for any method.
Although this research being an experimental study from which it is not appropriate to conduct general or universal inferences, the results demonstrate that the most common errors found in the literature regarding the use of MCDM ranking methods are easily found.Therefore, considering the results of this study and those found in the literature it is difficult to advocate the use of a unique MCDM ranking method for ranking alternatives.In this sense, this research is a warning for the choice of MCDM ranking methods.It is suggested that special care must be taken in the choice of ranking methods and, besides axiomatic comparisons, ranking comparisons could be a useful way to enhance the decision making process, since MCDM methods are tools for learning about the problem and do not prescribe solutions that necessarily translate to the real state of the world.
The use of spreadsheets or software that perform the calculations for different methods is relevant to reducing the impacts of such ranking inconsistency, such as the examples of software mentioned in this paper.
study and the Figueira & Roy (2009) note is replaced by a discussion regarding the user's expectations of MCDM ranking methods in relation to the predicting capability of the user initial assessment.Pesquisa Operacional, Vol.36(2), 2016

Figure 1 -
Figure 1 -Flowchart for the sensitivity analysis programmed in Excel macro.

Figure 2 -
Figure 2 -Decision matrix set up as input for the various methods available in SANNA.

Figure 3 -
Figure 3 -PROMETHEE II function for indifference and preference.

Figure 4 -
Figure 4 -Average correlation with the initial ranking among the 20 participants.

Figure 5 -
Figure 5 -Average correlation with the first MCDM ranking among the 100 iterations.

Table 1 -
Decision matrix for participants evaluation.

Table 2 -
Weighting vectors for the participants.

Table 4 -
Summary of the data used in the descriptive analysis.
2. The ELECTRE III, PROMETHEE II and TODIM methods did not present