Acessibilidade / Reportar erro

Scale development: ten main limitations and recommendations to improve future research practices

Abstract

The scale development process is critical to building knowledge in human and social sciences. The present paper aimed (a) to provide a systematic review of the published literature regarding current practices of the scale development process, (b) to assess the main limitations reported by the authors in these processes, and (c) to provide a set of recommendations for best practices in future scale development research. Papers were selected in September 2015, with the search terms “scale development” and “limitations” from three databases: Scopus, PsycINFO, and Web of Science, with no time restriction. We evaluated 105 studies published between 1976 and 2015. The analysis considered the three basic steps in scale development: item generation, theoretical analysis, and psychometric analysis. The study identified ten main types of limitation in these practices reported in the literature: sample characteristic limitations, methodological limitations, psychometric limitations, qualitative research limitations, missing data, social desirability bias, item limitations, brevity of the scale, difficulty controlling all variables, and lack of manual instructions. Considering these results, various studies analyzed in this review clearly identified methodological weaknesses in the scale development process (e.g., smaller sample sizes in psychometric analysis), but only a few researchers recognized and recorded these limitations. We hope that a systematic knowledge of the difficulties usually reported in scale development will help future researchers to recognize their own limitations and especially to make the most appropriate choices among different conceptions and methodological strategies.

Assessment; Measurement; Psychometrics; Reliability; Validity

Introduction

In recent years, numerous measurement scales have been developed to assess attitudes, techniques, and interventions in a variety of scientific applications (Meneses et al. 201481. Meneses, J., Barrios, M., Bonillo, A., Cosculluela, A., Lozano, L. M., Turbany, J., & Valero, S. (2014). Psicometría. Barcelona: Editorial UOC.). Measurement is a fundamental activity of science, since it enables researchers to acquire knowledge about people, objects, events, and processes. Measurement scales are useful tools to attribute scores in some numerical dimension to phenomena that cannot be measured directly. They consist of sets of items revealing levels of theoretical variables otherwise unobservable by direct means (DeVellis 200338. DeVellis, R. F. (2003). Scale development: theory and applications (2nd ed.). Newbury Park: Sage Publications.).

A variety of authors (Clark and Watson 199531. Clark, L. A., & Watson, D. (1995). Constructing validity: basic issues in objective scale development. Psychological Assessment, 7(3), 309–319. http://dx.doi.org/10.1037/1040-3590.7.3.309.
http://dx.doi.org/10.1037/1040-3590.7.3....
; DeVellis 200338. DeVellis, R. F. (2003). Scale development: theory and applications (2nd ed.). Newbury Park: Sage Publications.; Nunnally 196788. Nunnally, J. C. (1967). Psychometric theory. New York: McGraw Hill.; Pasquali 201093. Pasquali, L. (2010). Instrumentação psicológica: fundamentos e práticas. Porto Alegre: Artmed.) have agreed that the scale development process involves complex and systematic procedures that require theoretical and methodological rigor. According to these authors, the scale development process can be carried out in three basic steps.

In the first step, commonly referred as “item generation,” the researcher provides theoretical support for the initial item pool (Hutz et al. 201564. Hutz, CS, Bandeira, DR, & Trentini, CM. (Org.). (2015). Psicometria. Porto Alegre, Artmed). Methods for the initial item generation can be classified as deductive, inductive, or a combination of the two. Deductive methods involve item generation based on an extensive literature review and pre-existing scales (Hinkin 199562. Hinkin, T. R. (1995). A review of scale development practices in the study of organizations. Journal of Management, 21(5), 967–988. http://dx.doi.org/10.1177/014920639502100509.
http://dx.doi.org/10.1177/01492063950210...
). On the other hand, inductive methods base item development on qualitative information regarding a construct obtained from opinions gathered from the target population—e.g., focus groups, interviews, expert panels, and qualitative exploratory research methodologies (Kapuscinski and Masters 201066. Kapuscinski, A. N., & Masters, K. S. (2010). The current status of measures of spirituality: a critical review of scale development. Psychology of Religion and Spirituality, 2(4), 191–205. http://dx.doi.org/10.1037/a0020498.
http://dx.doi.org/10.1037/a0020498...
). The researcher is also concerned with a variety of parameters that regulate the setting of each item and of the scale as a whole. For example, suitable scale instructions, an appropriate number of items, adequate display format, appropriate item redaction (all items should be simple, clear, specific, ensure the variability of response, remain unbiased, etc.), among other parameters (DeVellis 200338. DeVellis, R. F. (2003). Scale development: theory and applications (2nd ed.). Newbury Park: Sage Publications.; Pasquali 201093. Pasquali, L. (2010). Instrumentação psicológica: fundamentos e práticas. Porto Alegre: Artmed.).

In the second step, usually referred to as the “theoretical analysis,” the researcher assesses the content validity of the new scale, ensuring that the initial item pool reflects the desired construct (Arias et al. 20145. Arias, M. R. M., Lloreda, M. J. H., & Lloreda, M. V. H. (2014). Psicometría. S.A.: Alianza Editorial). A content validity assessment is required, since inferences are made based on the final scale items. The item content must be deemed valid to instill confidence in all consequent inferences. In order to ensure the content validity, the researcher seeks other opinions about the operationalized items. The opinions can be those of expert judges (experts in the development scales or experts in the target construct) or target population judges (potential users of the scale), enabling the researcher to ensure that the hypothesis elaborated in the research appropriately represents the construct of interest (Nunnally 196788. Nunnally, J. C. (1967). Psychometric theory. New York: McGraw Hill.).

In the last step, psychometric analysis, the researcher should assess whether the new scale has construct validity and reliability. Construct validity is most directly related to the question of what the instrument is in fact measuring—what construct, trait, or concept underlies an individual’s performance or score on a measure (Churchill 197929. Churchill, G. (1979). A paradigm for developing better measures of marketing constructs. Journal of Marketing Research, 16(1), 64–73. http://dx.doi.org/10.2307/3150876.
http://dx.doi.org/10.2307/3150876...
). This refers to the degree to which inferences can be legitimately made from the observed scores to the theoretical constructs about which these observations are supposed to contain information (Podsakoff et al. 201396. Podsakoff, N. P., Podsakoff, P. M., MacKenzie, S. B., & Klinger, R. L. (2013). Are we really measuring what we say we’re measuring? Using video techniques to supplement traditional construct validation procedures. Journal of Applied Psychology, 98(1), 99–113. http://dx.doi.org/10.1037/a0029570.
http://dx.doi.org/10.1037/a0029570...
). Construct validity can be assessed with the use of exploratory factor analysis (EFA), confirmatory factor analysis (CFA), or with convergent, discriminant, predictive/nomological, criterion, internal, and external validity. In turn, reliability is a measure of score consistency, usually measured by use of internal consistency, test-retest reliability, split-half, item-total correlation/inter-item reliability, and inter-observer reliability (DeVellis 200338. DeVellis, R. F. (2003). Scale development: theory and applications (2nd ed.). Newbury Park: Sage Publications.). To ensure construct validity and reliability, the data should be collected in a large and appropriately representative sample of the target population. It is a common rule of thumb that there should be at least 10 participants for each item of the scale, making an ideal of 15:1 or 20:1 (Clark and Watson 199531. Clark, L. A., & Watson, D. (1995). Constructing validity: basic issues in objective scale development. Psychological Assessment, 7(3), 309–319. http://dx.doi.org/10.1037/1040-3590.7.3.309.
http://dx.doi.org/10.1037/1040-3590.7.3....
; DeVellis 200338. DeVellis, R. F. (2003). Scale development: theory and applications (2nd ed.). Newbury Park: Sage Publications.; Hair Junior et al. 200954. Hair Junior, J. F., Black, W. C., Babin, N. J., Anderson, R. E., & Tatham, R. L. (2009). Análise multivariada de dados (6ath ed.). São Paulo: Bookman.).

Although the literature on theoretical and methodological care in scale development is extensive, many limitations have been identified in the process. These include failure to adequately define the construct domain, failure to correctly specify the measurement model, underutilization of some techniques that are helpful in establishing construct validity (MacKenzie et al. 201177. MacKenzie, S. B., Podsakoff, P. M., & Podsakoff, N. P. (2011). Construct measurement and validation procedures in MIS and behavioral research: integrating new and existing techniques. MIS Quarterly, 35(2), 293–334.), relatively weak psychometric properties, applicability to only a single form of treatment or manual, extensive time required to fill out the questionnaire (Hilsenroth et al. 200561. Hilsenroth, M. J., Blagys, M. D., Ackerman, S. J., Bonge, D. R., & Blais, M. A. (2005). Measuring psychodynamic-interpersonal and cognitive-behavioral techniques: development of the comparative psychotherapy process scale. Psychotherapy: Theory, Research, Practice, Training, 42(3), 340–356. http://dx.doi.org/10.1037/0033-3204.42.3.340.
http://dx.doi.org/10.1037/0033-3204.42.3...
), inappropriate item redaction, too few items and participants in the construction and analysis, an imbalance between items that assess positive beliefs and those that assess negative beliefs (Prados 200798. Prados, J. M. (2007). Development of a new scale of beliefs about the worry consequences. Annals of Psychology, 23(2), 226–230.), social desirability bias (King and Bruner 200071. King, M. F., & Bruner, G. C. (2000). Social desirability bias: a neglected aspect of validity testing. Psychology and Marketing, 17(2), 79–103. http://dx.doi.org/10.1002/(SICI)1520-6793(200002)17:2<79::AID-MAR2>3.0.CO;2-0.
http://dx.doi.org/10.1002/(SICI)1520-679...
), among others.

These limitations in the scale development process weaken the obtained psychometric results, limiting the future applicability of the new scale and hindering its generalizability. In this sense, knowledge of the most often reported limitations is fundamental in providing essential information to help develop best practices for future research in this area. The purpose of this article is threefold: (a) to provide a systematic review of the published literature regarding some current practices of the scale development process, (b) to assess the main limitations reported by the authors in this process, and (c) to provide a set of recommendations for best practices in future scale development research.

Review

Method

This systematic review identified and selected papers from three databases: Scopus, PsycINFO, and Web of Science. There was no time restriction in the literature search, which was completed in September 1, 2015. The following search term was used: “scale development.” In the set of databases analyzed, the search was done inclusively in “Any Field” (PsycINFO), in “Article Title, Abstract, Keywords” (Scopus), or in any “Topic” (Web of Science). In addition, we used an advanced search to filter the articles in (search within results), with the search term “limitations” identified in “Any Field” in all databases. Both terms were used in English only. Four reviewers evaluated the papers in an independent and blinded way. Any disagreements on eligibility of a particular study were resolved through consensus among reviewers.

Figure 1 shows a flowchart summarizing the strategy adopted for identification and selection of studies. We used only one inclusion criteria for the evaluation of the studies: (a) articles that aim to develop and validate self-administered measurement scales for humans. We excluded (a) unavailable full-text papers in the analyzed databases, (b) papers in languages other than English, Portuguese, or Spanish, (c) articles which were not clearly aimed at the development of a new scale (i.e., we excluded articles investigating only the reliability, validity, or revisions of existing scales and studies that describe the validation of instruments for other languages), (d) papers with unvalidated scales, and (e) articles that did not declare the limitations of the study.

Fig. 1
Flowchart showing summary of the systematic process of identifying and selecting article

Results

In all, this systematic review evaluated 105 studies published between 1976 and 2015. Most (88.5%) was published between 2005 and 2015, and only two studies date from the last century. We analyzed two major issues: (a) current practices of the scale development process—considering the three steps usually reported in the literature (step 1—item generation, step 2—theoretical analysis, step 3—psychometric analysis), the number of participants in step 3, the number of items in the beginning scale, and the number of items in the final scale; (b) main limitations reported by the authors in the scale development process—considering the limitations observed and recorded by the authors during the scale development process. The description of these results can be found in Table 1.

Table 1
Systematic review of the scale development process recorded in 105 included studies

Current practices of the scale development process

Step 1—item generation In the first step, 35.2% (n= 37) of the studies reported using exclusively deductive methods to write items, 7.6% (n= 8) used only inductive methods, and 56.2% (n= 59) combined deductive and inductive strategies. The majority of the studies used a literature review (84.7%, n= 89) as the deductive method in item generation. In inductive methods, 26.6% of studies (n= 28) chose to conduct an interview.

Step 2—theoretical analysis In order to theoretically refine the items, several studies used opinions of experts (74.2%, n= 78), whereas others used target population opinions (43.8%, n= 46). In addition, 63.8% (n= 67) of the studies used only one of these approaches (expert or population judges).

Step 3—psychometric analysis The most common analyses that have been used to assess construct validity are EFA (88.6%, n= 93), CFA (72.3%, n= 76), convergent validity (72.3%, n= 76), and discriminant validity (56.2%, n= 59). Most studies opted to combine EFA and CFA (65.7%, n= 69). Only 4.7% (n= 5) failed to use factor analysis in their research. In relation to study reliability, internal consistency checks were used by all studies and test-retest reliability was the second most commonly used technique (22.8%, n= 24).

Sample size in step 3 and number of items Interestingly, 50.4% (n= 53) of the studies used sample sizes smaller than the rule of thumb, which is a minimum of 10 participants for each item in the scale. Regarding number of items, the majority of the studies (49.6%, n= 52) lost more than 50% of the initial item pool during the validation process.

Table 2 summarizes and provides more details on our findings regarding the current practices in the scale development.

Table 2
Summary of current practices of the scale development process

Main limitations reported in the scale development process

As result of this systematic review, we found ten main limitations commonly referenced in the scale development process: (1) sample characteristic limitations—cited by 81% of the studies, (2) methodological limitations—33.2%, (3) psychometric limitations—30.4%, (4) qualitative research limitations—5.6%, (5) missing data—2.8%, (6) social desirability bias—1.9%, (7) item limitations—1.9%, (8) brevity of the scale—1.9%, (9) difficulty controlling all variables—0.9%, and (10) lack of manual instructions—0.9%. Table 3 summarizes these findings.

Table 3
Scale development process—ten main limitations

Discussion

This systematic review was primarily directed at identifying the published literature regarding current practices of the scale development. The results show a variety of practices that have been used to generate and assess items, both theoretically and psychometrically. We evaluated these current practices, considering three distinct steps (item generation, theoretical analysis, and psychometric analysis). We also considered the relationship between sample size and number of items, since this is considered an important methodological aspect to be evaluated during the scale development process. The results are discussed together with recommendations for best practices in future scale development research.

Current practices of the scale development process—findings and research implications

Regarding step 1, item generation, our results show that, although several studies used exclusively deductive methods (e.g., Henderson-King and Henderson-King 200558. Henderson-King, D., & Henderson-King, E. (2005). Acceptance of cosmetic surgery: scale development and validation. Body Image, 2, 137–149. http://dx.doi.org/10.1016/j.bodyim.2005.03.003.
http://dx.doi.org/10.1016/j.bodyim.2005....
; Kim et al. 201169. Kim, S., Cha, J., Knutson, B. J., & Beck, J. A. (2011). Development and testing of the Consumer Experience Index (CEI). Managing Service Quality: An International Journal, 21(2), 112–132. http://dx.doi.org/10.1108/09604521111113429.
http://dx.doi.org/10.1108/09604521111113...
), the majority (e.g., Bakar and Mustaffa 201310. Bakar, H. A., & Mustaffa, C. S. (2013). Organizational communication in Malaysia organizations. Corporate Communications: An International Journal, 18(1), 87–109. http://dx.doi.org/10.1108/13563281311294146.
http://dx.doi.org/10.1108/13563281311294...
; Uzunboylu and Ozdamli 2011126. Uzunboylu, H., & Ozdamli, F. (2011). Teacher perception for m-learning: scale development and teachers’ perceptions. Journal of Computer Assisted Learning, 27, 544–556. http://dx.doi.org/10.1111/j.1365-2729.2011.00415.x.
http://dx.doi.org/10.1111/j.1365-2729.20...
) combined deductive and inductive methods, a combination consistent with the recommended strategy for the creation of new measures (DeVellis 200338. DeVellis, R. F. (2003). Scale development: theory and applications (2nd ed.). Newbury Park: Sage Publications.). These findings, however, differ from previous critical reviews of scale development practices, which found that most of the reported studies used exclusively deductive methods (Hinkin 199562. Hinkin, T. R. (1995). A review of scale development practices in the study of organizations. Journal of Management, 21(5), 967–988. http://dx.doi.org/10.1177/014920639502100509.
http://dx.doi.org/10.1177/01492063950210...
; Kapuscinski and Masters 201066. Kapuscinski, A. N., & Masters, K. S. (2010). The current status of measures of spirituality: a critical review of scale development. Psychology of Religion and Spirituality, 2(4), 191–205. http://dx.doi.org/10.1037/a0020498.
http://dx.doi.org/10.1037/a0020498...
; Ladhari 201073. Ladhari, R. (2010). Developing e-service quality scales: a literature review. Journal of Retailing and Consumer Services, 17, 464–477. http://dx.doi.org/10.1016/j.jretconser.2010.06.003.
http://dx.doi.org/10.1016/j.jretconser.2...
). This is particularly important since the quality of generated items depends on the way that the construct is defined. Failing to adequately define the conceptual domain of a construct causes several problems related to poor construct definition, leading to, for example, (a) confusion about what the construct does and does not refer to, including the similarities and differences between it and other constructs that already exist in the field, (b) indicators that may either be deficient or contaminated, and (c) invalid conclusions about relationships with other constructs (MacKenzie et al. 201177. MacKenzie, S. B., Podsakoff, P. M., & Podsakoff, N. P. (2011). Construct measurement and validation procedures in MIS and behavioral research: integrating new and existing techniques. MIS Quarterly, 35(2), 293–334.). Considering that item generation may be the most important part of the scale development process, future measures should be developed using the appropriate definition of the conceptual domain based on the combination of both deductive and inductive approaches.

Our results suggest that literature review was the most widely used deductive method (e.g., Bolton and Lane 201217. Bolton, D. L., & Lane, M. D. (2012). Individual entrepreneurial orientation: development of a measurement instrument. Education + Training, 54(2/3), 219–233. http://dx.doi.org/10.1108/00400911211210314.
http://dx.doi.org/10.1108/00400911211210...
; Henderson-King and Henderson-King 200558. Henderson-King, D., & Henderson-King, E. (2005). Acceptance of cosmetic surgery: scale development and validation. Body Image, 2, 137–149. http://dx.doi.org/10.1016/j.bodyim.2005.03.003.
http://dx.doi.org/10.1016/j.bodyim.2005....
). This is consistent with the views of several other researchers who have systematically reviewed scales (Bastos et al. 201012. Bastos, J. L., Celeste, R. K., Faerstein, E., & Barros, A. J. D. (2010). Racial discrimination and health: a systematic review of scales with a focus on their psychometric properties. Social Science and Medicine, 70, 1091–1099. http://dx.doi.org/10.1016/j.socscimed.2009.12.20.
http://dx.doi.org/10.1016/j.socscimed.20...
; Ladhari 201073. Ladhari, R. (2010). Developing e-service quality scales: a literature review. Journal of Retailing and Consumer Services, 17, 464–477. http://dx.doi.org/10.1016/j.jretconser.2010.06.003.
http://dx.doi.org/10.1016/j.jretconser.2...
; Sveinbjornsdottir and Thorsteinsson 2008120. Sveinbjornsdottir, S., & Thorsteinsson, E. B. (2008). Adolescent coping scales: a critical psychometric review. Scandinavian Journal of Psychology, 49(6), 533–548. http://dx.doi.org/10.1111/j.1467-9450.2008.00669.x.
http://dx.doi.org/10.1111/j.1467-9450.20...
). Nevertheless, this finding differs from another study (Kapuscinski and Masters 201066. Kapuscinski, A. N., & Masters, K. S. (2010). The current status of measures of spirituality: a critical review of scale development. Psychology of Religion and Spirituality, 2(4), 191–205. http://dx.doi.org/10.1037/a0020498.
http://dx.doi.org/10.1037/a0020498...
) that found that the most common deductive strategies were reading works by spiritual leaders, theory written by psychologists, and discussion among authors. Literature review should be considered central for the enumeration of the constructs. It also serves to clarify the nature and variety of the target construct content. In addition, literature reviews help to identify existing measures that can be used as references to create new scales (Clark and Watson 199531. Clark, L. A., & Watson, D. (1995). Constructing validity: basic issues in objective scale development. Psychological Assessment, 7(3), 309–319. http://dx.doi.org/10.1037/1040-3590.7.3.309.
http://dx.doi.org/10.1037/1040-3590.7.3....
; DeVellis 200338. DeVellis, R. F. (2003). Scale development: theory and applications (2nd ed.). Newbury Park: Sage Publications.). In this sense, future research should consider the literature review as the initial and necessary deductive step foundational to building a new scale.

This review also highlights the fact that interviews and focus groups were the most widely used inductive methods (e.g., Lin and Hsieh 201174. Lin, J. C., & Hsieh, P. (2011). Assessing the self-service technology encounters: development and validation of SSTQUAL scale. Journal of Retailing, 87(2), 194–206. http://dx.doi.org/10.1016/j.jretai.2011.02.006.
http://dx.doi.org/10.1016/j.jretai.2011....
; Sharma 2010112. Sharma, P. (2010). Measuring personal cultural orientations: scale development and validation. Journal of the Academy of Marketing Science, 38, 787–806. http://dx.doi.org/.1007/s11747-009-0184-7.
http://dx.doi.org/.1007/s11747-009-0184-...
). Similar results were found in the systematic review by Kapuscinski and Masters (2010)66. Kapuscinski, A. N., & Masters, K. S. (2010). The current status of measures of spirituality: a critical review of scale development. Psychology of Religion and Spirituality, 2(4), 191–205. http://dx.doi.org/10.1037/a0020498.
http://dx.doi.org/10.1037/a0020498...
, Sveinbjornsdottir and Thorsteinsson (2008)120. Sveinbjornsdottir, S., & Thorsteinsson, E. B. (2008). Adolescent coping scales: a critical psychometric review. Scandinavian Journal of Psychology, 49(6), 533–548. http://dx.doi.org/10.1111/j.1467-9450.2008.00669.x.
http://dx.doi.org/10.1111/j.1467-9450.20...
, and Ladhari (2010)73. Ladhari, R. (2010). Developing e-service quality scales: a literature review. Journal of Retailing and Consumer Services, 17, 464–477. http://dx.doi.org/10.1016/j.jretconser.2010.06.003.
http://dx.doi.org/10.1016/j.jretconser.2...
. These findings have particular relevance to future researchers, since they emphasize the importance of using methodological strategies that consider the opinions of the target population. Despite the fact that a panel of experts contributes widely to increasing the researchers’ confidence in the content validity of the new scale, it is important to also consider the most original and genuine information about the construct of interest, which can be best obtained through reports obtained from interviews and focus groups with the target population.

Related to step 2, theoretical analysis, the results of this review indicate that expert judges have been the most widely utilized tool for analyzing content validity (e.g., Uzunboylu and Ozdamli 2011126. Uzunboylu, H., & Ozdamli, F. (2011). Teacher perception for m-learning: scale development and teachers’ perceptions. Journal of Computer Assisted Learning, 27, 544–556. http://dx.doi.org/10.1111/j.1365-2729.2011.00415.x.
http://dx.doi.org/10.1111/j.1365-2729.20...
; Zheng et al. 2010137. Zheng, J, You, L, Lou, T, Chen, N, Lai, D, Liang, Y, … Zhai, C. (2010). Development and psychometric evaluation of the dialysis patient-perceived exercise benefits and barriers scale. International Journal of Nursing Studies, 47, 166-180. doi: http://dx.doi.org/10.1016/j.ijnurstu.2009.05.023
http://dx.doi.org/10.1016/j.ijnurstu.200...
). Previous studies have also found expert opinion to be the most common qualitative method for the elimination of unsuitable items (Kapuscinski and Masters 201066. Kapuscinski, A. N., & Masters, K. S. (2010). The current status of measures of spirituality: a critical review of scale development. Psychology of Religion and Spirituality, 2(4), 191–205. http://dx.doi.org/10.1037/a0020498.
http://dx.doi.org/10.1037/a0020498...
; Ladhari 201073. Ladhari, R. (2010). Developing e-service quality scales: a literature review. Journal of Retailing and Consumer Services, 17, 464–477. http://dx.doi.org/10.1016/j.jretconser.2010.06.003.
http://dx.doi.org/10.1016/j.jretconser.2...
). In the literature review conducted by Hardesty and Bearden (2004)57. Hardesty, D. M., & Bearden, W. O. (2004). The use of expert judges in scale development Implications for improving face validity of measures of unobservable constructs. Journal of Business Research, 57, 98–107. http://dx.doi.org/10.1016/S0148-2963(01)00295-8.
http://dx.doi.org/10.1016/S0148-2963(01)...
, the authors highlighted the importance of these experts to carefully analyze the initial item pool. They suggested that any research using new, changed, or previously unexamined scale items, should at a minimum be judged by a panel of experts. However, the authors also point out the apparent lack of consistency in the literature in terms of how researchers use the opinions of expert judges in aiding the decision of whether or not to retain items for a scale. Given this inconsistency, the authors developed guidelines regarding the application of different decision rules to use for item retention. For example, the “sumscore decision rule,” defined as the total score for an item across all judges, is considered by the authors to be the most effective in predicting whether an item should be included in a scale and appears, therefore, to be a reasonable rule for researchers to employ.

Future research in developing scales should be concerned, not only with opinions from experts but also with the opinions of the target population. The results of this review show that only a minority of studies considered the review of the scales’ items by members of the target population (e.g., Uzunboylu and Ozdamli 2011126. Uzunboylu, H., & Ozdamli, F. (2011). Teacher perception for m-learning: scale development and teachers’ perceptions. Journal of Computer Assisted Learning, 27, 544–556. http://dx.doi.org/10.1111/j.1365-2729.2011.00415.x.
http://dx.doi.org/10.1111/j.1365-2729.20...
; Zheng et al. 2010137. Zheng, J, You, L, Lou, T, Chen, N, Lai, D, Liang, Y, … Zhai, C. (2010). Development and psychometric evaluation of the dialysis patient-perceived exercise benefits and barriers scale. International Journal of Nursing Studies, 47, 166-180. doi: http://dx.doi.org/10.1016/j.ijnurstu.2009.05.023
http://dx.doi.org/10.1016/j.ijnurstu.200...
). In addition, a smaller minority combined the two approaches in the assessment of item content (e.g., Mahudin et al. 201278. Mahudin, N. D. M., Cox, T., & Griffiths, A. (2012). Measuring rail passenger crowding: scale development and psychometric properties. Transportation Research Part, F 15, 38–51. http://dx.doi.org/10.1016/j.trf.2011.11.006.
http://dx.doi.org/10.1016/j.trf.2011.11....
; Morgado et al. 201483. Morgado, F. F. R., Campana, A. N. N. B., & Tavares, M. C. G. C. F. (2014). Development and validation of the self-acceptance scale for persons with early blindness: the SAS-EB. PloS One, 9(9), e106848. http://dx.doi.org/10.1371/journal.pone.0106848.
http://dx.doi.org/10.1371/journal.pone.0...
). The limited use of target population opinions is a problem. A previous study of systematic scale development reviews found that the opinion of these people is the basis for content validity (Bastos et al. 201012. Bastos, J. L., Celeste, R. K., Faerstein, E., & Barros, A. J. D. (2010). Racial discrimination and health: a systematic review of scales with a focus on their psychometric properties. Social Science and Medicine, 70, 1091–1099. http://dx.doi.org/10.1016/j.socscimed.2009.12.20.
http://dx.doi.org/10.1016/j.socscimed.20...
). As highlighted by Clark and Watson (1995)31. Clark, L. A., & Watson, D. (1995). Constructing validity: basic issues in objective scale development. Psychological Assessment, 7(3), 309–319. http://dx.doi.org/10.1037/1040-3590.7.3.309.
http://dx.doi.org/10.1037/1040-3590.7.3....
and Malhotra (2004)79. Malhotra, N. K. (2004). Pesquisa de marketing: Uma orientação aplicada (4ath ed.). Porto Alegre: Bookman., it is essential for the new scale to undergo prior review by members of the target population. Pre-test or pilot study procedures make it possible to determine respondents’ opinions of, and reactions to, each item on the scale, enabling researchers to identify and eliminate potential problems in the scale before it is applied at large.

Another problem noted in this systematic review was that some studies failed to clearly report how they performed the theoretical analysis of the items (e.g., Glynn et al. 201552. Glynn, N. W., Santanasto, A. J., Simonsick, E. M., Boudreau, R. M., Beach, S. R., Schulz, R., & Newman, A. B. (2015). The Pittsburgh fatigability scale for older adults: development and validation. Journal of American Geriatrics Society, 63, 130–135. http://dx.doi.org/10.1111/jgs.13191.
http://dx.doi.org/10.1111/jgs.13191...
; Gottlieb et al. 201453. Gottlieb, U., Brown, M., & Ferrier, L. (2014). Consumer perceptions of trade show effectiveness. European Journal of Marketing, 48(1/2), 89–107. http://dx.doi.org/10.1108/EJM-06-2011-0310.
http://dx.doi.org/10.1108/EJM-06-2011-03...
). We hypothesized that the authors either did not perform this analysis or found it unimportant to record. Future research should consider this analysis, as well as all subsequent analyses, necessary and relevant for reporting.

Almost all studies (95.3%) reported using at least one type of factor analysis—EFA or CFA—in step 3, psychometric analysis (e.g., Sewitch et al. 2003111. Sewitch, M. J., Abrahamowicz, M., Dobkin, P. L., & Tamblyn, R. (2003). Measuring differences between patients’ and physicians’ health perceptions: the patient–physician discordance scale. Journal of Behavioral Medicine, 26(3), 245–263. http://dx.doi.org/10.1023/A:1023412604715.
http://dx.doi.org/10.1023/A:102341260471...
; Tanimura et al. 2011122. Tanimura, C., Morimoto, M., Hiramatsu, K., & Hagino, H. (2011). Difficulties in the daily life of patients with osteoarthritis of the knee: scale development and descriptive study. Journal of Clinical Nursing, 20, 743–753. http://dx.doi.org/10.1111/j.1365-2702.2010.03536.x.
http://dx.doi.org/10.1111/j.1365-2702.20...
). Clark and Watson (1995)31. Clark, L. A., & Watson, D. (1995). Constructing validity: basic issues in objective scale development. Psychological Assessment, 7(3), 309–319. http://dx.doi.org/10.1037/1040-3590.7.3.309.
http://dx.doi.org/10.1037/1040-3590.7.3....
consider that “unfortunately, many test developers are hesitant to use factor analysis, either because it requires a relatively large number of respondents or because it involves several perplexing decisions” (p. 17). They emphasized the importance of the researcher’s need to understand and apply this analysis, “it is important that test developers either learn about the technique or consult with a psychometrician during the scale development process” (Clark and Watson 199531. Clark, L. A., & Watson, D. (1995). Constructing validity: basic issues in objective scale development. Psychological Assessment, 7(3), 309–319. http://dx.doi.org/10.1037/1040-3590.7.3.309.
http://dx.doi.org/10.1037/1040-3590.7.3....
, p. 17). This question seems to have been almost overcome in recent studies, since the vast majority of the analyzed studies used the factor analysis method.

Among the studies than used factor analysis, the majority chose to use EFA (e.g., Bakar and Mustaffa 201310. Bakar, H. A., & Mustaffa, C. S. (2013). Organizational communication in Malaysia organizations. Corporate Communications: An International Journal, 18(1), 87–109. http://dx.doi.org/10.1108/13563281311294146.
http://dx.doi.org/10.1108/13563281311294...
; Turker 2009125. Turker, D. (2009). Measuring corporate social responsibility: a scale development study. Journal of Business Ethics, 85, 411–427. http://dx.doi.org/10.1007/s10551-008-9780-6.
http://dx.doi.org/10.1007/s10551-008-978...
). Similar to our findings, Bastos et al. (2010)12. Bastos, J. L., Celeste, R. K., Faerstein, E., & Barros, A. J. D. (2010). Racial discrimination and health: a systematic review of scales with a focus on their psychometric properties. Social Science and Medicine, 70, 1091–1099. http://dx.doi.org/10.1016/j.socscimed.2009.12.20.
http://dx.doi.org/10.1016/j.socscimed.20...
and Ladhari (2010)73. Ladhari, R. (2010). Developing e-service quality scales: a literature review. Journal of Retailing and Consumer Services, 17, 464–477. http://dx.doi.org/10.1016/j.jretconser.2010.06.003.
http://dx.doi.org/10.1016/j.jretconser.2...
found EFA to be the more commonly utilized construct validity method when compared to CFA. EFA has extensive value because it is considered to be effective in identifying the underlying latent variables or factors of a measure by exploring relationships among observed variables. However, it allows for more subjectivity in the decision-making process than many other statistical procedures, which can be considered a problem (Roberson et al. 2014104. Roberson, R. B., III, Elliott, T. R., Chang, J. E., & Hill, J. N. (2014). Exploratory factor analysis in rehabilitation psychology: a content analysis. Rehabilitation Psychology, 59(4), 429–438. http://dx.doi.org/10.1037/a0037899.
http://dx.doi.org/10.1037/a0037899...
).

For more consistent results on the psychometric indices of the new scale, DeVellis (2003)38. DeVellis, R. F. (2003). Scale development: theory and applications (2nd ed.). Newbury Park: Sage Publications. indicates the combined use of EFA and CFA, as was performed with most studies evaluated in this review. In CFA, the specific hypothesized factor structure proposed in EFA (including the correlations among the factors) is statistically evaluated. If the estimated model fits the data, then a researcher concludes that the factor structure replicates. If not, the modification indices are used to identify where constraints placed on the factor pattern are causing a misfit (Reise et al. 2000101. Reise, S. P., Waller, N. G., & Comrey, A. L. (2000). Factor analysis and scale revision. Psychological Assessment, 12(3), 287–297. http://dx.doi.org/10.1037//1040-3590.12.3.287.
http://dx.doi.org/10.1037//1040-3590.12....
). Future studies should consider the combined use of EFA and CFA during the evaluation of construct validity of the new measure, and should also apply a combination of multiple fit indices (e.g., modification indices) in order to provide more consistent psychometric results.

After EFA and CFA, convergent validity was the preferred technique used in the vast majority of the studies included in this review (e.g., Brun et al. 201422. Brun, I., Rajaobelina, L., & Ricard, L. (2014). Online relationship quality: scale development and initial testing. International Journal of Bank Marketing, 32(1), 5–27. http://dx.doi.org/10.1108/IJBM-02-2013-0022.
http://dx.doi.org/10.1108/IJBM-02-2013-0...
; Cicero et al. 201030. Cicero, D. C., Kerns, J. G., & McCarthy, D. M. (2010). The Aberrant Salience Inventory: a new measure of psychosis proneness. Psychological Assessment, 22(3), 688–701. http://dx.doi.org/10.1037/a0019913.
http://dx.doi.org/10.1037/a0019913...
). This finding is consistent with prior research (Bastos et al. 201012. Bastos, J. L., Celeste, R. K., Faerstein, E., & Barros, A. J. D. (2010). Racial discrimination and health: a systematic review of scales with a focus on their psychometric properties. Social Science and Medicine, 70, 1091–1099. http://dx.doi.org/10.1016/j.socscimed.2009.12.20.
http://dx.doi.org/10.1016/j.socscimed.20...
). Convergent validity consists in examining whether a scale’s score is associated with the other variables and measures of the same construct to which it should be related. It is verified either by calculating the average variance extracted for each factor when the shared variance accounted for 0.50 or more of the total variance or by correlating their scales with a measure of overall quality (Ladhari 201073. Ladhari, R. (2010). Developing e-service quality scales: a literature review. Journal of Retailing and Consumer Services, 17, 464–477. http://dx.doi.org/10.1016/j.jretconser.2010.06.003.
http://dx.doi.org/10.1016/j.jretconser.2...
). In the sequence of convergent validity, the following methods were identified as favorites in the assessment of construct validity: discriminant validity (the extent to which the scale’s score does not correlate with unrelated constructs) (e.g., Coker et al. 201132. Coker, B. L. S., Ashill, N. J., & Hope, B. (2011). Measuring internet product purchase risk. European Journal of Marketing, 45(7/8), 1130–1151. http://dx.doi.org/10.1108/03090561111137642.
http://dx.doi.org/10.1108/03090561111137...
), predictive/nomological validity (the extent to which the scores of one construct are empirically related to the scores of other conceptually related constructs) (e.g., Sharma 2010112. Sharma, P. (2010). Measuring personal cultural orientations: scale development and validation. Journal of the Academy of Marketing Science, 38, 787–806. http://dx.doi.org/.1007/s11747-009-0184-7.
http://dx.doi.org/.1007/s11747-009-0184-...
), criterion validity (the empirical association that the new scale has with a gold standard criterion concerned with the prediction of a certain behavior) (e.g., Tanimura et al. 2011122. Tanimura, C., Morimoto, M., Hiramatsu, K., & Hagino, H. (2011). Difficulties in the daily life of patients with osteoarthritis of the knee: scale development and descriptive study. Journal of Clinical Nursing, 20, 743–753. http://dx.doi.org/10.1111/j.1365-2702.2010.03536.x.
http://dx.doi.org/10.1111/j.1365-2702.20...
), internal (signifies whether the study results and conclusions are valid for the study population), and external validity (generalizability of study) (e.g., Bolton and Lane 201217. Bolton, D. L., & Lane, M. D. (2012). Individual entrepreneurial orientation: development of a measurement instrument. Education + Training, 54(2/3), 219–233. http://dx.doi.org/10.1108/00400911211210314.
http://dx.doi.org/10.1108/00400911211210...
; Khorsan and Crawford 201468. Khorsan, R, & Crawford, C. (2014). External validity and model validity: a conceptual approach for systematic review methodology. Evidence-Based Complementary and Alternative Medicine, 2014, Article ID 694804, 12 pages. doi: http://dx.doi.org/10.1155/2014/694804
http://dx.doi.org/10.1155/2014/694804...
). Considering the importance of validity to ensure the quality of the collected data and the generalized potential of the new instrument, future studies should allow different ways to assess the validity of the new scale, thus increasing the psychometric rigor of the analysis.

With regard to reliability, all studies reported internal consistency statistics (Cronbach’s alpha) for all subscales and/or the final version of the full scale (e.g., Schlosser and McNaughton 2009110. Schlosser, F. K., & McNaughton, R. B. (2009). Using the I-MARKOR scale to identify market-oriented individuals in the financial services sector. Journal of Services Marketing, 23(4), 236–248. http://dx.doi.org/10.1108/08876040910965575.
http://dx.doi.org/10.1108/08876040910965...
; Sewitch et al. 2003111. Sewitch, M. J., Abrahamowicz, M., Dobkin, P. L., & Tamblyn, R. (2003). Measuring differences between patients’ and physicians’ health perceptions: the patient–physician discordance scale. Journal of Behavioral Medicine, 26(3), 245–263. http://dx.doi.org/10.1023/A:1023412604715.
http://dx.doi.org/10.1023/A:102341260471...
). These findings are consistent with those of previous review studies (Bastos et al. 201012. Bastos, J. L., Celeste, R. K., Faerstein, E., & Barros, A. J. D. (2010). Racial discrimination and health: a systematic review of scales with a focus on their psychometric properties. Social Science and Medicine, 70, 1091–1099. http://dx.doi.org/10.1016/j.socscimed.2009.12.20.
http://dx.doi.org/10.1016/j.socscimed.20...
; Kapuscinski and Masters 201066. Kapuscinski, A. N., & Masters, K. S. (2010). The current status of measures of spirituality: a critical review of scale development. Psychology of Religion and Spirituality, 2(4), 191–205. http://dx.doi.org/10.1037/a0020498.
http://dx.doi.org/10.1037/a0020498...
). DeVellis (2003)38. DeVellis, R. F. (2003). Scale development: theory and applications (2nd ed.). Newbury Park: Sage Publications. explains that internal consistency is the most widely used measure of reliability. It is concerned with the homogeneity of the items within a scale. Given its importance, future studies should to consider alpha evaluation as a central point of measurement reliability, and yet, as much as possible, involve the assessment of internal consistency with other measures of reliability. In the sequence of internal consistency, the following methods were identified by this review: test-retest reliability (analysis of the temporal stability; items are applied on two separate occasions, and the scores could be correlated) (e.g., Forbush et al. 201346. Forbush, KT, Wildes, JE, Pollack, LO, Dunbar, D, Luo, J, Patterson, P, Petruzzi, L, … Watson, D. (2013). Development and validation of the Eating Pathology Symptoms Inventory (EPSI). Psychological Assessment, 25(3), 859-878. doi: http://dx.doi.org/10.1037/a0032639.
http://dx.doi.org/10.1037/a0032639...
), item-total/inter-item correlation reliability (analysis of the correlation of each item with the total score of the scale or subscales/analysis of the correlation of each item with another item) (e.g., Rodrigues and Bastos 2012105. Rodrigues, A. C. A., & Bastos, A. V. B. (2012). Organizational entrenchment: scale development and validation. Psicologia: Reflexão e Crítica, 25(4), 688–700. http://dx.doi.org/10.1590/S0102-79722012000400008.
http://dx.doi.org/10.1590/S0102-79722012...
), split-half reliability (the scale is split in half and the first half of the items are compared to the second half) (e.g., Uzunboylu and Ozdamli 2011126. Uzunboylu, H., & Ozdamli, F. (2011). Teacher perception for m-learning: scale development and teachers’ perceptions. Journal of Computer Assisted Learning, 27, 544–556. http://dx.doi.org/10.1111/j.1365-2729.2011.00415.x.
http://dx.doi.org/10.1111/j.1365-2729.20...
), and inter-judge reliability (analysis of the consistency between two different observers when they assess the same measure in the same individual) (e.g., Akter et al. 20133. Akter, S., D’Ambra, J., & Ray, P. (2013). Development and validation of an instrument to measure user perceived service quality of mHealth. Information and Management, 50, 181–195. http://dx.doi.org/10.1016/j.im.2013.03.001.
http://dx.doi.org/10.1016/j.im.2013.03.0...
; DeVellis 200338. DeVellis, R. F. (2003). Scale development: theory and applications (2nd ed.). Newbury Park: Sage Publications.; Nunnally 196788. Nunnally, J. C. (1967). Psychometric theory. New York: McGraw Hill.).

Regarding sample size in step 3 and number of items, a particularly noteworthy finding was that most studies utilized sample sizes smaller than the rule of thumb that the minimum required ratio should be 10:1 (e.g., Turker 2009125. Turker, D. (2009). Measuring corporate social responsibility: a scale development study. Journal of Business Ethics, 85, 411–427. http://dx.doi.org/10.1007/s10551-008-9780-6.
http://dx.doi.org/10.1007/s10551-008-978...
; Zheng et al. 2010137. Zheng, J, You, L, Lou, T, Chen, N, Lai, D, Liang, Y, … Zhai, C. (2010). Development and psychometric evaluation of the dialysis patient-perceived exercise benefits and barriers scale. International Journal of Nursing Studies, 47, 166-180. doi: http://dx.doi.org/10.1016/j.ijnurstu.2009.05.023
http://dx.doi.org/10.1016/j.ijnurstu.200...
). DeVellis (2003)38. DeVellis, R. F. (2003). Scale development: theory and applications (2nd ed.). Newbury Park: Sage Publications. and Hair Junior et al. (2009)54. Hair Junior, J. F., Black, W. C., Babin, N. J., Anderson, R. E., & Tatham, R. L. (2009). Análise multivariada de dados (6ath ed.). São Paulo: Bookman. comment that the sample size should be as large as possible to ensure factor stability. The ‘observations to variables’ ratio is ideal at 15:1, or even 20:1. However, most of the studies included in this review failed to adopt this rule. Some studies looked for justification on evidence related to the effectiveness of much smaller observations to variables ratios. For example, Nagy et al. (2014)84. Nagy, B. G., Blair, E. S., & Lohrke, F. T. (2014). Developing a scale to measure liabilities and assets of newness after start-up. International Entrepreneurship and Management Journal, 10, 277–295. http://dx.doi.org/10.1007/s11365-012-0219-2.
http://dx.doi.org/10.1007/s11365-012-021...
justified the small sample size used in their investigation based on the findings of Barrett and Kline (1981)11. Barrett, P. T., & Kline, P. (1981). The observation to variable ratio in factor analysis. Personality Study and Group Behavior, 1, 23–33., concluding that the difference in ratios 1.25:1 and 31:1 was not a significant contributor to results obtained in the factor stability. Additionally, Arrindell and van der Ende (1985)7. Arrindell, W. A., & van der Ende, J. (1985). An empirical-test of the utility of the observations-to-variables ratio in factor and components analysis. Applied Psychological Measurement, 9(2), 165–178. http://dx.doi.org/10.1177/014662168500900205.
http://dx.doi.org/10.1177/01466216850090...
concluded that ratios of 1.3:1 and 19.8:1 did not impact the factor stability. Although the rules of thumb vary enormously, ten participants to each item has widely been considered safe recommended (Sveinbjornsdottir and Thorsteinsson 2008120. Sveinbjornsdottir, S., & Thorsteinsson, E. B. (2008). Adolescent coping scales: a critical psychometric review. Scandinavian Journal of Psychology, 49(6), 533–548. http://dx.doi.org/10.1111/j.1467-9450.2008.00669.x.
http://dx.doi.org/10.1111/j.1467-9450.20...
).

Finally, several studies had their number final of items reduced by more than 50%. For example, Flight et al. (2011)45. Flight, R. L., D’Souza, G., & Allaway, A. W. (2011). Characteristics-based innovation adoption: scale and model validation. Journal of Product and Brand Management, 20(5), 343–355. http://dx.doi.org/10.1108/10610421111157874.
http://dx.doi.org/10.1108/10610421111157...
developed an initial item pool composed of 122 items and finished the scale with only 43. Pommer et al. (2013)97. Pommer, AM, Prins, L, van Ranst, D, Meijer, J, Hul, AV, Janssen, J, … Pop, VJM. (2013). Development and validity of the Patient-Centred COPD Questionnaire (PCQ). Journal of Psychosomatic Research, 75, 563-571. doi: http://dx.doi.org/10.1016/j.jpsychores.2013.10.001
http://dx.doi.org/10.1016/j.jpsychores.2...
developed 391 initial items and finished with only 18. Our findings clearly indicate that a significant amount of items can get lost during the development of a new scale. These results are consistent with previous literature which states both that the initial number of items must be twice the desired number in the final scale, since, during the process of analysis of the items, many may be excluded for inadequacy (Nunnally 196788. Nunnally, J. C. (1967). Psychometric theory. New York: McGraw Hill.), and that the initial set of items should be three or four times more numerous than the number of items desired, as a good way to ensure internal consistency of the scale (DeVellis 200338. DeVellis, R. F. (2003). Scale development: theory and applications (2nd ed.). Newbury Park: Sage Publications.). Future research should consider these issues and expect significant loss of items during the scale development process.

Ten main limitations reported in the scale development process—findings and research implications

In addition to identifying the current practices of the scale development process, this review also aims to assess the main limitations reported by the authors. Ten limitations were found, which will be discussed together with recommendations for best practices in future scale development research (Table 3).

Sample characteristic limitations The above-mentioned limitations were recorded in the majority of the studies, in two main ways. The first and the most representative way was related to the sample type. Several studies used homogeneous sampling (e.g., Forbush et al. 201346. Forbush, KT, Wildes, JE, Pollack, LO, Dunbar, D, Luo, J, Patterson, P, Petruzzi, L, … Watson, D. (2013). Development and validation of the Eating Pathology Symptoms Inventory (EPSI). Psychological Assessment, 25(3), 859-878. doi: http://dx.doi.org/10.1037/a0032639.
http://dx.doi.org/10.1037/a0032639...
; Morean et al. 201282. Morean, M. E., Corbin, W. R., & Treat, T. A. (2012). The anticipated effects of alcohol scale: development and psychometric evaluation of a novel assessment tool for measuring alcohol expectancies. Psychological Assessment, 24(4), 1008–1023. http://dx.doi.org/10.1037/a0028982.
http://dx.doi.org/10.1037/a0028982...
), whereas others used convenience sampling (e.g., Coker et al. 201132. Coker, B. L. S., Ashill, N. J., & Hope, B. (2011). Measuring internet product purchase risk. European Journal of Marketing, 45(7/8), 1130–1151. http://dx.doi.org/10.1108/03090561111137642.
http://dx.doi.org/10.1108/03090561111137...
; Flight et al. 201145. Flight, R. L., D’Souza, G., & Allaway, A. W. (2011). Characteristics-based innovation adoption: scale and model validation. Journal of Product and Brand Management, 20(5), 343–355. http://dx.doi.org/10.1108/10610421111157874.
http://dx.doi.org/10.1108/10610421111157...
). Both homogeneous and convenience samples were related to limitations of generalization. For example, Atkins and Kim (2012)8. Atkins, K. G., & Kim, Y. (2012). Smart shopping: conceptualization and measurement. International Journal of Retail and Distribution Management, 40(5), 360–375. http://dx.doi.org/10.1108/09590551211222349.
http://dx.doi.org/10.1108/09590551211222...
pointed out that “the participants for all stages of the study were US consumers; therefore, this study cannot be generalized to other cultural contexts.” Or indeed, “convenience samples are weaknesses of this study, as they pose generalizability questions,” as highlighted by Blankson et al. (2012)16. Blankson, C., Paswan, A., & Boakye, K. G. (2012). College students’ consumption of credit cards. International Journal of Bank Marketing, 30(7), 567–585. http://dx.doi.org/10.1108/02652321211274327.
http://dx.doi.org/10.1108/02652321211274...
. Nunnally (1967)88. Nunnally, J. C. (1967). Psychometric theory. New York: McGraw Hill. suggested that, to extend the generalizability of the new scale, sample diversification should be considered in terms of data collection, particularly in the psychometric evaluation step. Future studies should consider this suggestion, recruiting heterogeneous and truly random samples for the evaluation of construct validity and the reliability of the new measure.

The second way was related to small sample size. As previously described, most of the analyzed studies utilized sample sizes less than 10:1. Only some of the authors recognized this flaw. For example, Nagy et al. (2014)84. Nagy, B. G., Blair, E. S., & Lohrke, F. T. (2014). Developing a scale to measure liabilities and assets of newness after start-up. International Entrepreneurship and Management Journal, 10, 277–295. http://dx.doi.org/10.1007/s11365-012-0219-2.
http://dx.doi.org/10.1007/s11365-012-021...
reported that “the sample size employed in conducting the exploratory factor analysis is another potential limitation of the study,” Rosenthal (2011)107. Rosenthal, S. (2011). Measuring knowledge of indoor environmental hazards. Journal of Environmental Psychology, 31, 137–146. http://dx.doi.org/10.1016/j.jenvp.2010.08.003.
http://dx.doi.org/10.1016/j.jenvp.2010.0...
described, “the current study was limited by the relatively small nonprobability sample of university students,” and Ho and Lin (2010)63. Ho, C. B., & Lin, W. (2010). Measuring the service quality of internet banking: scale development and validation. European Business Review, 22(1), 5–24. http://dx.doi.org/10.1108/09555341011008981.
http://dx.doi.org/10.1108/09555341011008...
recognized that “the respondent sample size was small.” Based in these results, we emphasize that future research should seek a larger sample size (minimum ratio of 10:1) to increase the credibility of the results and thus obtain a more exact outcome in the psychometric analysis.

Methodological limitations Cross-sectional methods were the main methodological limitations reported by other studies (e.g., Schlosser and McNaughton 2009110. Schlosser, F. K., & McNaughton, R. B. (2009). Using the I-MARKOR scale to identify market-oriented individuals in the financial services sector. Journal of Services Marketing, 23(4), 236–248. http://dx.doi.org/10.1108/08876040910965575.
http://dx.doi.org/10.1108/08876040910965...
; Tombaugh et al. 2011124. Tombaugh, J. R., Mayfield, C., & Durand, R. (2011). Spiritual expression at work: exploring the active voice of workplace spirituality. International Journal of Organizational Analysis, 19(2), 146–170. http://dx.doi.org/10.1108/19348831111135083.
http://dx.doi.org/10.1108/19348831111135...
). Data collected under a cross-sectional study design contains the typical limitation associated with this type of research methodology, namely inability to determine the causal relationship. If cross-sectional methods are used to estimate models whose parameters do in fact vary over time, the resulting estimation may fail to yield statistically valid results, fail to identify the true model parameters, and produce inefficient estimates (Bowen and Wiersema 199919. Bowen, H. P., & Wiersema, M. F. (1999). Matching method to paradigm in strategy research: limitations of cross-sectional analysis and some methodological alternatives. Strategic Management Journal, 20, 625–636.). In this way, different authors (e.g., Akter et al. 20133. Akter, S., D’Ambra, J., & Ray, P. (2013). Development and validation of an instrument to measure user perceived service quality of mHealth. Information and Management, 50, 181–195. http://dx.doi.org/10.1016/j.im.2013.03.001.
http://dx.doi.org/10.1016/j.im.2013.03.0...
; Boyar et al. 201420. Boyar, S. L., Campbell, N. S., Mosley, D. C., Jr., & Carson, C. M. (2014). Development of a work/family social support measure. Journal of Managerial Psychology, 29(7), 901–920. http://dx.doi.org/10.1108/JMP-06-2012-0189.
http://dx.doi.org/10.1108/JMP-06-2012-01...
) recognized that employing instruments at one point in time limits the ability to assess causal relationships. With the goal of remediating these issues and gaining a deeper understanding of the construct of interest, different studies (e.g., Morean et al. 201282. Morean, M. E., Corbin, W. R., & Treat, T. A. (2012). The anticipated effects of alcohol scale: development and psychometric evaluation of a novel assessment tool for measuring alcohol expectancies. Psychological Assessment, 24(4), 1008–1023. http://dx.doi.org/10.1037/a0028982.
http://dx.doi.org/10.1037/a0028982...
; Schlosser and McNaughton 2009110. Schlosser, F. K., & McNaughton, R. B. (2009). Using the I-MARKOR scale to identify market-oriented individuals in the financial services sector. Journal of Services Marketing, 23(4), 236–248. http://dx.doi.org/10.1108/08876040910965575.
http://dx.doi.org/10.1108/08876040910965...
) suggest conducting a longitudinal study during the scale development. Using the longitudinal studies in this process may also allow the assessment of the scale’s predictive validity, since longitudinal designs evaluate whether the proposed interpretation of test scores can predict outcomes of interest over time. Therefore, future studies should consider the longitudinal approach in the scale development, both to facilitate greater understanding of the analyzed variables and to assess the predictive validity.

Self-reporting methodologies were also cited as limitations in some studies (e.g., Fisher et al. 201444. Fisher, R., Maritz, A., & Lobo, A. (2014). Evaluating entrepreneurs’ perception of success. International Journal of Entrepreneurial Behavior and Research, 20(5), 478–492. http://dx.doi.org/10.1108/IJEBR-10-2013-0157.
http://dx.doi.org/10.1108/IJEBR-10-2013-...
; Pan et al. 201392. Pan, J., Wong, D. F. K., & Ye, S. (2013). Post-migration growth scale for Chinese international students: development and validation. Journal of Happiness Studies, 14, 1639–1655. http://dx.doi.org/10.1007/s10902-012-9401-z.
http://dx.doi.org/10.1007/s10902-012-940...
). Mahudin et al. (2012)78. Mahudin, N. D. M., Cox, T., & Griffiths, A. (2012). Measuring rail passenger crowding: scale development and psychometric properties. Transportation Research Part, F 15, 38–51. http://dx.doi.org/10.1016/j.trf.2011.11.006.
http://dx.doi.org/10.1016/j.trf.2011.11....
clarified that the self-reporting nature of quantitative studies raises the possibility of participant bias, social desirability, demand characteristics, and response sets. Such possibilities may, in turn, affect the validity of the findings. We agree with the authors’ suggestion that future research may also incorporate other objective or independent measures to supplement the subjective evaluation of the variables studied in the development of the new scale and to improve the interpretation of findings.

In addition, web-based surveys were another methodological limitation reported in some studies (e.g., Kim et al. 201169. Kim, S., Cha, J., Knutson, B. J., & Beck, J. A. (2011). Development and testing of the Consumer Experience Index (CEI). Managing Service Quality: An International Journal, 21(2), 112–132. http://dx.doi.org/10.1108/09604521111113429.
http://dx.doi.org/10.1108/09604521111113...
; Reed et al. 2011100. Reed, L. L., Vidaver-Cohen, D., & Colwell, S. R. (2011). A new scale to measure executive servant leadership: development, analysis, and implications for research. Journal of Business Ethics, 101, 415–434. http://dx.doi.org/10.1007/s10551-010-0729-1.
http://dx.doi.org/10.1007/s10551-010-072...
). Although this particular method has time- and cost-saving elements for data collection, its limitations are also highlighted. Researchers have observed that important concerns include coverage bias (bias due to sampled individuals not having—or choosing not to access—the Internet) and nonresponse bias (bias due to participants of a survey differing from those who did not respond in terms of demographic or attitudinal variables) (Kim et al. 201169. Kim, S., Cha, J., Knutson, B. J., & Beck, J. A. (2011). Development and testing of the Consumer Experience Index (CEI). Managing Service Quality: An International Journal, 21(2), 112–132. http://dx.doi.org/10.1108/09604521111113429.
http://dx.doi.org/10.1108/09604521111113...
). Alternatives to minimize the problem in future research would be in-person surveys or survey interviews. Although more costly and more time consuming, these methods reduce problems related to concerns about confidentiality and the potential for coverage and nonresponse bias (Reed et al. 2011100. Reed, L. L., Vidaver-Cohen, D., & Colwell, S. R. (2011). A new scale to measure executive servant leadership: development, analysis, and implications for research. Journal of Business Ethics, 101, 415–434. http://dx.doi.org/10.1007/s10551-010-0729-1.
http://dx.doi.org/10.1007/s10551-010-072...
). Therefore, whenever possible, in-person surveys or survey interviews should be given priority in future research rather than web surveys.

Psychometric limitations Consistent with previous reports (MacKenzie et al. 201177. MacKenzie, S. B., Podsakoff, P. M., & Podsakoff, N. P. (2011). Construct measurement and validation procedures in MIS and behavioral research: integrating new and existing techniques. MIS Quarterly, 35(2), 293–334.; Prados 200798. Prados, J. M. (2007). Development of a new scale of beliefs about the worry consequences. Annals of Psychology, 23(2), 226–230.), this systematic review found distinct psychometric limitations reported in the scale development process. The lack of a more robust demonstration of construct validity and/or reliability was the most often mentioned limitation in the majority of the analyzed studies. For example, Alvarado-Herrera et al. (2015)4. Alvarado-Herrera, A, Bigne, E, Aldas-Manzano, J, & Curras-Perez, R. (2015). A scale for measuring consumer perceptions of corporate social responsibility following the sustainable development paradigm. Journal of Business Ethics, 1-20. doi: http://dx.doi.org/10.1007/s10551-015-2654-9
http://dx.doi.org/10.1007/s10551-015-265...
reported the lack of a more robust demonstration of the predictive validity whereas Kim et al. (2011)69. Kim, S., Cha, J., Knutson, B. J., & Beck, J. A. (2011). Development and testing of the Consumer Experience Index (CEI). Managing Service Quality: An International Journal, 21(2), 112–132. http://dx.doi.org/10.1108/09604521111113429.
http://dx.doi.org/10.1108/09604521111113...
of the nomological validity. Caro and Garcia (2007)24. Caro, L. M., & García, J. A. M. (2007). Measuring perceived service quality in urgent transport service. Journal of Retailing and Consumer Services, 14, 60–72. http://dx.doi.org/10.1016/j.jretconser.2006.04.001.
http://dx.doi.org/10.1016/j.jretconser.2...
noted that the relationships of the scale with other constructs were not analyzed. Saxena et al. (2015)108. Saxena, S., Ayers, C. R., Dozier, M. E., & Maidment, K. M. (2015). The UCLA Hoarding Severity Scale: development and validation. Journal of Affective Disorders, 175, 488–493. http://dx.doi.org/10.1016/j.jad.2015.01.030.
http://dx.doi.org/10.1016/j.jad.2015.01....
and Pan et al. (2013)92. Pan, J., Wong, D. F. K., & Ye, S. (2013). Post-migration growth scale for Chinese international students: development and validation. Journal of Happiness Studies, 14, 1639–1655. http://dx.doi.org/10.1007/s10902-012-9401-z.
http://dx.doi.org/10.1007/s10902-012-940...
described the lack of demonstrable temporal stability (e.g., test-retest reliability). Imprecise or incomplete psychometric procedures that are employed during scale development are likely to obscure the outcome. Therefore, it is necessary for future research to consider adverse consequences for the reliability and validity of any construct, caused by poor test-theoretical practices. Only through detailed information and explanation of the rationale for statistical choices can the new measures be shown to have sufficient psychometric adjustments (Sveinbjornsdottir and Thorsteinsson 2008120. Sveinbjornsdottir, S., & Thorsteinsson, E. B. (2008). Adolescent coping scales: a critical psychometric review. Scandinavian Journal of Psychology, 49(6), 533–548. http://dx.doi.org/10.1111/j.1467-9450.2008.00669.x.
http://dx.doi.org/10.1111/j.1467-9450.20...
).

Additionally, the inadequate choice of the instruments or variables to be correlated with the variable of interest was another psychometric limitation cited in some studies (e.g., Bakar and Mustaffa 201310. Bakar, H. A., & Mustaffa, C. S. (2013). Organizational communication in Malaysia organizations. Corporate Communications: An International Journal, 18(1), 87–109. http://dx.doi.org/10.1108/13563281311294146.
http://dx.doi.org/10.1108/13563281311294...
; Tanimura et al. 2011122. Tanimura, C., Morimoto, M., Hiramatsu, K., & Hagino, H. (2011). Difficulties in the daily life of patients with osteoarthritis of the knee: scale development and descriptive study. Journal of Clinical Nursing, 20, 743–753. http://dx.doi.org/10.1111/j.1365-2702.2010.03536.x.
http://dx.doi.org/10.1111/j.1365-2702.20...
). This kind of limitation directly affects the convergent validity, which is a problem since, as has already been shown in this review, this type of validity has been one of the most recurrent practices in scale development. One hypothesis for this limitation may be the lack of gold standard measures to assess similar constructs as those of a new scale. In such cases, a relatively recent study by Morgado et al. (2014)83. Morgado, F. F. R., Campana, A. N. N. B., & Tavares, M. C. G. C. F. (2014). Development and validation of the self-acceptance scale for persons with early blindness: the SAS-EB. PloS One, 9(9), e106848. http://dx.doi.org/10.1371/journal.pone.0106848.
http://dx.doi.org/10.1371/journal.pone.0...
offers a valid alternative. The authors used information collected on sociodemographic questionnaires (e.g., level of education and intensity of physical activity) to correlate with the constructs of interest. Future researchers should seek support from the literature on the constructs that would be theoretically associated with the construct of interest, searching for alternatives in information collected on, for example, sociodemographic questionnaires, to assess the convergent validity of the new scale.

Another psychometric limitation reported in some studies was related to factor analysis. These limitations were identified in five main forms: (1) EFA and CFA were conducted using the data from the same sample (Zheng et al. 2010137. Zheng, J, You, L, Lou, T, Chen, N, Lai, D, Liang, Y, … Zhai, C. (2010). Development and psychometric evaluation of the dialysis patient-perceived exercise benefits and barriers scale. International Journal of Nursing Studies, 47, 166-180. doi: http://dx.doi.org/10.1016/j.ijnurstu.2009.05.023
http://dx.doi.org/10.1016/j.ijnurstu.200...
)—when this occurs, good model fit in the CFA is expected, as a consequence, the added strength of the CFA in testing a hypothesized structure for a new data set based on theory or previous findings is lost (Khine 200867. Khine, M. S. (2008). Knowing, knowledge and beliefs: epistemological studies across diverse cultures. New York: Springer.); (2) lack of CFA (Bolton and Lane 201217. Bolton, D. L., & Lane, M. D. (2012). Individual entrepreneurial orientation: development of a measurement instrument. Education + Training, 54(2/3), 219–233. http://dx.doi.org/10.1108/00400911211210314.
http://dx.doi.org/10.1108/00400911211210...
)—if this happens, the researcher loses the possibility of assigning items to factors, testing the hypothesized structure of the data, and statistically comparing alternative models (Khine 200867. Khine, M. S. (2008). Knowing, knowledge and beliefs: epistemological studies across diverse cultures. New York: Springer.); (3) a certain amount of subjectivity was necessary in identifying and labeling factors in EFA (Lombaerts et al. 200975. Lombaerts, K., Backer, F., Engels, N., Van Braak, J., & Athanasou, J. (2009). Development of the self-regulated learning teacher belief scale. European Journal of Psychology of Education, 24(1), 79–96. http://dx.doi.org/10.1007/BF03173476.
http://dx.doi.org/10.1007/BF03173476...
)—since a factor is qualitative, it is common practice to label each factor based on an interpretation of the variables loading most heavily on it; the problem is that these labels are subjective in nature, represent the authors’ interpretation, and can vary typically from 0.30 to 0.50 (Gottlieb et al. 201453. Gottlieb, U., Brown, M., & Ferrier, L. (2014). Consumer perceptions of trade show effectiveness. European Journal of Marketing, 48(1/2), 89–107. http://dx.doi.org/10.1108/EJM-06-2011-0310.
http://dx.doi.org/10.1108/EJM-06-2011-03...
; Khine 200867. Khine, M. S. (2008). Knowing, knowledge and beliefs: epistemological studies across diverse cultures. New York: Springer.); (4) the initial unsatisfactory factor analysis output (Lombaerts et al. 200975. Lombaerts, K., Backer, F., Engels, N., Van Braak, J., & Athanasou, J. (2009). Development of the self-regulated learning teacher belief scale. European Journal of Psychology of Education, 24(1), 79–96. http://dx.doi.org/10.1007/BF03173476.
http://dx.doi.org/10.1007/BF03173476...
); and (5) lack of a more robust CFA level (Jong et al. 201465. Jong, N., Van Leeuwen, R. G. J., Hoekstra, H. A., & van der Zee, K. I. (2014). CRIQ: an innovative measure using comparison awareness to avoid self-presentation tactics. Journal of Vocational Behavior, 84, 199–214. http://dx.doi.org/10.1016/j.jvb.2014.01.003.
http://dx.doi.org/10.1016/j.jvb.2014.01....
) taken together—when the study result distances itself from statistical results expected for EFA (e.g., KMO, Bartlett test of sphericity) and/or CFA (e.g., CFI, GFI, RMSEA), it results in an important limitation, since the tested exploratory and theoretical models are not considered valid (Khine 200867. Khine, M. S. (2008). Knowing, knowledge and beliefs: epistemological studies across diverse cultures. New York: Springer.). Taking these results, future studies should consider the use of separate samples for EFA and CFA, the combination of EFA and CFA, the definition of objective parameters to label factors, and about the consideration for unsatisfactory results of EFA and CFA, seeking alternatives to better fit the model.

Qualitative research limitations This review also found reported limitations on the qualitative approach of the analyzed studies. The first limitation was related to the exclusive use of the deductive method to generate items. It is noteworthy that, although most of the studies included in this review used exclusively deductive methods to generate items, only two studies recognized this as a limitation (Coleman et al. 201133. Coleman, D., Chernatony, L., & Christodoulides, G. (2011). B2B service brand identity: scale development and validation. Industrial Marketing Management, 40, 1063–1071. http://dx.doi.org/10.1016/j.indmarman.2011.09.010.
http://dx.doi.org/10.1016/j.indmarman.20...
; Song et al. 2011117. Song, J. H., Uhm, D., & Yoon, S. W. (2011). Organizational knowledge creation practice. Leadership and Organization Development Journal, 32(3), 243–259. http://dx.doi.org/10.1108/01437731111123906.
http://dx.doi.org/10.1108/01437731111123...
). Both studies used only the literature review to generate and operationalize the initial item pool. The authors recognized the importance of this deductive method to theoretically operationalize the target construct, but they noted that, “for further research, more diverse views should be considered to reflect more comprehensive perspectives of human knowledge-creating behaviors to strengthen the validity of the developed scales” (Song et al. 2011117. Song, J. H., Uhm, D., & Yoon, S. W. (2011). Organizational knowledge creation practice. Leadership and Organization Development Journal, 32(3), 243–259. http://dx.doi.org/10.1108/01437731111123906.
http://dx.doi.org/10.1108/01437731111123...
, p. 256) and, “a qualitative stage could have been used to generate additional items […]. This could also have reduced measurement error by using specific language the population used to communicate” (Coleman et al. 201133. Coleman, D., Chernatony, L., & Christodoulides, G. (2011). B2B service brand identity: scale development and validation. Industrial Marketing Management, 40, 1063–1071. http://dx.doi.org/10.1016/j.indmarman.2011.09.010.
http://dx.doi.org/10.1016/j.indmarman.20...
; p. 1069). Thus, the combination of deductive and inductive approaches (e.g., focus groups or interviews) in item generation is again suggested in future research.

In addition, it is also necessary that the researcher consider the quality of the reviewed literature. Napoli et al. (201485. Napoli, J., Dickinson, S. J., Beverland, M. B., & Farrelly, F. (2014). Measuring consumer-based brand authenticity. Journal of Business Research, 67, 1090–1098. http://dx.doi.org/10.1016/j.jbusres.2013.06.001.
http://dx.doi.org/10.1016/j.jbusres.2013...
, p. 1096) reported limitations related to the loss of a more robust literature review, suggesting that the scale developed in the study may have been incorrectly operationalized: “Yet some question remains as to whether cultural symbolism should form part of this scale. Perhaps the way in which the construct was initially conceptualized and operationalized was incorrect.” The incorrect operation of the construct compromises the psychometric results of scale and its applicability in future studies.

Another limitation involves the subjective analysis of the qualitative research. Fisher et al. (201444. Fisher, R., Maritz, A., & Lobo, A. (2014). Evaluating entrepreneurs’ perception of success. International Journal of Entrepreneurial Behavior and Research, 20(5), 478–492. http://dx.doi.org/10.1108/IJEBR-10-2013-0157.
http://dx.doi.org/10.1108/IJEBR-10-2013-...
, p. 488) pointed out that the qualitative methods (literature reviews and interviews) used to develop and conceptualize the construct were the main weaknesses of the study, “this research is limited by […] the nature of qualitative research in which the interpretations of one researcher may not reflect those of another.” The authors explained that, due to the potential for researcher bias when interpreting data, it has been recognized that credible results are difficult to achieve. Nevertheless, subjective analysis is the essence and nature of qualitative studies. Some precautions in future studies can be taken to rule out potential researcher bias, such as attempts at neutrality. This is not always possible, however, and this limitation will remain a common problem in any qualitative study.

In turn, Sewitch et al. (2003111. Sewitch, M. J., Abrahamowicz, M., Dobkin, P. L., & Tamblyn, R. (2003). Measuring differences between patients’ and physicians’ health perceptions: the patient–physician discordance scale. Journal of Behavioral Medicine, 26(3), 245–263. http://dx.doi.org/10.1023/A:1023412604715.
http://dx.doi.org/10.1023/A:102341260471...
, p. 260) reported that failure to formally assess content validity was a limitation. The reason given was budgetary constraints. It is worthwhile to remember that the content validity is an important step to ensure confidence in any inferences made using the final scale form. Therefore, it is necessarily required in any scale development process.

An additional limitation was reported by Lucas-Carrasco et al. (2011)76. Lucas-Carrasco, R., Eser, E., Hao, Y., McPherson, K. M., Green, A., & Kullmann, L. (2011). The quality of care and support (QOCS) for people with disability scale: development and psychometric properties. Research in Developmental Disabilities, 32, 1212–1225. http://dx.doi.org/10.1016/j.ridd.2010.12.030.
http://dx.doi.org/10.1016/j.ridd.2010.12...
in the recruitment of a larger number of interviewers, which may have affected the quality of the data collected. In order to minimize this limitation, the authors reported, “all interviewers had sufficient former education, received training on the study requirements, and were provided with a detailed guide” (p. 1223). Future studies planning the use of multiple interviewers should consider potential resulting bias.

Missing data In connection, missing data was another issue reported by some studies included in this systematic review (e.g., Glynn et al. 201552. Glynn, N. W., Santanasto, A. J., Simonsick, E. M., Boudreau, R. M., Beach, S. R., Schulz, R., & Newman, A. B. (2015). The Pittsburgh fatigability scale for older adults: development and validation. Journal of American Geriatrics Society, 63, 130–135. http://dx.doi.org/10.1111/jgs.13191.
http://dx.doi.org/10.1111/jgs.13191...
; Ngorsuraches et al. 200787. Ngorsuraches, S., Lerkiatbundit, S., Li, S. C., Treesak, C., Sirithorn, R., & Korwiwattanakarn, M. (2007). Development and validation of the patient trust in community pharmacists (TRUST-Ph) scale: results from a study conducted in Thailand. Research in Social and Administrative Pharmacy, 4, 272–283. http://dx.doi.org/10.1016/j.sapharm.2007.10.002.
http://dx.doi.org/10.1016/j.sapharm.2007...
). Such limitations typically occur across different fields of scientific research. Missing data includes numbers that have been grouped, aggregated, rounded, censored, or truncated, resulting in partial loss of information (Schafer and Graham 2002109. Schafer, J. L., & Graham, J. W. (2002). Missing data: our view of the state of the Art. Psychological Methods, 7(2), 147–177. http://dx.doi.org/10.1037//1082-989X.7.2.147.
http://dx.doi.org/10.1037//1082-989X.7.2...
). Collins et al. (2001)34. Collins, L. M., Schafer, J. L., & Kam, C.-M. (2001). A comparison of inclusive and restrictive strategies in modern missing data procedures. Psychological Methods, 6(4), 330–351. http://dx.doi.org/10.1037/1082-989X.6.4.330.
http://dx.doi.org/10.1037/1082-989X.6.4....
clarified that when researchers are confronted with missing data, they run an increased risk of reaching incorrect conclusions. This is because missing data may bias parameter estimates, inflate type I and type II error rates, and degrade the performance of confidence intervals. The authors also explained that, “because a loss of data is nearly always accompanied by a loss of information, missing values may dramatically reduce statistical power” (p. 330). Therefore, future researchers who wish to mitigate these risks during the scale development must pay close attention to the missing data aspect of the analysis and choose their strategy carefully.

Statistical methods to solve the problem of missing data have improved significantly, as demonstrated by Schafer and Graham (2002)109. Schafer, J. L., & Graham, J. W. (2002). Missing data: our view of the state of the Art. Psychological Methods, 7(2), 147–177. http://dx.doi.org/10.1037//1082-989X.7.2.147.
http://dx.doi.org/10.1037//1082-989X.7.2...
, although misconceptions still remain abundant. Several methods to deal with missing data were reviewed, issues raised, and advice offered for those that remain unresolved. Considering the fact that a more detailed discussion of the statistics dealing with missing data is beyond of the scope of this article, more details about missing data analysis can be found in Schafer and Graham (2002)109. Schafer, J. L., & Graham, J. W. (2002). Missing data: our view of the state of the Art. Psychological Methods, 7(2), 147–177. http://dx.doi.org/10.1037//1082-989X.7.2.147.
http://dx.doi.org/10.1037//1082-989X.7.2...
.

Social desirability bias Another limitation reported in some studies (Bova et al. 200618. Bova, C., Fennie, K. P., Watrous, E., Dieckhaus, K., & Williams, A. B. (2006). The health care relationship (HCR) trust scale: development and psychometric evaluation. Research in Nursing and Health, 29, 477–488. http://dx.doi.org/10.1002/nur.20158.
http://dx.doi.org/10.1002/nur.20158...
; Ngorsuraches et al. 200787. Ngorsuraches, S., Lerkiatbundit, S., Li, S. C., Treesak, C., Sirithorn, R., & Korwiwattanakarn, M. (2007). Development and validation of the patient trust in community pharmacists (TRUST-Ph) scale: results from a study conducted in Thailand. Research in Social and Administrative Pharmacy, 4, 272–283. http://dx.doi.org/10.1016/j.sapharm.2007.10.002.
http://dx.doi.org/10.1016/j.sapharm.2007...
) and identified in this systematic review is social desirability bias. This type of bias is considered to be a systematic error in self-reporting measures resulting from the desire of respondents to avoid embarrassment and project a favorable image to others (Fisher 199343. Fisher, R. J. (1993). Social desirability bias and the validity of indirect questioning. Journal of Consumer Research, 20(2), 303–315. http://dx.doi.org/10.1086/209351.
http://dx.doi.org/10.1086/209351...
). According to King and Bruner (2000)71. King, M. F., & Bruner, G. C. (2000). Social desirability bias: a neglected aspect of validity testing. Psychology and Marketing, 17(2), 79–103. http://dx.doi.org/10.1002/(SICI)1520-6793(200002)17:2<79::AID-MAR2>3.0.CO;2-0.
http://dx.doi.org/10.1002/(SICI)1520-679...
, social desirability bias is an important threat to the validity of research employing multi-item scales. Provision of socially desirable responses in self-reported data may lead to spurious correlations between variables, as well as the suppression or moderation of relationships between the constructs of interest. Thus, one aspect of scale validity, which should be of particular concern to researchers, is the potential threat of contamination due to social-desirability response bias. To remedy this problem, we agree with the authors that it is incumbent upon researchers to identify situations in which data may be systematically biased toward the respondents’ perceptions of what is socially acceptable, to determine the extent to which this represents contamination of the data, and to implement the most appropriate methods of control. Details on methods for identifying, testing for, and/or preventing social desirability bias are beyond the scope of this article, but can be found at King and Bruner (2000)71. King, M. F., & Bruner, G. C. (2000). Social desirability bias: a neglected aspect of validity testing. Psychology and Marketing, 17(2), 79–103. http://dx.doi.org/10.1002/(SICI)1520-6793(200002)17:2<79::AID-MAR2>3.0.CO;2-0.
http://dx.doi.org/10.1002/(SICI)1520-679...
.

Item limitations In comparison with at least one previous study (Prados 200798. Prados, J. M. (2007). Development of a new scale of beliefs about the worry consequences. Annals of Psychology, 23(2), 226–230.), our findings reflect some potential item limitations. Firstly, items that were ambiguous or difficult to answer were the main weaknesses reported by Gottlieb et al. (2014)53. Gottlieb, U., Brown, M., & Ferrier, L. (2014). Consumer perceptions of trade show effectiveness. European Journal of Marketing, 48(1/2), 89–107. http://dx.doi.org/10.1108/EJM-06-2011-0310.
http://dx.doi.org/10.1108/EJM-06-2011-03...
. On this issue, the literature dealing with the necessary caution in wording the items is extensive. For example, items must clearly define the problem being addressed, must be as simple as possible, express a single idea, and use common words that reflect the vocabulary level of the target population. Items should not be inductors or have alternative or underlying assumptions. They must be free of generalizations and estimates, and be written to ensure the variability of responses. In writing the items, the researcher should avoid using fashionable expressions and colloquialisms or other words or phrases that impair understanding for groups of varying ages, ethnicities, religions, or genders. Furthermore, the items should be organized properly. For example, the opening questions should be simple and interesting to win the trust of the subjects. The most delicate, complex, or dull questions should be asked at the end of the sequence (Clark and Watson 199531. Clark, L. A., & Watson, D. (1995). Constructing validity: basic issues in objective scale development. Psychological Assessment, 7(3), 309–319. http://dx.doi.org/10.1037/1040-3590.7.3.309.
http://dx.doi.org/10.1037/1040-3590.7.3....
; Malhotra 200479. Malhotra, N. K. (2004). Pesquisa de marketing: Uma orientação aplicada (4ath ed.). Porto Alegre: Bookman.; Pasquali 201093. Pasquali, L. (2010). Instrumentação psicológica: fundamentos e práticas. Porto Alegre: Artmed.).

Furthermore, Cicero et al. (2010)30. Cicero, D. C., Kerns, J. G., & McCarthy, D. M. (2010). The Aberrant Salience Inventory: a new measure of psychosis proneness. Psychological Assessment, 22(3), 688–701. http://dx.doi.org/10.1037/a0019913.
http://dx.doi.org/10.1037/a0019913...
reported that the main limitation of their study was the fact that none of the items were reverse-scored. Although some methodologists claim that reverse scoring is necessary to avoid acquiescence among participants, this advice should be taken with caution. There are reports that the reverse-scored items may be confusing to participants, that the opposite of a construct reverse-scored may be fundamentally different than the construct, that reverse-scored items tend to be the worst fitting items in factor analyses, or that the factor structure of scales includes a factor with straightforward wording compared to a reverse-scored factor (Cicero et al. 201030. Cicero, D. C., Kerns, J. G., & McCarthy, D. M. (2010). The Aberrant Salience Inventory: a new measure of psychosis proneness. Psychological Assessment, 22(3), 688–701. http://dx.doi.org/10.1037/a0019913.
http://dx.doi.org/10.1037/a0019913...
). Awareness of these issues is necessary for future researchers to choose between avoiding acquiescence among participants or preventing a number of other problems related to the use of reverse scores.

Brevity of the scale Limitations on the scale size were also identified in this review. Studies by Negra and Mzoughi (2012)86. Negra, A., & Mzoughi, M. N. (2012). How wise are online procrastinators? A scale development. Internet Research, 22(4), 426–442. http://dx.doi.org/10.1108/10662241211250971.
http://dx.doi.org/10.1108/10662241211250...
and Tombaugh et al. (2011)124. Tombaugh, J. R., Mayfield, C., & Durand, R. (2011). Spiritual expression at work: exploring the active voice of workplace spirituality. International Journal of Organizational Analysis, 19(2), 146–170. http://dx.doi.org/10.1108/19348831111135083.
http://dx.doi.org/10.1108/19348831111135...
mentioned the short version of the scale as their main limitation. In both studies, the final version of the new scale included only five items. Generally, short scales are good, because they require less time from respondents. However, very short scales can in fact seriously compromise the reliability of the instrument (Raykov 200899. Raykov, T. (2008). Alpha if item deleted: a note on loss of criterion validity in scale development if maximizing coefficient alpha. British Journal of Mathematical and Statistical Psychology, 61, 275–285. http://dx.doi.org/10.1348/000711007X188520.
http://dx.doi.org/10.1348/000711007X1885...
). To the extent that the researcher removes items of the scale, the Cronbach’s alpha tends to decrease. It is valuable to remember that the minimum acceptable alpha should be at least 0.7, while an alpha value between 0.8 and 0.9 is considered ideal. Scales with many items tend to be more reliable, with higher alpha values (DeVellis 200338. DeVellis, R. F. (2003). Scale development: theory and applications (2nd ed.). Newbury Park: Sage Publications.). In this context, future researchers should prioritize scales with enough items to keep the alpha within the acceptable range. Although many items may be lost during theoretical and psychometric analysis, an alternative already mentioned in this study would be to begin the initial item pool with at least twice the desired items of the final scale.

Difficulty controlling all variables In addition to all limitations reported, Gottlieb et al. (2014)53. Gottlieb, U., Brown, M., & Ferrier, L. (2014). Consumer perceptions of trade show effectiveness. European Journal of Marketing, 48(1/2), 89–107. http://dx.doi.org/10.1108/EJM-06-2011-0310.
http://dx.doi.org/10.1108/EJM-06-2011-03...
mentioned a common limitation in different research fields—the difficulty of controlling all the variables that could influence the central construct of the study. The authors reported that “it may be that there are other variables that influence visitors’ perception of trade show effectiveness that were not uncovered in the research” and suggest “future research might yield insights that are not provided here” (p. 104). The reported limitation calls attention to the importance of the first step—item generation—in the scale development process. A possible remedy to this issue would be to know the target construct in detail during the item generation, allowing for all possible and important variables to be investigated and controlled. However, this is not always possible. Even using inductive and deductive approaches to generate items (literature review and interview), the authors still reported that limitation. In this light, future researchers must use care in hypothesizing and testing potential variables that could be controlled during construction of the scale development process.

Lack of manual instructions Finally, this review found a weakness reported on the loss of manualized instructions that regulate the data analysis. Saxena et al. (2015108. Saxena, S., Ayers, C. R., Dozier, M. E., & Maidment, K. M. (2015). The UCLA Hoarding Severity Scale: development and validation. Journal of Affective Disorders, 175, 488–493. http://dx.doi.org/10.1016/j.jad.2015.01.030.
http://dx.doi.org/10.1016/j.jad.2015.01....
, p. 492) pointed out that the initial version of the new scale “did not contain manualized instructions for raters, so it lacked objective anchor points for choosing specific ratings on many of its questions”. Therefore, an important detail that should have the attention of future researchers are instructions that determine the application methods of the new scale. Pasquali (2010)93. Pasquali, L. (2010). Instrumentação psicológica: fundamentos e práticas. Porto Alegre: Artmed. suggests that when drafting the instructions, the researcher should define the development of operational strategies that will enable the application of the instrument and the format in which it will be presented and decide both how the subject’s response will be given for each item and the way that the respondent should answer each item. The researcher should also define how the scale scores would be analyzed. In addition, the instructions need to be as short as possible without confusion to the subjects of the target population, should contain one or more examples of how the items should be answered, and should ensure that the subject is free of any related tension or anxiety.

Study limitations and strengths

This review itself is subject to some limitations that should be taken into consideration. First, during the selection of the articles included in the analysis, we may have missed some studies that could have been identified by using other terms related to “scale development.” This may have impacted our findings. However, application of this term alone was recommended by its widespread use by researchers in the area (Clark and Watson 199531. Clark, L. A., & Watson, D. (1995). Constructing validity: basic issues in objective scale development. Psychological Assessment, 7(3), 309–319. http://dx.doi.org/10.1037/1040-3590.7.3.309.
http://dx.doi.org/10.1037/1040-3590.7.3....
; DeVellis 200338. DeVellis, R. F. (2003). Scale development: theory and applications (2nd ed.). Newbury Park: Sage Publications.; Hinkin 199562. Hinkin, T. R. (1995). A review of scale development practices in the study of organizations. Journal of Management, 21(5), 967–988. http://dx.doi.org/10.1177/014920639502100509.
http://dx.doi.org/10.1177/01492063950210...
; Nunnally 196788. Nunnally, J. C. (1967). Psychometric theory. New York: McGraw Hill.) and by the large number of publications identified with this descriptor in the period evaluated, as compared with those screened with correlates (e.g., “development of questionnaire” and “development of measure”). In the same way, we may also have missed numerous studies that, despite recording their weaknesses, did not have the search term “limitations” indexed in the analyzed databases. We could have reduced this limitation by also using the search term ‘weakness’ or a similar word for selection and inclusion of several other articles. However, a larger number of included studies would hinder the operationalization of our findings.

Second, particularly regarding analysis of items and reliability, we lost information about the basic theories that support the scale development process: classical test theory (CTT)—known as classical psychometry—and item response theory (IRT)—known as modern psychometry (PASQUALI 201093. Pasquali, L. (2010). Instrumentação psicológica: fundamentos e práticas. Porto Alegre: Artmed.). Although it was beyond the scope of this article to examine these theories, information on the employability of one or the other could contribute to a deeper understanding of their main limitations. Future studies could focus on CTT and IRT, compare the applicability of both, and identify their main limitations in the scale development process.

Still, our review is current with studies published until September 2015. As new evidence emerges on current practices and limitations reported in the scale development process, revisions to this systematic review and practice guideline would be required in future studies.

Despite its weaknesses, the strengths of this study should be highlighted. First, this study reviews the updated and consistent literature on scale development practices to be applied in, not only a specific field of knowledge as carried out in most systematic review studies, but across various fields. With this variety of conceptions, we hope to assist future researchers in different areas of human and social sciences in making the most appropriate choice between strategies.

Second, this study differs from most studies of scale development revision, since it primarily considers the conceptions of the authors themselves about the main difficulties and mistakes made during the scale development process in their own studies. We hope to contribute to the efforts of future researchers, based on the knowledge of previous mistakes. While several weaknesses in scale development research were identified, specific recommendations for future research relevant to particular previously dimensions discussed were embedded within the appropriate sections throughout the article.

We observe that, although some weaknesses have been clearly identified in the scale development practices of many studies, only a few researchers recognized and recorded these limitations. This was evidenced in the large number of studies using exclusively deductive approaches to generate the initial item pool and the limited number of studies that recognized this as a limitation, or there were a large number of studies using smaller sample sizes than recommended in the literature for psychometric analysis and the limited number of studies that reported this issue as a limitation. Considering the observed distance between the limitation and its recognition, it is important that future researchers are comfortable with the detailed process of developing a new measure, especially as it pertains to avoiding theoretical and/or methodological mistakes, or at least, if they occur, to mention them as limitations.

Conclusions

In conclusion, the present research reviews numerous studies that both proposed current practices of the scale development process and also reported its main limitations. A variety of conceptions and methodological strategies and ten mains limitations were identified and discussed along with suggestions for future research. In this way, we believe that this paper makes important contributions to the literature, especially because it provides a comprehensive set of recommendations to increase the quality of future practices in the scale development process.

References

  • 1
    Aagja, J. P., & Garg, R. (2010). Measuring perceived service quality for public hospitals (PubHosQual) in the Indian context. International Journal of Pharmaceutical and Healthcare Marketing, 4(10), 60–83. http://dx.doi.org/10.1108/17506121011036033
    » http://dx.doi.org/10.1108/17506121011036033
  • 2
    Ahmad, N., Awan, M. U., Raouf, A., & Sparks, L. (2009). Development of a service quality scale for pharmaceutical supply chains. International Journal of Pharmaceutical and Healthcare Marketing, 3(1), 26–45. http://dx.doi.org/10.1108/17506120910948494
    » http://dx.doi.org/10.1108/17506120910948494
  • 3
    Akter, S., D’Ambra, J., & Ray, P. (2013). Development and validation of an instrument to measure user perceived service quality of mHealth. Information and Management, 50, 181–195. http://dx.doi.org/10.1016/j.im.2013.03.001
    » http://dx.doi.org/10.1016/j.im.2013.03.001
  • 4
    Alvarado-Herrera, A, Bigne, E, Aldas-Manzano, J, & Curras-Perez, R. (2015). A scale for measuring consumer perceptions of corporate social responsibility following the sustainable development paradigm. Journal of Business Ethics, 1-20. doi: http://dx.doi.org/10.1007/s10551-015-2654-9
    » http://dx.doi.org/10.1007/s10551-015-2654-9
  • 5
    Arias, M. R. M., Lloreda, M. J. H., & Lloreda, M. V. H. (2014). Psicometría S.A.: Alianza Editorial
  • 6
    Armfield, J. M. (2010). Development and psychometric evaluation of the Index of Dental Anxiety and Fear (IDAF-4C+). Psychological Assessment, 22(2), 279–287. http://dx.doi.org/10.1037/a0018678
    » http://dx.doi.org/10.1037/a0018678
  • 7
    Arrindell, W. A., & van der Ende, J. (1985). An empirical-test of the utility of the observations-to-variables ratio in factor and components analysis. Applied Psychological Measurement, 9(2), 165–178. http://dx.doi.org/10.1177/014662168500900205
    » http://dx.doi.org/10.1177/014662168500900205
  • 8
    Atkins, K. G., & Kim, Y. (2012). Smart shopping: conceptualization and measurement. International Journal of Retail and Distribution Management, 40(5), 360–375. http://dx.doi.org/10.1108/09590551211222349
    » http://dx.doi.org/10.1108/09590551211222349
  • 9
    Bagdare, S., & Jain, R. (2013). Measuring retail customer experience. International Journal of Retail and Distribution Management, 41(10), 790–804. http://dx.doi.org/10.1108/IJRDM-08-2012-0084
    » http://dx.doi.org/10.1108/IJRDM-08-2012-0084
  • 10
    Bakar, H. A., & Mustaffa, C. S. (2013). Organizational communication in Malaysia organizations. Corporate Communications: An International Journal, 18(1), 87–109. http://dx.doi.org/10.1108/13563281311294146
    » http://dx.doi.org/10.1108/13563281311294146
  • 11
    Barrett, P. T., & Kline, P. (1981). The observation to variable ratio in factor analysis. Personality Study and Group Behavior, 1, 23–33.
  • 12
    Bastos, J. L., Celeste, R. K., Faerstein, E., & Barros, A. J. D. (2010). Racial discrimination and health: a systematic review of scales with a focus on their psychometric properties. Social Science and Medicine, 70, 1091–1099. http://dx.doi.org/10.1016/j.socscimed.2009.12.20
    » http://dx.doi.org/10.1016/j.socscimed.2009.12.20
  • 13
    Beaudreuil, J, Allard, A, Zerkak, D, Gerber, RA, Cappelleri, JC, Quintero, N, Lasbleiz, S, … Bardin, T. (2011). Unite’ Rhumatologique des Affections de la Main (URAM) Scale: development and validation of a tool to assess Dupuytren’s disease–specific disability. Arthritis Care & Research, 63(10), 1448-1455. doi: http://dx.doi.org/10.1002/acr.20564
    » http://dx.doi.org/10.1002/acr.20564
  • 14
    Bhattacherjee, A. (2002). Individual trust in online firms: scale development and initial test. Journal of Management Information Systems, 19(1), 211–241. http://dx.doi.org/10.1080/07421222.2002.11045715
    » http://dx.doi.org/10.1080/07421222.2002.11045715
  • 15
    Blankson, C., Cheng, J. M., & Spears, N. (2007). Determinants of banks selection in USA, Taiwan and Ghana. International Journal of Bank Marketing, 25(7), 469–489. http://dx.doi.org/10.1108/02652320710832621
    » http://dx.doi.org/10.1108/02652320710832621
  • 16
    Blankson, C., Paswan, A., & Boakye, K. G. (2012). College students’ consumption of credit cards. International Journal of Bank Marketing, 30(7), 567–585. http://dx.doi.org/10.1108/02652321211274327
    » http://dx.doi.org/10.1108/02652321211274327
  • 17
    Bolton, D. L., & Lane, M. D. (2012). Individual entrepreneurial orientation: development of a measurement instrument. Education + Training, 54(2/3), 219–233. http://dx.doi.org/10.1108/00400911211210314
    » http://dx.doi.org/10.1108/00400911211210314
  • 18
    Bova, C., Fennie, K. P., Watrous, E., Dieckhaus, K., & Williams, A. B. (2006). The health care relationship (HCR) trust scale: development and psychometric evaluation. Research in Nursing and Health, 29, 477–488. http://dx.doi.org/10.1002/nur.20158
    » http://dx.doi.org/10.1002/nur.20158
  • 19
    Bowen, H. P., & Wiersema, M. F. (1999). Matching method to paradigm in strategy research: limitations of cross-sectional analysis and some methodological alternatives. Strategic Management Journal, 20, 625–636.
  • 20
    Boyar, S. L., Campbell, N. S., Mosley, D. C., Jr., & Carson, C. M. (2014). Development of a work/family social support measure. Journal of Managerial Psychology, 29(7), 901–920. http://dx.doi.org/10.1108/JMP-06-2012-0189
    » http://dx.doi.org/10.1108/JMP-06-2012-0189
  • 21
    Brock, J. K., & Zhou, Y. (2005). Organizational use of the internet. Internet Research, 15(1), 67–87. http://dx.doi.org/10.1108/10662240510577077
    » http://dx.doi.org/10.1108/10662240510577077
  • 22
    Brun, I., Rajaobelina, L., & Ricard, L. (2014). Online relationship quality: scale development and initial testing. International Journal of Bank Marketing, 32(1), 5–27. http://dx.doi.org/10.1108/IJBM-02-2013-0022
    » http://dx.doi.org/10.1108/IJBM-02-2013-0022
  • 23
    Butt, M. M., & Run, E. C. (2010). Private healthcare quality: applying a SERVQUAL model. International Journal of Health Care Quality Assurance, 23(7), 658–673. http://dx.doi.org/10.1108/09526861011071580
    » http://dx.doi.org/10.1108/09526861011071580
  • 24
    Caro, L. M., & García, J. A. M. (2007). Measuring perceived service quality in urgent transport service. Journal of Retailing and Consumer Services, 14, 60–72. http://dx.doi.org/10.1016/j.jretconser.2006.04.001
    » http://dx.doi.org/10.1016/j.jretconser.2006.04.001
  • 25
    Chahal, H., & Kumari, N. (2012). Consumer perceived value. International Journal of Pharmaceutical and Healthcare Marketing, 6(2), 167–190. http://dx.doi.org/10.1108/17506121211243086
    » http://dx.doi.org/10.1108/17506121211243086
  • 26
    Chen, H., Tian, Y., & Daugherty, P. J. (2009). Measuring process orientation. The International Journal of Logistics Management, 20(2), 213–227. http://dx.doi.org/10.1108/09574090910981305
    » http://dx.doi.org/10.1108/09574090910981305
  • 27
    Choi, S. W., Victorson, D. E., Yount, S., Anton, S., & Cella, D. (2011). Development of a conceptual framework and calibrated item banks to measure patient-reported dyspnea severity and related functional limitations. Value in Health, 14, 291–306. http://dx.doi.org/10.1016/j.jval.2010.06.001
    » http://dx.doi.org/10.1016/j.jval.2010.06.001
  • 28
    Christophersen, T., & Konradt, U. (2012). Development and validation of a formative and a reflective measure for the assessment of online store usability. Behaviour and Information Technology, 31(9), 839–857. http://dx.doi.org/10.1080/0144929X.2010.529165
    » http://dx.doi.org/10.1080/0144929X.2010.529165
  • 29
    Churchill, G. (1979). A paradigm for developing better measures of marketing constructs. Journal of Marketing Research, 16(1), 64–73. http://dx.doi.org/10.2307/3150876
    » http://dx.doi.org/10.2307/3150876
  • 30
    Cicero, D. C., Kerns, J. G., & McCarthy, D. M. (2010). The Aberrant Salience Inventory: a new measure of psychosis proneness. Psychological Assessment, 22(3), 688–701. http://dx.doi.org/10.1037/a0019913
    » http://dx.doi.org/10.1037/a0019913
  • 31
    Clark, L. A., & Watson, D. (1995). Constructing validity: basic issues in objective scale development. Psychological Assessment, 7(3), 309–319. http://dx.doi.org/10.1037/1040-3590.7.3.309
    » http://dx.doi.org/10.1037/1040-3590.7.3.309
  • 32
    Coker, B. L. S., Ashill, N. J., & Hope, B. (2011). Measuring internet product purchase risk. European Journal of Marketing, 45(7/8), 1130–1151. http://dx.doi.org/10.1108/03090561111137642
    » http://dx.doi.org/10.1108/03090561111137642
  • 33
    Coleman, D., Chernatony, L., & Christodoulides, G. (2011). B2B service brand identity: scale development and validation. Industrial Marketing Management, 40, 1063–1071. http://dx.doi.org/10.1016/j.indmarman.2011.09.010
    » http://dx.doi.org/10.1016/j.indmarman.2011.09.010
  • 34
    Collins, L. M., Schafer, J. L., & Kam, C.-M. (2001). A comparison of inclusive and restrictive strategies in modern missing data procedures. Psychological Methods, 6(4), 330–351. http://dx.doi.org/10.1037/1082-989X.6.4.330
    » http://dx.doi.org/10.1037/1082-989X.6.4.330
  • 35
    Colwell, S. R., Aung, M., Kanetkar, V., & Holden, A. L. (2008). Toward a measure of service convenience: multiple-item scale development and empirical test. Journal of Services Marketing, 22(2), 160–169. http://dx.doi.org/10.1108/08876040810862895
    » http://dx.doi.org/10.1108/08876040810862895
  • 36
    Cossette, S., Cara, C., Ricard, N., & Pepin, J. (2005). Assessing nurse–patient interactions from a caring perspective: report of the development and preliminary psychometric testing of the Caring Nurse–Patient Interactions Scale. International Journal of Nursing Studies, 42, 673–686. http://dx.doi.org/10.1016/j.ijnurstu.2004.10.004
    » http://dx.doi.org/10.1016/j.ijnurstu.2004.10.004
  • 37
    Dennis, R. S., & Bocarnea, M. (2005). Development of the servant leadership assessment instrument. Leadership and Organization Development Journal, 26(8), 600–615. http://dx.doi.org/10.1108/01437730510633692
    » http://dx.doi.org/10.1108/01437730510633692
  • 38
    DeVellis, R. F. (2003). Scale development: theory and applications (2nd ed.). Newbury Park: Sage Publications.
  • 39
    Devlin, J. F., Roy, S. K., & Sekhon, H. (2014). Perceptions of fair treatment in financial services. European Journal of Marketing, 48(7/8), 1315–1332. http://dx.doi.org/10.1108/EJM-08-2012-0469
    » http://dx.doi.org/10.1108/EJM-08-2012-0469
  • 40
    Dunham, A., & Burt, C. (2014). Understanding employee knowledge: the development of an organizational memory scale. The Learning Organization, 21(2), 126–145. http://dx.doi.org/10.1108/TLO-04-2011-0026
    » http://dx.doi.org/10.1108/TLO-04-2011-0026
  • 41
    Edwards, J. R., Knight, D. K., Broome, K. M., & Flynn, P. M. (2010). The development and validation of a transformational leadership survey for substance use treatment programs. Substance Use and Misuse, 45, 1279–1302. http://dx.doi.org/10.3109/10826081003682834
    » http://dx.doi.org/10.3109/10826081003682834
  • 42
    Feuerstein, M., Nicholas, R. A., Huang, G. D., Haufler, A. J., Pransky, G., & Robertson, M. (2005). Workstyle: development of a measure of response to work in those with upper extremity pain. Journal of Occupational Rehabilitation, 15(2), 87–104. http://dx.doi.org/10.1007/s10926-005-3420-0
    » http://dx.doi.org/10.1007/s10926-005-3420-0
  • 43
    Fisher, R. J. (1993). Social desirability bias and the validity of indirect questioning. Journal of Consumer Research, 20(2), 303–315. http://dx.doi.org/10.1086/209351
    » http://dx.doi.org/10.1086/209351
  • 44
    Fisher, R., Maritz, A., & Lobo, A. (2014). Evaluating entrepreneurs’ perception of success. International Journal of Entrepreneurial Behavior and Research, 20(5), 478–492. http://dx.doi.org/10.1108/IJEBR-10-2013-0157
    » http://dx.doi.org/10.1108/IJEBR-10-2013-0157
  • 45
    Flight, R. L., D’Souza, G., & Allaway, A. W. (2011). Characteristics-based innovation adoption: scale and model validation. Journal of Product and Brand Management, 20(5), 343–355. http://dx.doi.org/10.1108/10610421111157874
    » http://dx.doi.org/10.1108/10610421111157874
  • 46
    Forbush, KT, Wildes, JE, Pollack, LO, Dunbar, D, Luo, J, Patterson, P, Petruzzi, L, … Watson, D. (2013). Development and validation of the Eating Pathology Symptoms Inventory (EPSI). Psychological Assessment, 25(3), 859-878. doi: http://dx.doi.org/10.1037/a0032639
    » http://dx.doi.org/10.1037/a0032639
  • 47
    Foster, J. D., McCain, J. L., Hibberts, M. F., Brunell, A. B., & Johnson, B. (2015). The grandiose narcissism scale: a global and facet-level measure of grandiose narcissism. Personality and Individual Differences, 73, 12–16. http://dx.doi.org/10.1016/j.paid.2014.08.042
    » http://dx.doi.org/10.1016/j.paid.2014.08.042
  • 48
    Franche, R., Corbière, M., Lee, H., Breslin, F. C., & Hepburn, G. (2007). The readiness for return-to-work (RRTW) scale: development and validation of a self-report staging scale in lost-time claimants with musculoskeletal disorders. Journal of Occupational Rehabilitation, 17, 450–472. http://dx.doi.org/10.1007/s10926-007-9097-9
    » http://dx.doi.org/10.1007/s10926-007-9097-9
  • 49
    Gesten, E. L. (1976). A health resources inventory: the development of a measure of the personal and social competence of primary-grade children. Journal of Consulting and Clinical Psychology, 44(5), 775–786. http://dx.doi.org/10.1037/0022-006X.44.5.775
    » http://dx.doi.org/10.1037/0022-006X.44.5.775
  • 50
    Gibbons, C. J., Kenning, C., Coventry, P. A., Bee, P., Bundy, C., Fisher, L., & Bower, P. (2013). Development of a Multimorbidity Illness Perceptions Scale (MULTIPleS). PloS One, 8(12), e81852. http://dx.doi.org/10.1371/journal.pone.0081852
    » http://dx.doi.org/10.1371/journal.pone.0081852
  • 51
    Gligor, D. M., & Holcomb, M. (2014). The road to supply chain agility: an RBV perspective on the role of logistics capabilities. The International Journal of Logistics Management, 25(1), 160–179. http://dx.doi.org/10.1108/IJLM-07-2012-0062
    » http://dx.doi.org/10.1108/IJLM-07-2012-0062
  • 52
    Glynn, N. W., Santanasto, A. J., Simonsick, E. M., Boudreau, R. M., Beach, S. R., Schulz, R., & Newman, A. B. (2015). The Pittsburgh fatigability scale for older adults: development and validation. Journal of American Geriatrics Society, 63, 130–135. http://dx.doi.org/10.1111/jgs.13191
    » http://dx.doi.org/10.1111/jgs.13191
  • 53
    Gottlieb, U., Brown, M., & Ferrier, L. (2014). Consumer perceptions of trade show effectiveness. European Journal of Marketing, 48(1/2), 89–107. http://dx.doi.org/10.1108/EJM-06-2011-0310
    » http://dx.doi.org/10.1108/EJM-06-2011-0310
  • 54
    Hair Junior, J. F., Black, W. C., Babin, N. J., Anderson, R. E., & Tatham, R. L. (2009). Análise multivariada de dados (6ath ed.). São Paulo: Bookman.
  • 55
    Hall, M. A., Camacho, F., Dugan, E., & Balkrishnan, R. (2002). Trust in the medical profession: conceptual and measurement issues. Health Services Research, 37(5), 1419–1439. http://dx.doi.org/10.1111/1475-6773.01070
    » http://dx.doi.org/10.1111/1475-6773.01070
  • 56
    Han, H., Back, K., & Kim, Y. (2011). A multidimensional scale of switching barriers in the full-service restaurant industry. Cornell Hospitality Quarterly, 52(1), 54–-63. http://dx.doi.org/10.1177/1938965510389261
    » http://dx.doi.org/10.1177/1938965510389261
  • 57
    Hardesty, D. M., & Bearden, W. O. (2004). The use of expert judges in scale development Implications for improving face validity of measures of unobservable constructs. Journal of Business Research, 57, 98–107. http://dx.doi.org/10.1016/S0148-2963(01)00295-8
    » http://dx.doi.org/10.1016/S0148-2963(01)00295-8
  • 58
    Henderson-King, D., & Henderson-King, E. (2005). Acceptance of cosmetic surgery: scale development and validation. Body Image, 2, 137–149. http://dx.doi.org/10.1016/j.bodyim.2005.03.003
    » http://dx.doi.org/10.1016/j.bodyim.2005.03.003
  • 59
    Hernandez, J. M. C., & Santos, C. C. (2010). Development-based trust: proposing and validating a new trust measurement model for buyer-seller relationships. Brazilian Administration Review, 7(2), 172–197. http://dx.doi.org/10.1590/S1807-76922010000200005
    » http://dx.doi.org/10.1590/S1807-76922010000200005
  • 60
    Hildebrandt, T., Langenbucher, J., & Schlundt, D. G. (2004). Muscularity concerns among men: development of attitudinal and perceptual measures. Body Image, 1, 169–181. http://dx.doi.org/10.1016/j.bodyim.2004.01.001
    » http://dx.doi.org/10.1016/j.bodyim.2004.01.001
  • 61
    Hilsenroth, M. J., Blagys, M. D., Ackerman, S. J., Bonge, D. R., & Blais, M. A. (2005). Measuring psychodynamic-interpersonal and cognitive-behavioral techniques: development of the comparative psychotherapy process scale. Psychotherapy: Theory, Research, Practice, Training, 42(3), 340–356. http://dx.doi.org/10.1037/0033-3204.42.3.340
    » http://dx.doi.org/10.1037/0033-3204.42.3.340
  • 62
    Hinkin, T. R. (1995). A review of scale development practices in the study of organizations. Journal of Management, 21(5), 967–988. http://dx.doi.org/10.1177/014920639502100509
    » http://dx.doi.org/10.1177/014920639502100509
  • 63
    Ho, C. B., & Lin, W. (2010). Measuring the service quality of internet banking: scale development and validation. European Business Review, 22(1), 5–24. http://dx.doi.org/10.1108/09555341011008981
    » http://dx.doi.org/10.1108/09555341011008981
  • 64
    Hutz, CS, Bandeira, DR, & Trentini, CM. (Org.). (2015). Psicometria Porto Alegre, Artmed
  • 65
    Jong, N., Van Leeuwen, R. G. J., Hoekstra, H. A., & van der Zee, K. I. (2014). CRIQ: an innovative measure using comparison awareness to avoid self-presentation tactics. Journal of Vocational Behavior, 84, 199–214. http://dx.doi.org/10.1016/j.jvb.2014.01.003
    » http://dx.doi.org/10.1016/j.jvb.2014.01.003
  • 66
    Kapuscinski, A. N., & Masters, K. S. (2010). The current status of measures of spirituality: a critical review of scale development. Psychology of Religion and Spirituality, 2(4), 191–205. http://dx.doi.org/10.1037/a0020498
    » http://dx.doi.org/10.1037/a0020498
  • 67
    Khine, M. S. (2008). Knowing, knowledge and beliefs: epistemological studies across diverse cultures New York: Springer.
  • 68
    Khorsan, R, & Crawford, C. (2014). External validity and model validity: a conceptual approach for systematic review methodology. Evidence-Based Complementary and Alternative Medicine, 2014, Article ID 694804, 12 pages. doi: http://dx.doi.org/10.1155/2014/694804
    » http://dx.doi.org/10.1155/2014/694804
  • 69
    Kim, S., Cha, J., Knutson, B. J., & Beck, J. A. (2011). Development and testing of the Consumer Experience Index (CEI). Managing Service Quality: An International Journal, 21(2), 112–132. http://dx.doi.org/10.1108/09604521111113429
    » http://dx.doi.org/10.1108/09604521111113429
  • 70
    Kim, D., Lee, Y., Lee, J., Nam, J. K., & Chung, Y. (2014). Development of Korean smartphone addiction proneness scale for youth. PloS One, 9(5), e97920. http://dx.doi.org/10.1371/journal.pone.0097920
    » http://dx.doi.org/10.1371/journal.pone.0097920
  • 71
    King, M. F., & Bruner, G. C. (2000). Social desirability bias: a neglected aspect of validity testing. Psychology and Marketing, 17(2), 79–103. http://dx.doi.org/10.1002/(SICI)1520-6793(200002)17:2<79::AID-MAR2>3.0.CO;2-0
    » http://dx.doi.org/10.1002/(SICI)1520-6793(200002)17:2<79::AID-MAR2>3.0.CO;2-0
  • 72
    Kwon, W., & Lennon, S. J. (2011). Assessing college women’s associations of American specialty apparel brands. Journal of Fashion Marketing and Management: An International Journal, 15(2), 242–256. http://dx.doi.org/10.1108/13612021111132663
    » http://dx.doi.org/10.1108/13612021111132663
  • 73
    Ladhari, R. (2010). Developing e-service quality scales: a literature review. Journal of Retailing and Consumer Services, 17, 464–477. http://dx.doi.org/10.1016/j.jretconser.2010.06.003
    » http://dx.doi.org/10.1016/j.jretconser.2010.06.003
  • 74
    Lin, J. C., & Hsieh, P. (2011). Assessing the self-service technology encounters: development and validation of SSTQUAL scale. Journal of Retailing, 87(2), 194–206. http://dx.doi.org/10.1016/j.jretai.2011.02.006
    » http://dx.doi.org/10.1016/j.jretai.2011.02.006
  • 75
    Lombaerts, K., Backer, F., Engels, N., Van Braak, J., & Athanasou, J. (2009). Development of the self-regulated learning teacher belief scale. European Journal of Psychology of Education, 24(1), 79–96. http://dx.doi.org/10.1007/BF03173476
    » http://dx.doi.org/10.1007/BF03173476
  • 76
    Lucas-Carrasco, R., Eser, E., Hao, Y., McPherson, K. M., Green, A., & Kullmann, L. (2011). The quality of care and support (QOCS) for people with disability scale: development and psychometric properties. Research in Developmental Disabilities, 32, 1212–1225. http://dx.doi.org/10.1016/j.ridd.2010.12.030
    » http://dx.doi.org/10.1016/j.ridd.2010.12.030
  • 77
    MacKenzie, S. B., Podsakoff, P. M., & Podsakoff, N. P. (2011). Construct measurement and validation procedures in MIS and behavioral research: integrating new and existing techniques. MIS Quarterly, 35(2), 293–334.
  • 78
    Mahudin, N. D. M., Cox, T., & Griffiths, A. (2012). Measuring rail passenger crowding: scale development and psychometric properties. Transportation Research Part, F 15, 38–51. http://dx.doi.org/10.1016/j.trf.2011.11.006
    » http://dx.doi.org/10.1016/j.trf.2011.11.006
  • 79
    Malhotra, N. K. (2004). Pesquisa de marketing: Uma orientação aplicada (4ath ed.). Porto Alegre: Bookman.
  • 80
    Medina-Pradas, C., Navarro, J. B., López, S. R., Grau, A., & Obiols, J. E. (2011). Further development of a scale of perceived expressed emotion and its evaluation in a sample of patients with eating disorders. Psychiatry Research, 190, 291–296. http://dx.doi.org/10.1016/j.psychres.2011.06.011
    » http://dx.doi.org/10.1016/j.psychres.2011.06.011
  • 81
    Meneses, J., Barrios, M., Bonillo, A., Cosculluela, A., Lozano, L. M., Turbany, J., & Valero, S. (2014). Psicometría Barcelona: Editorial UOC.
  • 82
    Morean, M. E., Corbin, W. R., & Treat, T. A. (2012). The anticipated effects of alcohol scale: development and psychometric evaluation of a novel assessment tool for measuring alcohol expectancies. Psychological Assessment, 24(4), 1008–1023. http://dx.doi.org/10.1037/a0028982
    » http://dx.doi.org/10.1037/a0028982
  • 83
    Morgado, F. F. R., Campana, A. N. N. B., & Tavares, M. C. G. C. F. (2014). Development and validation of the self-acceptance scale for persons with early blindness: the SAS-EB. PloS One, 9(9), e106848. http://dx.doi.org/10.1371/journal.pone.0106848
    » http://dx.doi.org/10.1371/journal.pone.0106848
  • 84
    Nagy, B. G., Blair, E. S., & Lohrke, F. T. (2014). Developing a scale to measure liabilities and assets of newness after start-up. International Entrepreneurship and Management Journal, 10, 277–295. http://dx.doi.org/10.1007/s11365-012-0219-2
    » http://dx.doi.org/10.1007/s11365-012-0219-2
  • 85
    Napoli, J., Dickinson, S. J., Beverland, M. B., & Farrelly, F. (2014). Measuring consumer-based brand authenticity. Journal of Business Research, 67, 1090–1098. http://dx.doi.org/10.1016/j.jbusres.2013.06.001
    » http://dx.doi.org/10.1016/j.jbusres.2013.06.001
  • 86
    Negra, A., & Mzoughi, M. N. (2012). How wise are online procrastinators? A scale development. Internet Research, 22(4), 426–442. http://dx.doi.org/10.1108/10662241211250971
    » http://dx.doi.org/10.1108/10662241211250971
  • 87
    Ngorsuraches, S., Lerkiatbundit, S., Li, S. C., Treesak, C., Sirithorn, R., & Korwiwattanakarn, M. (2007). Development and validation of the patient trust in community pharmacists (TRUST-Ph) scale: results from a study conducted in Thailand. Research in Social and Administrative Pharmacy, 4, 272–283. http://dx.doi.org/10.1016/j.sapharm.2007.10.002
    » http://dx.doi.org/10.1016/j.sapharm.2007.10.002
  • 88
    Nunnally, J. C. (1967). Psychometric theory New York: McGraw Hill.
  • 89
    Oh, H. (2005). Measuring affective reactions to print apparel advertisements: a scale development. Journal of Fashion Marketing and Management: An International Journal, 9(3), 283–305. http://dx.doi.org/10.1108/13612020510610426
    » http://dx.doi.org/10.1108/13612020510610426
  • 90
    Olaya, B, Marsà, F, Ochoa, S, Balanzá-Martínez, V, Barbeito, S, González-Pinto, A, … Haro, JM. (2012). Development of the insight scale for affective disorders (ISAD): modification from the scale to assess unawareness of mental disorder. Journal of Affective Disorders, 142, 65-71. doi: http://dx.doi.org/10.1016/j.jad.2012.03.041
    » http://dx.doi.org/10.1016/j.jad.2012.03.041
  • 91
    Omar, N. A., & Musa, R. (2011). Measuring service quality in retail loyalty programmes (LPSQual). International Journal of Retail and Distribution Management, 39(10), 759–784. http://dx.doi.org/10.1108/09590551111162257
    » http://dx.doi.org/10.1108/09590551111162257
  • 92
    Pan, J., Wong, D. F. K., & Ye, S. (2013). Post-migration growth scale for Chinese international students: development and validation. Journal of Happiness Studies, 14, 1639–1655. http://dx.doi.org/10.1007/s10902-012-9401-z
    » http://dx.doi.org/10.1007/s10902-012-9401-z
  • 93
    Pasquali, L. (2010). Instrumentação psicológica: fundamentos e práticas Porto Alegre: Artmed.
  • 94
    Patwardhan, H., & Balasubramanian, S. K. (2011). Brand romance: a complementary approach to explain emotional attachment toward brands. Journal of Product and Brand Management, 20(4), 297–308. http://dx.doi.org/10.1108/10610421111148315
    » http://dx.doi.org/10.1108/10610421111148315
  • 95
    Pimentel, C. E., Gouveia, V. V., & Pessoa, V. S. (2007). Escala de Preferência Musical: construção e comprovação da sua estrutura fatorial. Psico-USF, 12(2), 145–155.
  • 96
    Podsakoff, N. P., Podsakoff, P. M., MacKenzie, S. B., & Klinger, R. L. (2013). Are we really measuring what we say we’re measuring? Using video techniques to supplement traditional construct validation procedures. Journal of Applied Psychology, 98(1), 99–113. http://dx.doi.org/10.1037/a0029570
    » http://dx.doi.org/10.1037/a0029570
  • 97
    Pommer, AM, Prins, L, van Ranst, D, Meijer, J, Hul, AV, Janssen, J, … Pop, VJM. (2013). Development and validity of the Patient-Centred COPD Questionnaire (PCQ). Journal of Psychosomatic Research, 75, 563-571. doi: http://dx.doi.org/10.1016/j.jpsychores.2013.10.001
    » http://dx.doi.org/10.1016/j.jpsychores.2013.10.001
  • 98
    Prados, J. M. (2007). Development of a new scale of beliefs about the worry consequences. Annals of Psychology, 23(2), 226–230.
  • 99
    Raykov, T. (2008). Alpha if item deleted: a note on loss of criterion validity in scale development if maximizing coefficient alpha. British Journal of Mathematical and Statistical Psychology, 61, 275–285. http://dx.doi.org/10.1348/000711007X188520
    » http://dx.doi.org/10.1348/000711007X188520
  • 100
    Reed, L. L., Vidaver-Cohen, D., & Colwell, S. R. (2011). A new scale to measure executive servant leadership: development, analysis, and implications for research. Journal of Business Ethics, 101, 415–434. http://dx.doi.org/10.1007/s10551-010-0729-1
    » http://dx.doi.org/10.1007/s10551-010-0729-1
  • 101
    Reise, S. P., Waller, N. G., & Comrey, A. L. (2000). Factor analysis and scale revision. Psychological Assessment, 12(3), 287–297. http://dx.doi.org/10.1037//1040-3590.12.3.287
    » http://dx.doi.org/10.1037//1040-3590.12.3.287
  • 102
    Rice, S. M., Fallon, B. J., Aucote, H. M., & Möller-Leimkühler, A. M. (2013). Development and preliminary validation of the male depression risk scale: Furthering the assessment of depression in men. Journal of Affective Disorders, 151, 950–958. http://dx.doi.org/10.1016/j.jad.2013.08.013
    » http://dx.doi.org/10.1016/j.jad.2013.08.013
  • 103
    Riedel, M., Spellmann, I., Schennach-Wolff, R., Obermeier, M., & Musil, R. (2011). The RSM-scale: a pilot study on a new specific scale for self- and observer-rated quality of life in patients with schizophrenia. Quality of Life Research, 20, 263–272. http://dx.doi.org/10.1007/s11136-010-9744-z
    » http://dx.doi.org/10.1007/s11136-010-9744-z
  • 104
    Roberson, R. B., III, Elliott, T. R., Chang, J. E., & Hill, J. N. (2014). Exploratory factor analysis in rehabilitation psychology: a content analysis. Rehabilitation Psychology, 59(4), 429–438. http://dx.doi.org/10.1037/a0037899
    » http://dx.doi.org/10.1037/a0037899
  • 105
    Rodrigues, A. C. A., & Bastos, A. V. B. (2012). Organizational entrenchment: scale development and validation. Psicologia: Reflexão e Crítica, 25(4), 688–700. http://dx.doi.org/10.1590/S0102-79722012000400008
    » http://dx.doi.org/10.1590/S0102-79722012000400008
  • 106
    Rodríguez, I., Kozusznik, M. W., & Peiró, J. M. (2013). Development and validation of the Valencia Eustress-Distress Appraisal Scale. International Journal of Stress Management, 20(4), 279–308. http://dx.doi.org/10.1037/a0034330
    » http://dx.doi.org/10.1037/a0034330
  • 107
    Rosenthal, S. (2011). Measuring knowledge of indoor environmental hazards. Journal of Environmental Psychology, 31, 137–146. http://dx.doi.org/10.1016/j.jenvp.2010.08.003
    » http://dx.doi.org/10.1016/j.jenvp.2010.08.003
  • 108
    Saxena, S., Ayers, C. R., Dozier, M. E., & Maidment, K. M. (2015). The UCLA Hoarding Severity Scale: development and validation. Journal of Affective Disorders, 175, 488–493. http://dx.doi.org/10.1016/j.jad.2015.01.030
    » http://dx.doi.org/10.1016/j.jad.2015.01.030
  • 109
    Schafer, J. L., & Graham, J. W. (2002). Missing data: our view of the state of the Art. Psychological Methods, 7(2), 147–177. http://dx.doi.org/10.1037//1082-989X.7.2.147
    » http://dx.doi.org/10.1037//1082-989X.7.2.147
  • 110
    Schlosser, F. K., & McNaughton, R. B. (2009). Using the I-MARKOR scale to identify market-oriented individuals in the financial services sector. Journal of Services Marketing, 23(4), 236–248. http://dx.doi.org/10.1108/08876040910965575
    » http://dx.doi.org/10.1108/08876040910965575
  • 111
    Sewitch, M. J., Abrahamowicz, M., Dobkin, P. L., & Tamblyn, R. (2003). Measuring differences between patients’ and physicians’ health perceptions: the patient–physician discordance scale. Journal of Behavioral Medicine, 26(3), 245–263. http://dx.doi.org/10.1023/A:1023412604715
    » http://dx.doi.org/10.1023/A:1023412604715
  • 112
    Sharma, P. (2010). Measuring personal cultural orientations: scale development and validation. Journal of the Academy of Marketing Science, 38, 787–806. http://dx.doi.org/.1007/s11747-009-0184-7
    » http://dx.doi.org/.1007/s11747-009-0184-7
  • 113
    Sharma, D., & Gassenheimer, J. B. (2009). Internet channel and perceived cannibalization. European Journal of Marketing, 43(7/8), 1076–1091. http://dx.doi.org/10.1108/03090560910961524
    » http://dx.doi.org/10.1108/03090560910961524
  • 114
    Shawyer, F., Ratcliff, K., Mackinnon, A., Farhall, J., Hayes, S. C., & Copolov, D. (2007). The Voices Acceptance and Action Scale (VAAS): pilot data. Journal of Clinical Psychology, 63(6), 593–606. http://dx.doi.org/10.1002/jclp.20366
    » http://dx.doi.org/10.1002/jclp.20366
  • 115
    Sin, L. Y. M., Tse, A. C. B., & Yim, F. H. K. (2005). CRM: conceptualization and scale development. European Journal of Marketing, 39(11/12), 1264–1290. http://dx.doi.org/10.1108/03090560510623253
    » http://dx.doi.org/10.1108/03090560510623253
  • 116
    Sohn, D., & Choi, S. M. (2014). Measuring expected interactivity: scale development and validation. New Media and Society, 16(5), 856–870. http://dx.doi.org/10.1177/1461444813495808
    » http://dx.doi.org/10.1177/1461444813495808
  • 117
    Song, J. H., Uhm, D., & Yoon, S. W. (2011). Organizational knowledge creation practice. Leadership and Organization Development Journal, 32(3), 243–259. http://dx.doi.org/10.1108/01437731111123906
    » http://dx.doi.org/10.1108/01437731111123906
  • 118
    Staines, Z. (2013). Managing tacit investigative knowledge: measuring “investigative thinking styles”. Policing: An International Journal of Police Strategies and Management, 36(3), 604–619. http://dx.doi.org/10.1108/PIJPSM-07-2012-0072
    » http://dx.doi.org/10.1108/PIJPSM-07-2012-0072
  • 119
    Sultan, P., & Wong, H. (2010). Performance-based service quality model: an empirical study on Japanese universities. Quality Assurance in Education, 18(2), 126–143. http://dx.doi.org/10.1108/09684881011035349
    » http://dx.doi.org/10.1108/09684881011035349
  • 120
    Sveinbjornsdottir, S., & Thorsteinsson, E. B. (2008). Adolescent coping scales: a critical psychometric review. Scandinavian Journal of Psychology, 49(6), 533–548. http://dx.doi.org/10.1111/j.1467-9450.2008.00669.x
    » http://dx.doi.org/10.1111/j.1467-9450.2008.00669.x
  • 121
    Swaid, S. I., & Wigand, R. T. (2009). Measuring the quality of E-Service: scale development and initial validation. Journal of Electronic Commerce Research, 10(1), 13–28.
  • 122
    Tanimura, C., Morimoto, M., Hiramatsu, K., & Hagino, H. (2011). Difficulties in the daily life of patients with osteoarthritis of the knee: scale development and descriptive study. Journal of Clinical Nursing, 20, 743–753. http://dx.doi.org/10.1111/j.1365-2702.2010.03536.x
    » http://dx.doi.org/10.1111/j.1365-2702.2010.03536.x
  • 123
    Taute, H. A., & Sierra, J. (2014). Brand tribalism: an anthropological perspective. Journal of Product and Brand Management, 23(1), 2–15. http://dx.doi.org/10.1108/JPBM-06-2013-0340
    » http://dx.doi.org/10.1108/JPBM-06-2013-0340
  • 124
    Tombaugh, J. R., Mayfield, C., & Durand, R. (2011). Spiritual expression at work: exploring the active voice of workplace spirituality. International Journal of Organizational Analysis, 19(2), 146–170. http://dx.doi.org/10.1108/19348831111135083
    » http://dx.doi.org/10.1108/19348831111135083
  • 125
    Turker, D. (2009). Measuring corporate social responsibility: a scale development study. Journal of Business Ethics, 85, 411–427. http://dx.doi.org/10.1007/s10551-008-9780-6
    » http://dx.doi.org/10.1007/s10551-008-9780-6
  • 126
    Uzunboylu, H., & Ozdamli, F. (2011). Teacher perception for m-learning: scale development and teachers’ perceptions. Journal of Computer Assisted Learning, 27, 544–556. http://dx.doi.org/10.1111/j.1365-2729.2011.00415.x
    » http://dx.doi.org/10.1111/j.1365-2729.2011.00415.x
  • 127
    Van der Gaag, M, Schütz, C, ten Napel, A, Landa, Y, Delespaul, P, Bak, M, … Hert, M. (2013). Development of the Davos Assessment of Cognitive Biases Scale (DACOBS). Schizophrenia Research, 144, 63-71. doi: http://dx.doi.org/10.1016/j.schres.2012.12.010
    » http://dx.doi.org/10.1016/j.schres.2012.12.010
  • 128
    Von Steinbüchel, N, Wilson, L, Gibbons, H, Hawthorne, G, Höfer, S, Schmidt, S, … Truelle, J. (2010). Journal of Neurotrauma, 27, 1167-1185. doi: http://dx.doi.org/10.1089/neu.2009.1076
    » http://dx.doi.org/10.1089/neu.2009.1076
  • 129
    Voon, B. H., Abdullah, F., Lee, N., & Kueh, K. (2014). Developing a HospiSE scale for hospital service excellence. International Journal of Quality and Reliability Management, 31(3), 261–280. http://dx.doi.org/10.1108/IJQRM-10-2012-0143
    » http://dx.doi.org/10.1108/IJQRM-10-2012-0143
  • 130
    Walshe, M., Peach, R. K., & Miller, N. (2009). Dysarthria Impact Profile: development of a scale to measure psychosocial effects. International Journal of Language and Communication Disorders, 44(5), 693–715. http://dx.doi.org/10.1080/13682820802317536
    » http://dx.doi.org/10.1080/13682820802317536
  • 131
    Wang, C. L., & Mowen, J. C. (1997). The separateness-connectedness self-schema: scale development and application to message construction. Psychology and Marketing, 14(2), 185–207. http://dx.doi.org/10.1002/(SICI)1520-6793(199703)14:2<185::AID-MAR5>3.0.CO;2-9
    » http://dx.doi.org/10.1002/(SICI)1520-6793(199703)14:2<185::AID-MAR5>3.0.CO;2-9
  • 132
    Wepener, M., & Boshoff, C. (2015). An instrument to measure the customer-based corporate reputation of large service organizations. Journal of Services Marketing, 29(3), 163–172. http://dx.doi.org/10.1108/JSM-01-2014-0026
    » http://dx.doi.org/10.1108/JSM-01-2014-0026
  • 133
    Williams, Z., Ponder, N., & Autry, C. W. (2009). Supply chain security culture: measure development and validation. The International Journal of Logistics Management, 20(2), 243–260. http://dx.doi.org/10.1108/09574090910981323
    » http://dx.doi.org/10.1108/09574090910981323
  • 134
    Wilson, N. L., & Holmvall, C. M. (2013). The development and validation of the incivility from customers scale. Journal of Occupational Health Psychology, 18(3), 310–326. http://dx.doi.org/10.1037/a0032753
    » http://dx.doi.org/10.1037/a0032753
  • 135
    Yang, M., Weng, S., & Hsiao, P. (2014). Measuring blog service innovation in social media services. Internet Research, 24(1), 110–128. http://dx.doi.org/10.1108/IntR-12-2012-0253
    » http://dx.doi.org/10.1108/IntR-12-2012-0253
  • 136
    Zhang, X., & Hu, D. (2011). Farmer-buyer relationships in China: the effects of contracts, trust and market environment. China Agricultural Economic Review, 3(1), 42–53. http://dx.doi.org/10.1108/17561371111103534
    » http://dx.doi.org/10.1108/17561371111103534
  • 137
    Zheng, J, You, L, Lou, T, Chen, N, Lai, D, Liang, Y, … Zhai, C. (2010). Development and psychometric evaluation of the dialysis patient-perceived exercise benefits and barriers scale. International Journal of Nursing Studies, 47, 166-180. doi: http://dx.doi.org/10.1016/j.ijnurstu.2009.05.023
    » http://dx.doi.org/10.1016/j.ijnurstu.2009.05.023

Publication Dates

  • Publication in this collection
    2017

History

  • Received
    3 Aug 2016
  • Reviewed
    22 Dec 2016
  • Accepted
    25 Jan 2017
Curso de Pós-Graduação em Psicologia da Universidade Federal do Rio Grande do Sul Rua Ramiro Barcelos, 2600 - sala 110, 90035-003 Porto Alegre RS - Brazil, Tel.: +55 51 3308-5691 - Porto Alegre - RS - Brazil
E-mail: prc@springeropen.com