Review Article
Item reduction based on rigorous methodological guidelines is necessary to maintain validity when shortening composite measurement scales

https://doi.org/10.1016/j.jclinepi.2012.12.015Get rights and content

Abstract

Objective

To review current practice and update guidelines for the methodology of shortening composite measurement scales (CMSs).

Study Design and Setting

A literature review gathered data on 91 shortening processes from 1995 to 2009. The validity of the initial CMS, the shortening methods, and the validity of the derived short-form scales were examined. The results were compared with those from a previous literature review (articles from 1985 to 1995) to develop updated guidelines for CMS shortening.

Results

The literature review revealed a persisting lack of use of rigorous methodology for CMS shortening. Of the 91 cases of CMS shortening, 36 combined a content approach and a statistical approach; 45 used only a statistical approach and 10 (11%) only a content approach. The updated guidelines deal with the validity and conceptual model of the initial CMS, the preservation of content and psychometric properties during shortening, the selection of items, and the validation of the short form.

Conclusion

Item reduction based on a rigorous methodology is necessary if the short-form instrument aims to maintain the validity and other measurement properties of the parent instrument, which in turn supports application in research and clinical practice.

Introduction

Many health constructs are too complex to be captured by direct measurement. When these constructs need to be examined, one of the most popular methods requires the use of a composite measurement scale (CMS). The CMS generally consists of items or questions that assess one or several attributes scored by a scale.

Over the years, the measurement of health constructs has led to the production of a large number of scales, with often a high number of items [1].

The burden of long scales and the increasing need for multiple instruments in the same study have logically created a strong need to shorten CMSs. To shorten a CMS consists in reducing its number of items while trying to preserve or improve its psychometric properties. The methodology to develop new CMSs is well documented [2] and includes consideration of different psychometric properties but has limited applicability as a means of informing the shortening of instruments. Guidelines for shortening existing CMSs are scarce. In 1997, Coste et al. [3] noted that most articles reporting on scale shortening lacked rigorous methodology: shortening processes were often inadequately conceptualized, and excessive credit was given to statistical techniques. The authors recommended carefully choosing the original scale according to its content, its possibility for shortening, and its psychometric properties; focusing on criterion validity if the original scale can be considered the gold standard and if not, preferring an expert-based approach to content validity, possibly helped by statistical considerations; and finally, performing a validation study in an independent sample. In 2000, Smith et al. [4] also noted methodological pitfalls concerning CMS shortening in psychology and recommended using a validated original scale, clarifying the intended use of the short form, estimating a priori the properties of the short form to balance resource or time savings against the loss of validity, preserving content, and using an independent sample to validate the short form. In 2002, Stanton et al. [5] also observed a lack of methodological recommendations for scale reduction. The authors argued that focusing on internal consistency should be avoided and proposed a set of item “quality indices” to help conceptualize the competing issues that influence item retention decisions. However, new approaches and new statistical methods such as item response theory (IRT) are now used for scale shortening [6], but we lack published recommendations integrating these methods. Another shortfall in current guidelines concerns content validity; retaining content during the shortening process is important, but only methodological guidelines for developing new CMSs focus on content analysis [7], [8].

This article aims to describe the methodology currently used to shorten CMSs through a literature review and to compare with a previous review for proposing updated and structured guidelines for CMS shortening.

Section snippets

Search and identification of articles

Articles reporting on the development of a short form of an existing CMS concerning health or psychology domains, published between January 1, 1995 and December 31, 2009 and written in English, were selected from MedLine and Psycinfo. The following query was used to search title, abstract, and keyword fields: (“short form” or “brief form” or “short version” or “brief version”) and (questionnaire or scale or instrument) and (development or validation or reduction or shortening or “item reduction

Results

In total, 103 articles met the inclusion criteria (list available in Appendix at www.jclinepi.com), corresponding to 91 shortening processes. The process of selecting the articles is in Fig. 1. The number of published articles describing CMS shortening increased during the study period (Fig. 2). Only 17 (19%) of the shortening processes cited at least one of the three articles proposing guidelines for CMS shortening [3], [4], [5].

The main domains covered by the involved CMSs were psychology (n = 

Discussion

The analysis of the literature of CMS shortening points out to a persistent lack of rigorous methodology for such shortening. The original CMS on which short forms were developed were not sufficiently described. Their psychometric properties were not well described before than after publication of recommendations in 1997, 2000, and 2002 [3], [4], [5]. A shortening process aims to preserve the psychometrics properties of a CMS into a short form; thus, a shortening study must be started by

Document the validity of the original CMS and the objective of its shortening

The short form of an existing CMS may benefit from the quality, the validity, and possibly the popularity of the original CMS. However, the original CMS must have well-documented measurement properties.

Although the original CMS must have satisfactory measurement properties, the shortening process may be an opportunity to improve some of these. A short form should definitively not be developed from a CMS with insufficiently documented properties. The evaluation of a CMS includes content,

Conclusions

The methodology of CMS shortening too often lacks rigor, so their short forms may be at risk of poorer validity than the original long form. The guidelines we propose can help researchers identify the key steps to follow when developing a short-form CMS from an existing one. Following these guidelines to shorten a CMS will not lead to a perfect short-form CMS but will guarantee that the main pitfalls have been avoided. It will also let future users of the short-form CMS examine what choices

Acknowledgment

The authors thank F. Bonnetain, B. Fautrel, and T. Conroy for their careful review of the guidelines.

References (19)

  • J. Coste et al.

    Methodological approaches to shortening composite measurement scales

    J Clin Epidemiol

    (1997)
  • C.B. Terwee et al.

    Quality criteria were proposed for measurement properties of health status questionnaires

    J Clin Epidemiol

    (2007)
  • A. Garratt et al.

    Quality of life measurement: bibliographic study of patient assessed health outcome measures

    BMJ

    (2002)
  • D.L. Streiner et al.

    Health measurement scales, a practical guide to their development and use

    (2003)
  • G.T. Smith et al.

    On the sins of short-form development

    Psychol Assess

    (2000)
  • J.M. Stanton et al.

    Issues and strategies for reducing the length of self-report scales

    Pers Psychol

    (2002)
  • M.O. Edelen et al.

    Applying item response theory (IRT) modeling to questionnaire development, evaluation, and refinement

    Qual Life Res

    (2007)
  • R. Sartori et al.

    Quality and quantity in test validity: how can we be sure that psychological tests measure what they have to?

    Qual Quant

    (2007)
  • Standards for educational and psychological tests

    (1985)
There are more references available in the full text version of this article.

Cited by (111)

  • The validity and measurement equivalence of a brief safety climate questionnaire across casual and permanent workers

    2023, Safety Science
    Citation Excerpt :

    Similarly, future research may benefit from examining the NOSACQ-24 within organisations to determine if specific demographic subgroups can be identified using the measure, allowing for more accurate and specifically tailored intervention strategies to remediate identified safety concerns. The additional evidence in this study regarding dimension structure, external validity, construct validity, concurrent validity, and measurement equivalence of the NOSACQ-24 further supports the item-reduction approach outlined by both Goetz et al. (2013), and the addition of a practice-focussed criterion for item selection proposed in a previous study by Summers et al. (2022), which ultimately resulted in superior model fit indices. Given that this previous research identified a potential issue regarding negatively worded items and the impact they have on factor structure, future research may benefit from examining model fit indices in a version of the NOSACQ that contains only positively worded questions, particularly because one dimension of the NOSACQ, “Workers’ Safety Priority and Risk Non-Acceptance”, largely comprises (∼85 %) negatively worded questions.

  • Mental health risk factors for shift work disorder in paramedics: A longitudinal study

    2023, Sleep Health
    Citation Excerpt :

    This is further supported by our finding that baseline insomnia failed to significantly predict SWD at follow-up, although it did have a large odds ratio. With respect to the PHQ-9, it is important to note that examining individual questionnaire items to assess specific symptoms may lack the validity and reliability that multi-component measures possess.49 Likewise, reducing items from a measure without appropriate validation may nullify the measure's ability to detect the construct it was designed for.49

View all citing articles on Scopus

Conflict of interest statement: The authors declare no conflict of interest.

View full text