Abstract
There has been much philosophical interest regarding the ‘hierarchy of evidence’ used to determine which study designs are of most value for reporting on questions of effectiveness, prognosis, and so on. There has been much less philosophical interest in the choice of outcome measures with which the results of, say, an RCT or a cohort study are presented. In this paper, we examine the FDA’s recently published guidelines for assessing the psychometric adequacy of patient-reported outcome measures. We focus on their recommendations for demonstrating content validity and also for how researchers should weigh up the sum of psychometric evidence when choosing these measures. We argue that questions regarding judgment and understanding meaning of these measures should play a more central role in determining their adequacy.
Similar content being viewed by others
Notes
The potential for new insight into a text does not suggest that just any interpretation of a text is acceptable. For Gadamer two further conditions delimit understanding, namely, that a text is understood as coherent and taken as something from which we might learn.
This difference is not insignificant, but for our purposes here, it is what they share that distinguishes them.
True criterion validation of PROMs is rarely possible owing to the absence of accepted gold standards, and we do not consider it further in this paper.
Reliability was assessed in terms of internal consistency and test–retest reliability; validity was assessed in terms of content validity, criterion validity, construct validity (within scale analyses, convergent and discriminant, known groups), and other hypothesis testing.
Although it is beyond the scope of this paper, it is worth noting that the suggestion that measures of study quality are a sound pointer to the presence of bias is controversial. For instance, there is an ongoing debate within the evidence-based medicine movement about the extent to which measures of the quality of studies of clinical effectiveness can discriminate between biased and nonbiased studies. One example of this is the controversy around the quality of randomized controlled trials of screening for breast cancer with mammography; see Freedman et al. [22]. Indeed, we might also see John Worrall’s philosophical work critiquing randomization as an extension of this debate insofar as he questions the importance of randomization for ensuring high quality studies; see [1].
Alternative practices are used by, for example, health economists who would argue that PROMs designed for use in economic evaluation (e.g., the EQ-5D instrument) do not need to meet minimum psychometric standards, but must simply be ‘better’ than any alternative on offer [19].
It is arguable that this construct validity problem is caused by poor content validity in the VF-14 for the range of the construct in question. This example is illustrative of the iterative nature of construct and content validation. The potential flaw in content validity only emerges after construct validation. Moreover, the problem with the VF-14 can only be solved after further rounds of content and construct validation.
Our suggestion that the choice of a PROM may need to be determined on a case by case basis is not meant to suggest that expert judgment is infallible, but rather that attempts to codify such judgments does not overcome problems of fallibility.
References
Worrall, J. 2002. What evidence in evidence-based medicine? Philosophy of Science 69: S316–S330.
Goldenberg, M.J. 2009. Iconoclast or creed? Objectivism, pragmatism, and the hierarchy of evidence. Perspectives in Biological Medicine 52: 168–187.
Cartwright, N., and E. Munro. 2010. The limitations of randomized control trials in predicting effectiveness. Evaluation in Clinical Practice 16: 260–266.
Garratt, A., L. Schmidt, A. MacIntosh, et al. 2002. Quality of life measurement: Bibliographic study of patient assessed health outcome measures. British Medical Journal 324: 1417.
US Food and Drug Administration. 2009. Guidance for industry on patient-reported outcome measures: Use in medicinal product development to support labeling claims. Federal Register 74: 1–43.
Rotherman, M., L. Burke, P. Erickson, N.K. Leidy, D.L. Patrick, and C.D. Petrie. 2009. Use of existing Patient-Reported Outcome (PRO) instruments and their modification: The ISPOR good research practices for evaluating and documenting content validity for the use of existing instruments and their modification PRO Task Force Report. Value in Health 12(8): 1075–1083.
Mallinson, S. 2002. Listening to respondents: A qualitative assessment of the Short-Form 36 Health Status Questionnaire. Social Science and Medicine 54: 11–21.
Schwartz, C.E., and B.D. Rapkin. 2004. Reconsidering the psychometrics of quality of life assessment in light of response shift and appraisal. Health and Quality of Life Outcomes 2: 14.
Westerman, M.J., T. Hak, M.A. Sprangers, H.J. Groen, G. van der Wal, and A.M. The. 2008. Listen to their answers! Response Behavior in the measurement of physical and role functioning. Quality of Life Research 17: 549–558.
Rapkin, B.D., and C.E. Schwartz. 2004. Toward a theoretical model of quality-of-life appraisal: Implications of findings from studies of response shift. Health and Quality of Life Outcomes 2: 14.
McClimans, L. 2010. A theoretical framework for patient-reported outcome measures. Theoretical Medicine and Bioethics 31(3): 225–240.
Hunt, S.M. 1997. The problem of quality of life. Quality of Life Research 6: 205–212.
Lamping, D. 2008. Presidential address given at the 2008 International Society for Quality of Life Research (ISOQoL) annual conference, Montevideo, Uruguay, October 25.
Gadamer, H.G. 2003. Truth and method. 2nd ed. Trans. J. Weinsheimer and D.G. Marshall. New York: Continuum Press.
UK Department of Health. 2007. Guidance on the routine collection of patient reported outcome measures (PROMs). London: Department of Health.
McClimans, L. 2010. Towards self-determination in quality of life research. Medicine, Health Care and Philosophy 13: 67.
Alexandrova, A. 2008. First-person reports and the measurement of happiness. Philosophical Psychology 21: 657–669.
Fitzpatrick, R., A. Bowling, E. Gibbons, et al. 2006. A structured review of patient-reported measures in relation to selected chronic conditions, perceptions of quality of care and carer impact. University of Oxford Patient Reported Outcomes Measurement Group. http://phi.uhce.ox.ac.uk/pdf/Chronic Conditions/ch12_discussion.pdf.
Smith, S.C., S. Cano, D. Lamping, et al. 2005. Patient-reported outcome measures (PROMs) for routine use in treatment centres: Recommendations based on a review of the scientific evidence. Department of Public Health and Policy Documents, London School of Hygiene and Tropical Medicine. http://www.lshtm.ac.uk/php/documents/PROMS%20Final%20report%20Dec%2005.pdf.
Streiner, D.L., and G.R. Norman. 2003. Health measurement scales: A practical guide to their development and use. 3rd ed. Oxford: Oxford University Press.
Terwee, C., S.D.M. Bot, M.R. de Boer, et al. 2007. Quality criteria were proposed for measurement properties of health status questionnaires. Journal of Clinical Epidemiology 60: 34–42.
Freedman, D.A., D.B. Petitti, and J.M. Robins. 2004. On the efficacy of screening for breast cancer. International Journal of Epidemiology 33: 43–55.
Steinberg, E.P., J.M. Tielsch, O.D. Schein, et al. 1994. The VF-14: An index of functional impairment in patients with cataracts. Archives of Ophthalmology 112: 630–638.
Browne, J.P., L. Jamieson, J. Lewsey, et al. 2007. Patient-reported outcome measures (PROMs) in routine elective surgery. London: London School of Hygiene and Tropical Medicine. http://www.lshtm.ac.uk/hsru/research/PROMs-Report-12-Dec-07.pdf.
Black, N., J. Browne, J. van der Meulen, L. Jamieson, L. Copely, and J. Lewsey. 2009. Is there overutilization of cataract surgery in England? British Journal of Ophthalmology 93: 13–17.
Thronton, T. 2006. Tacit knowledge as the unifying factor in evidence based medicine and clinical judgment. Philosophy, Ethics, and Humanities 1: 2.
Wittgenstein, L. 1958. Philosophical investigations. Trans. G.E.M. Anscombe. Oxford: Blackwell Press.
Kant, I. 1998. The critique of pure reason. Trans. P. Guyer and A.W. Wood. Cambridge: Cambridge University Press.
McDowell, J. 1984. Wittgenstein on following a rule. Synthese 58: 325–363.
Kuhn, T. 1996. The structure of scientific revolutions. 3rd ed. Chicago: Chicago University Press.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
McClimans, L.M., Browne, J. Choosing a patient-reported outcome measure. Theor Med Bioeth 32, 47–60 (2011). https://doi.org/10.1007/s11017-010-9163-8
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11017-010-9163-8