Assessing the Quality of Decision Support Technologies Using the International Patient Decision Aid Standards instrument (IPDASi) Journal Articles uri icon

  •  
  • Overview
  •  
  • Research
  •  
  • Identity
  •  
  • Additional Document Info
  •  
  • View All
  •  

abstract

  • OBJECTIVES: To describe the development, validation and inter-rater reliability of an instrument to measure the quality of patient decision support technologies (decision aids). DESIGN: Scale development study, involving construct, item and scale development, validation and reliability testing. SETTING: There has been increasing use of decision support technologies--adjuncts to the discussions clinicians have with patients about difficult decisions. A global interest in developing these interventions exists among both for-profit and not-for-profit organisations. It is therefore essential to have internationally accepted standards to assess the quality of their development, process, content, potential bias and method of field testing and evaluation. METHODS: Scale development study, involving construct, item and scale development, validation and reliability testing. PARTICIPANTS: Twenty-five researcher-members of the International Patient Decision Aid Standards Collaboration worked together to develop the instrument (IPDASi). In the fourth Stage (reliability study), eight raters assessed thirty randomly selected decision support technologies. RESULTS: IPDASi measures quality in 10 dimensions, using 47 items, and provides an overall quality score (scaled from 0 to 100) for each intervention. Overall IPDASi scores ranged from 33 to 82 across the decision support technologies sampled (n = 30), enabling discrimination. The inter-rater intraclass correlation for the overall quality score was 0.80. Correlations of dimension scores with the overall score were all positive (0.31 to 0.68). Cronbach's alpha values for the 8 raters ranged from 0.72 to 0.93. Cronbach's alphas based on the dimension means ranged from 0.50 to 0.81, indicating that the dimensions, although well correlated, measure different aspects of decision support technology quality. A short version (19 items) was also developed that had very similar mean scores to IPDASi and high correlation between short score and overall score 0.87 (CI 0.79 to 0.92). CONCLUSIONS: This work demonstrates that IPDASi has the ability to assess the quality of decision support technologies. The existing IPDASi provides an assessment of the quality of a DST's components and will be used as a tool to provide formative advice to DSTs developers and summative assessments for those who want to compare their tools against an existing benchmark.

authors

  • Elwyn, Glyn
  • O'Connor, Annette M
  • Bennett, Carol
  • Newcombe, Robert G
  • Politi, Mary
  • Durand, Marie-Anne
  • Drake, Elizabeth
  • Joseph-Williams, Natalie
  • Khangura, Sara
  • Saarimaki, Anton
  • Sivell, Stephanie
  • Stiel, Mareike
  • Bernstein, Steven J
  • Col, Nananda
  • Coulter, Angela
  • Eden, Karen
  • Härter, Martin
  • Rovner, Margaret Holmes
  • Moumjid, Nora
  • Stacey, Dawn
  • Thomson, Richard
  • Whelan, Timothy
  • van der Weijden, Trudy
  • Edwards, Adrian

publication date

  • 2009