Development and testing of an assessment instrument for the formative peer review of significant event analyses
Autor: | John McKay, Schmuck Ml, Murray Lough, Murphy Dj, Eva Kw, Paul Bowie |
---|---|
Rok vydání: | 2007 |
Předmět: |
Safety Management
Quality Assurance Health Care Leadership and Management media_common.quotation_subject education Applied psychology Validity Formative assessment Content validity Humans Medicine Quality (business) General Nursing Reliability (statistics) Simulation media_common Analysis of Variance business.industry Health Policy Public Health Environmental and Occupational Health Reproducibility of Results Variance (accounting) Scotland Global Positioning System Developing Research and Practice Family Practice business Quality assurance |
Zdroj: | Quality and Safety in Health Care. 16:150-153 |
ISSN: | 1475-3901 1475-3898 |
DOI: | 10.1136/qshc.2006.020750 |
Popis: | Aim: To establish the content validity and specific aspects of reliability for an assessment instrument designed to provide formative feedback to general practitioners (GPs) on the quality of their written analysis of a significant event. Methods: Content validity was quantified by application of a content validity index. Reliability testing involved a nested design, with 5 cells, each containing 4 assessors, rating 20 unique significant event analysis (SEA) reports (10 each from experienced GPs and GPs in training) using the assessment instrument. The variance attributable to each identified variable in the study was established by analysis of variance. Generalisability theory was then used to investigate the instrument’s ability to discriminate among SEA reports. Results: Content validity was demonstrated with at least 8 of 10 experts endorsing all 10 items of the assessment instrument. The overall G coefficient for the instrument was moderate to good (G>0.70), indicating that the instrument can provide consistent information on the standard achieved by the SEA report. There was moderate inter-rater reliability (G>0.60) when four raters were used to judge the quality of the SEA. Conclusions: This study provides the first steps towards validating an instrument that can provide educational feedback to GPs on their analysis of significant events. The key area identified to improve instrument reliability is variation among peer assessors in their assessment of SEA reports. Further validity and reliability testing should be carried out to provide GPs, their appraisers and contractual bodies with a validated feedback instrument on this aspect of the general practice quality agenda. |
Databáze: | OpenAIRE |
Externí odkaz: |