The intra- and inter-assessor reliability of measurement of functional outcome by lameness scoring in horses.
Authors: Fuller Catherine J, Bladon Bruce M, Driver Adam J, Barr Alistair R S
Journal: Veterinary journal (London, England : 1997)
Summary
# Editorial Summary Fuller and colleagues (2006) investigated whether lameness scoring—a foundational tool in equine clinical practice—produces consistent and reproducible results, examining both intra-assessor reliability (one clinician scoring repeatedly) and inter-assessor reliability (multiple clinicians scoring the same cases). A single veterinary surgeon initially assessed 19 lame horses on four separate occasions using a 0–10 scale from sound to non-weight-bearing lameness, with videotaped gait recordings subsequently scored by the same assessor on two further occasions and independently by three different veterinary surgeons. Whilst intra-assessor reliability proved robust—meaning individual clinicians can score their own cases consistently—inter-assessor reliability was marginal, suggesting significant variation between different practitioners when evaluating identical lame horses; encouragingly, however, global scoring of overall change in lameness throughout a treatment period demonstrated acceptable reliability across assessors. This work carries considerable weight for equine professionals designing or interpreting clinical trials, as it reveals that whilst a single clinician can reliably track lameness progression in an individual horse, direct comparison of absolute lameness scores between different assessors or across multi-centre studies requires careful interpretation and standardisation protocols.
Read the full abstract on PubMed
Practical Takeaways
- •Individual lameness scores on a 0-10 scale are reasonably consistent when the same assessor re-evaluates a horse, but different veterinarians may score the same horse somewhat differently—use the same veterinarian when possible for objective comparisons
- •Global assessment of whether a horse is improving or worsening overall is more reliable than single-point lameness scores, so focus clinical discussions on directional change rather than absolute numerical values
- •Lameness scoring is sufficiently reliable for tracking treatment progress in clinical practice, but standardizing assessment methods and preferring serial evaluations by the same clinician will improve accuracy
Key Findings
- •Intra-assessor reliability for individual lameness scoring (0-10 scale) was good when one veterinary surgeon scored videotapes on two occasions
- •Inter-assessor reliability for individual lameness scores was only just within acceptable limits when three veterinary surgeons independently scored the same videotapes
- •Global scoring of change in lameness throughout the study period was found to be reliable across assessors
- •Clinician-based lameness scoring shows acceptable reliability for measuring changes in lameness over time, supporting its use in clinical studies