Examining the role of ChatGPT in the management of distal radius fractures: insights into its accuracy and consistency
View abstract on PubMed
Summary
This summary is machine-generated.ChatGPT shows significant limitations in accuracy and consistency for managing distal radius fractures. Its current capabilities offer limited utility for patient education and clinical decision-making in orthopaedics.
Area Of Science
- Orthopaedic Surgery
- Artificial Intelligence in Medicine
- Medical Information Systems
Background
- Distal radius fractures present ongoing management challenges for orthopaedic surgeons.
- Artificial Intelligence (AI) and Large Language Models (LLMs) show potential to enhance healthcare and research.
- ChatGPT, a prominent LLM, is being explored for its application in medical contexts.
Purpose Of The Study
- To evaluate the accuracy and consistency of ChatGPT's knowledge regarding distal radius fracture management.
- To assess ChatGPT's suitability for patient education and orthopaedic clinical decision support.
Main Methods
- ChatGPT was queried with seven questions on distal radius fracture management across two sessions, yielding 14 responses.
- Responses covered patient inquiries and clinical decision-making scenarios.
- Orthopaedic registrars and senior surgeons evaluated the accuracy, consistency, and reference validity of ChatGPT's outputs.
Main Results
- All 14 responses contained a mixture of accurate and inaccurate information.
- Of 47 cited references, only 13% were accurate; 28% were fabricated, and 57% were incorrect.
- Response consistency was observed in 71% of cases, but overall accuracy was limited.
Conclusions
- ChatGPT exhibits significant limitations in accuracy and consistency for information on distal radius fractures.
- Current ChatGPT performance restricts its utility for patient education and clinical decision-making in this orthopaedic domain.

