AI vs. Junior Residents: ChatGPT’s Role in Streamlining Patient History-Taking
Source: Graduate Institute of Data Science
Published on 2025-03-04
Summary
This study demonstrates that ChatGPT, particularly version 4.0, can match or exceed junior medical residents in medical documentation tasks during clinical simulations, offering promise for reducing healthcare administrative burdens while highlighting the need for human oversight to ensure accuracy and reliability.
In a study comparing junior medical interns to ChatGPT, researchers from Taipei Medical University (TMU) , have found that while AI excels at creating accurate and complete medical records, it struggles to match human clinicians’ empathy and nuanced communication. These findings highlight both the potential and the limitations of AI in healthcare.
The Novelty (What)
This research demonstrates that ChatGPT (versions 3.5 and 4.0) can match, and sometimes exceed, the performance of junior medical residents in medical documentation tasks during simulated clinical examinations. In a clinical simulation with standardized patients, the study revealed that ChatGPT 4.0 excels in producing grammatically accurate and comprehensive medical records, often achieving higher scores on documentation metrics than human counterparts. However, while its performance in documentation is promising, ChatGPT struggled with capturing critical “negative findings” and generated occasional irrelevant or fabricated information, underscoring the need for human oversight. The study highlights the potential for AI to support medical professionals in administrative and documentation tasks, with future possibilities for improving AI’s diagnostic accuracy and empathetic capabilities.

Comparative analysis on performance on the Objective Structured Clinical Examination, quality of medical records, and patient interview. LLM: large language model.

Comparative analysis on 5 metrics of medical documentation with IQR. LLM: large language model.
The Background (Why)
Medical documentation, a critical aspect of healthcare delivery, ensures continuity of care but often places a significant administrative burden on medical professionals, contributing to burnout and inefficiencies. Artificial intelligence (AI) systems like ChatGPT are increasingly explored for their potential to assist in such tasks. Yet, prior research has not adequately evaluated AI’s performance in real-world, clinically simulated environments. Past studies have typically focused on theoretical applications or narrow capabilities, leaving a gap in understanding how AI compares directly to human clinicians in patient history-taking and documentation. This study addresses these limitations using a simulated clinical model to assess ChatGPT’s effectiveness in tasks traditionally performed by junior medical residents. By demonstrating ChatGPT’s strengths and weaknesses, the research provides insights into how AI can serve as a supportive tool in healthcare, alleviating administrative workloads while maintaining the quality and reliability of medical documentation.

Clinician’s feedback on ChatGPT.
The SDG impact (Big Why)
Healthcare systems worldwide face increasing documentation demands, contributing to provider burnout and inefficiencies. This urgent issue requires innovative solutions to reduce inefficiencies and improve provider well-being. This study aligns with Sustainable Development Goal 3 (Good Health and Well-being) by exploring the potential of artificial intelligence to streamline medical documentation, allowing healthcare providers to focus more on patient care. Additionally, it supports SDG 9 (Industry, Innovation, and Infrastructure) by advancing technological innovation in healthcare, strengthening health systems, and fostering resilient infrastructure. Integrating AI tools like ChatGPT can enhance healthcare delivery efficiency while preserving the essential human elements of empathy and patient-centered care.
Look for More Information