Evaluating a large language model’s ability to answer clinicians’ requests for evidence summaries
Objective: This study investigated the performance of a generative artificial intelligence (AI) tool using GPT-4 in answering clinical questions in comparison with medical librarians’ gold-standard evidence syntheses. Methods: Questions were extracted from an in-house database of clinical evidenc...
Saved in:
Main Authors: | Mallory N. Blasingame, Taneya Y. Koonce, Annette M. Williams, Dario A. Giuse, Jing Su, Poppy A. Krump, Nunzia Bettinsoli Giuse |
---|---|
Format: | Article |
Language: | English |
Published: |
University Library System, University of Pittsburgh
2025-01-01
|
Series: | Journal of the Medical Library Association |
Subjects: | |
Online Access: | http://jmla.pitt.edu/ojs/jmla/article/view/1985 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Similar Items
-
Development of an open access systematic review instructional video series accessible through the SPI-HubTM website
by: Sheila Kusnoor, et al.
Published: (2025-01-01) -
Evaluating the effectiveness of prompt engineering for knowledge graph question answering
by: Catherine Kosten, et al.
Published: (2025-01-01) -
Large Language Models as Evaluators in Education: Verification of Feedback Consistency and Accuracy
by: Hyein Seo, et al.
Published: (2025-01-01) -
A Summary of the Best Evidence for Wet Pack Management
by: Yi L, et al.
Published: (2025-01-01) -
Summary of Evidence on Nutritional Management for Patients Undergoing Chemotherapy
by: Xin Dan, et al.
Published: (2024-12-01)