Large language models outperform general practitioners in identifying complex cases of childhood anxiety
Objective Anxiety is prevalent in childhood but often remains undiagnosed due to its physical manifestations and significant comorbidity. Despite the availability of effective treatments, including medication and psychotherapy, research indicates that physicians struggle to identify childhood anxiet...
Saved in:
| Main Authors: | , , , |
|---|---|
| Format: | Article |
| Language: | English |
| Published: |
SAGE Publishing
2024-12-01
|
| Series: | Digital Health |
| Online Access: | https://doi.org/10.1177/20552076241294182 |
| Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
| _version_ | 1846120643471343616 |
|---|---|
| author | Inbar Levkovich Eyal Rabin Michal Brann Zohar Elyoseph |
| author_facet | Inbar Levkovich Eyal Rabin Michal Brann Zohar Elyoseph |
| author_sort | Inbar Levkovich |
| collection | DOAJ |
| description | Objective Anxiety is prevalent in childhood but often remains undiagnosed due to its physical manifestations and significant comorbidity. Despite the availability of effective treatments, including medication and psychotherapy, research indicates that physicians struggle to identify childhood anxiety, particularly in complex and challenging cases. This study aims to explore the potential effectiveness of artificial intelligence (AI) language models in diagnosing childhood anxiety compared to general practitioners (GPs). Methods During February 2024, we evaluated the ability of several large language models (LLMs; ChatGPT-3.5 and ChatGPT-4, Claude.AI, Gemini) to identify cases childhood anxiety disorder, compared with reports of GPs. Results AI tools exhibited significantly higher rates of identifying anxiety than GPs. Each AI tool accurately identified anxiety in at least one case: Claude.AI and Gemini identified at least four cases, ChatGPT-3 identified three cases, and ChatGPT-4 identified one or two cases. Additionally, 40% of GPs preferred to manage the cases within their practice, often with the help of a practice nurse, whereas AI tools generally recommended referral to specialized mental or somatic health services. Conclusion Preliminary findings indicate that LLMs, specifically Claude.AI and Gemini, exhibit notable diagnostic capabilities in identifying child anxiety, demonstrating a comparative advantage over GPs. |
| format | Article |
| id | doaj-art-3c3fd37e4ce24dc1b3a025b02769760e |
| institution | Kabale University |
| issn | 2055-2076 |
| language | English |
| publishDate | 2024-12-01 |
| publisher | SAGE Publishing |
| record_format | Article |
| series | Digital Health |
| spelling | doaj-art-3c3fd37e4ce24dc1b3a025b02769760e2024-12-16T08:04:08ZengSAGE PublishingDigital Health2055-20762024-12-011010.1177/20552076241294182Large language models outperform general practitioners in identifying complex cases of childhood anxietyInbar Levkovich0Eyal Rabin1Michal Brann2Zohar Elyoseph3 The Faculty of Education, , Upper Galilee, Israel Department of Psychology and Education, , Ra'anana, Israel Department of Psychology and Educational Counseling, , Afula, Israel Faculty of Education, , Haifa, IsraelObjective Anxiety is prevalent in childhood but often remains undiagnosed due to its physical manifestations and significant comorbidity. Despite the availability of effective treatments, including medication and psychotherapy, research indicates that physicians struggle to identify childhood anxiety, particularly in complex and challenging cases. This study aims to explore the potential effectiveness of artificial intelligence (AI) language models in diagnosing childhood anxiety compared to general practitioners (GPs). Methods During February 2024, we evaluated the ability of several large language models (LLMs; ChatGPT-3.5 and ChatGPT-4, Claude.AI, Gemini) to identify cases childhood anxiety disorder, compared with reports of GPs. Results AI tools exhibited significantly higher rates of identifying anxiety than GPs. Each AI tool accurately identified anxiety in at least one case: Claude.AI and Gemini identified at least four cases, ChatGPT-3 identified three cases, and ChatGPT-4 identified one or two cases. Additionally, 40% of GPs preferred to manage the cases within their practice, often with the help of a practice nurse, whereas AI tools generally recommended referral to specialized mental or somatic health services. Conclusion Preliminary findings indicate that LLMs, specifically Claude.AI and Gemini, exhibit notable diagnostic capabilities in identifying child anxiety, demonstrating a comparative advantage over GPs.https://doi.org/10.1177/20552076241294182 |
| spellingShingle | Inbar Levkovich Eyal Rabin Michal Brann Zohar Elyoseph Large language models outperform general practitioners in identifying complex cases of childhood anxiety Digital Health |
| title | Large language models outperform general practitioners in identifying complex cases of childhood anxiety |
| title_full | Large language models outperform general practitioners in identifying complex cases of childhood anxiety |
| title_fullStr | Large language models outperform general practitioners in identifying complex cases of childhood anxiety |
| title_full_unstemmed | Large language models outperform general practitioners in identifying complex cases of childhood anxiety |
| title_short | Large language models outperform general practitioners in identifying complex cases of childhood anxiety |
| title_sort | large language models outperform general practitioners in identifying complex cases of childhood anxiety |
| url | https://doi.org/10.1177/20552076241294182 |
| work_keys_str_mv | AT inbarlevkovich largelanguagemodelsoutperformgeneralpractitionersinidentifyingcomplexcasesofchildhoodanxiety AT eyalrabin largelanguagemodelsoutperformgeneralpractitionersinidentifyingcomplexcasesofchildhoodanxiety AT michalbrann largelanguagemodelsoutperformgeneralpractitionersinidentifyingcomplexcasesofchildhoodanxiety AT zoharelyoseph largelanguagemodelsoutperformgeneralpractitionersinidentifyingcomplexcasesofchildhoodanxiety |