Evaluation of Chatbots in the Emergency Management of Avulsion Injuries

被引:0
|
作者
Mustuloglu, Seyma [1 ]
Deniz, Busra Pinar [1 ]
机构
[1] Mersin Univ, Fac Dent, Dept Paediat Dent, Mersin, Turkiye
关键词
artificial intelligence; chatbot; emergency; large language models; tooth avulsion;
D O I
10.1111/edt.13041
中图分类号
R78 [口腔科学];
学科分类号
1003 ;
摘要
BackgroundThis study assessed the accuracy and consistency of responses provided by six Artificial Intelligence (AI) applications, ChatGPT version 3.5 (OpenAI), ChatGPT version 4 (OpenAI), ChatGPT version 4.0 (OpenAI), Perplexity (Perplexity.AI), Gemini (Google), and Copilot (Bing), to questions related to emergency management of avulsed teeth.Materials and MethodsTwo pediatric dentists developed 18 true or false questions regarding dental avulsion and asked public chatbots for 3 days. The responses were recorded and compared with the correct answers. The SPSS program was used to calculate the obtained accuracies and their consistency.ResultsChatGPT 4.0 achieved the highest accuracy rate of 95.6% over the entire time frame, while Perplexity (Perplexity.AI) had the lowest accuracy rate of 67.2%. ChatGPT version 4.0 (OpenAI) was the only AI that achieved perfect agreement with real answers, except at noon on day 1. ChatGPT version 3.5 (OpenAI) was the AI that showed the weakest agreement (6 times).ConclusionsWith the exception of ChatGPT's paid version, 4.0, AI chatbots do not seem ready for use as the main resource in managing avulsed teeth during emergencies. It might prove beneficial to incorporate the International Association of Dental Traumatology (IADT) guidelines in chatbot databases, enhancing their accuracy and consistency.
引用
收藏
页数:8
相关论文
共 50 条