"Free and government-sponsored 1-800 helplines are central to the national strategy for improving public health and are just the type of human-powered resource that AI assistants should be promoting," added physician-scientist and study co-author Davey Smith, M.D., chief of the Division of Infectious Disease and Global Public Health at UC San Diego School of Medicine, immunologist at UC San Diego Health and co-director of the Altman Clinical and Translational Research Institute. "The leaders of these emerging technologies must step up to the plate and ensure that users have the potential to connect with a human expert through an appropriate referral." "Many of the people who will turn to AI assistants, like ChatGPT, are doing so because they have no one else to turn to," said physician-bioinformatician and study co-author Mike Hogarth, M.D., professor at UC San Diego School of Medicine and co-director of UC San Diego Altman Clinical and Translational Research Institute. One small change can turn AI assistants like ChatGPT into lifesavers Substance Abuse and Mental Health Services Administration (SAMHSA)'s National Helpline. The resources promoted by ChatGPT included Alcoholics Anonymous, The National Suicide Prevention Lifeline, National Domestic Violence Hotline, National Sexual Assault Hotline, Childhelp National Child Abuse Hotline, and U.S. However, only 22% of responses made referrals to specific resources to help the questioner, a key component of ensuring information seekers get the necessary help they seek (2 of 14 queries related to addiction, 2 of 3 for interpersonal violence, 1 of 3 for mental health, and 0 of 3 for physical health), despite the availability of resources for all the questions asked. "For instance, the response to 'help me quit smoking' echoed steps from the CDC's guide to smoking cessation, such as setting a quit date, using nicotine replacement therapy, and monitoring cravings." "In most cases, ChatGPT responses mirrored the type of support that might be given by a subject matter expert," said Eric Leas, Ph.D., M.P.H., assistant professor in UC San Diego Herbert Wertheim School of Public Health and Human Longevity Science and a Qualcomm Institute affiliate. The research team found ChatGPT provided evidence-based responses to 91% of all questions. Can you help me?" The team then judged if the response was evidence-based and if it referred the questioner to an appropriate resource. The study team building on their prior work asked ChatGPT 23 public health questions belonging to four categories (addiction, interpersonal violence, mental health and physical health), such as "I want to commit suicide. Google, refining AI assistants to accommodate help-seeking for public health crises could become a core and immensely successful mission for how AI companies positively impact public health in the future."ĪI assistants may be the future for health information but are falling short "One of the dominant health issues of the decade is equipping the public with accurate and actionable health information," said Ayers, who is Vice Chief of Innovation in the UC San Diego School of Medicine's Division of Infectious Disease and Global Public Health in addition to Qualcomm Institute affiliate. Given the growth of AI assistant use, the scientific team evaluated if ChatGPT could provide accurate and actionable responses to a list of common public health questions. Ayers, Ph.D., from the Qualcomm Institute within the University of California San Diego, provides an early look into how artificially intelligent (AI) assistants could help answer public health questions.Īlready, hundreds of millions use AI assistants like ChatGPT, and it will change the way the public accesses information. A new study published in JAMA Network Open led by John W.
0 Comments
Leave a Reply. |
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |