Automated machine learning tool scans social media to detect health product risks

A new artificial intelligence tool can scan social media data to discover adverse events associated with consumer health products, according to a study published September 30th in the open-access journal PLOS Digital Health by John Ayers of the University of California, San Diego, U.S., and colleagues.

The constant post-market surveillance of the safety of consumer products is crucial for public health and safety. However, current adverse-event reporting systems for approved prescription medications and medical devices depend on voluntary submissions from doctors and manufactures to the U.S. Food and Drug Administration. The rapid growth in consumer health products, such as cannabis-derived products and dietary supplements, has led to the need for new adverse event detection systems.

In the new study, researchers tested the efficacy of a new automated machine learning tool, "Waldo," that can sift through social media text to find consumer descriptions of adverse events. The tool was tested on its ability to scan Reddit posts to find adverse events (AEs) of cannabis-derived products.

When compared to human AE annotations of a set of Reddit posts, Waldo had an accuracy of 99.7%, far better than a general-purpose ChatGPT chatbot that was given the same set of posts. In a broader dataset of 437,132 Reddit posts, Waldo identified 28,832 potential reports of harm. When the researchers manually validated a random sample of these posts, they found that 86% were true AEs. The team has made Waldo open-source so that anyone-researchers, clinicians, or regulators-can use it.

"Waldo represents a significant advancement in social media-based AE detection, achieving superior performance compared to existing approaches," the authors say. "Additionally, Waldo's automated approach has broad applicability beyond cannabis-derived products to other consumer health products that similarly lack regulatory oversight."

Waldo shows that the health experiences people share online are not just noise, they're valuable safety signals. By capturing these voices, we can surface real-world harms that are invisible to traditional reporting systems."

Karan Desai, Lead Author

John Ayers adds, "This project highlights how digital health tools can transform post-market surveillance. By making Waldo open-source, we're ensuring that anyone, from regulators to clinicians, can use it to protect patients."

Second author Vijay Tiyyala notes, "From a technical perspective, we demonstrated that a carefully trained model like RoBERTa can outperform state-of-the-art chatbots for AE detection. Waldo's accuracy was surprising and encouraging."

"By democratizing access to Waldo, the team hopes to accelerate open science and improve safety for patients."

Source:
Journal reference:

Desai, K. S., et al. (2025). Waldo: Automated discovery of adverse events from unstructured self reports. PLOS Digital Health. doi.org/10.1371/journal.pdig.0001011

Comments

The opinions expressed here are the views of the writer and do not necessarily reflect the views and opinions of News Medical.
Post a new comment
Post

While we only use edited and approved content for Azthena answers, it may on occasions provide incorrect responses. Please confirm any data provided with the related suppliers or authors. We do not provide medical advice, if you search for medical information you must always consult a medical professional before acting on any information provided.

Your questions, but not your email details will be shared with OpenAI and retained for 30 days in accordance with their privacy principles.

Please do not ask questions that use sensitive or confidential information.

Read the full Terms & Conditions.

You might also like...
Adolescents with mental health conditions show different social media habits