'GREAT PLEA' system proposed for responsible use of generative AI in healthcare

NewsGuard 100/100 Score

In a recent study published in the journal NPJ Digital Medicine, researchers reviewed current guidelines on the ethical implementation of Artificial Intelligence (AI) in military and healthcare applications. Their discussions focus on ‘generative AI,’ a novel technology aimed at efficiently generating information, and attempts to overcome currently prevalent limitations on the ethical use of the technology. They develop and propose a novel system for the ethical application of AI to military and clinical research, named the “GREAT PLEA.”

Image Credit: CHIEW / Shutterstock

AI and the need for ethics

Artificial Intelligence (AI) is the simulation of human intelligence processes by machines, especially computer systems. With the increasing computational power of modern hardware and the widespread adoption of smart devices, access to and the adoption of AI is higher than ever. AI has percolated into and revolutionized almost every aspect of modern human society, with machine learning tools seeing widespread utilization in online advertisements, scientific research, and military simulations.

The military sector particularly benefits from AI tools, with the ongoing conflicts between Russia and Ukraine expected to pave the way for autonomous AI-guided weapon systems. Unfortunately, the development of AI tools is rarely accompanied by an ethical evaluation of the system, with collateral damages to non-combatants or friend forces left unconsidered.

“Seeing the rapid emergence of AI and its applications in the military, the United States Department of Defense (DOD) disclosed ethical principles for AI in 2020.”

These principles comprise five critical aspects of AI-centric ethics: responsibility, equitable treatment, traceability, reliability, and governance. The North Atlantic Treaty Organization (NATO) adopted these principles and expanded them to include lawfulness, explainability, and bias mitigation. These ethical introductions highlight the caution prominent military bodies impose on AI tools and the steps they are taking to mitigate unwanted AI outcomes.

Similar to the military importance of AI tools, these technologies share similar merits and cautions in the medical sector. AI applications in medicine usually involve using these technologies to aid clinicians in disease diagnoses and suggest treatments. However, in some cases, AI has begun replacing entire departments of formerly human staff.

The figure illustrates the commonalities and differences in ethical principles between military and healthcare. In our assessment, traceability, reliability, lawfulness, accountability, governability, and equity are the ethical principles that both fields have in common. At the same time, ethical principles, such as empathy and privacy, are emphasized in healthcare, whereas ethical principles, such as national security and defense, are emphasized in the military.The figure illustrates the commonalities and differences in ethical principles between military and healthcare. In our assessment, traceability, reliability, lawfulness, accountability, governability, and equity are the ethical principles that both fields have in common. At the same time, ethical principles, such as empathy and privacy, are emphasized in healthcare, whereas ethical principles, such as national security and defense, are emphasized in the military.

Generative AI (GenAI) is a novel type of artificial intelligence (AI) designed to produce new content. It learns patterns from existing data and uses this knowledge to generate new outputs. Despite proving extremely useful in drug discovery, evidence-based medicine summarization, and equipment design, these models are designed to optimize desired ‘black and white’ outcomes without concern for any potential (‘grey’) collateral damages. This necessitates ethically driven curbs on AI results, which, unfortunately, remains under-discussed. Furthermore, malicious use of AI-developed technologies remains rarely considered, much less discussed.

Recently, the World Health Organization (WHO) released a document on ethical considerations in the medical adoption of AI. However, this document remains in its nascency with much work, both research and discussion, pending before trust in AI can sufficiently allow it to take over formerly human roles.

We propose the “GREAT PLEA” ethical principles for generative AI in healthcare, namely Governability, Reliability, Equity, Accountability, Traceability, Privacy, Lawfulness, Empathy, and Autonomy. The GREAT PLEA ethical principles demonstrate our great plea for the community to prioritize these ethical principles when implementing and utilizing generative AI in practical healthcare settings.We propose the “GREAT PLEA” ethical principles for generative AI in healthcare, namely Governability, Reliability, Equity, Accountability, Traceability, Privacy, Lawfulness, Empathy, and Autonomy. The GREAT PLEA ethical principles demonstrate our great plea for the community to prioritize these ethical principles when implementing and utilizing generative AI in practical healthcare settings.

The current scenario

Since discussions on ethics in AI in the medical field remain controversial and rarely touched upon, most research into bias reductions and algorithm optimizations is focused on the military sphere. Studies and reports, most notably by the RAND Corporation, have aimed to consolidate legal obligations (such as the Geneva Conventions and the Law of Armed Conflict [LOAC]) with public opinion in developing critical guidelines for the ethical development and use of AI algorithms, especially generative AI.

Ongoing military research proposes that, while AI can be used for most military applications, AI is currently not advanced enough to accurately distinguish between combatants and civilians, necessitating a Human-In-The-Loop (HITL) prior to the activation of lethal weapon systems. Simultaneously, the aforementioned principles laid out by NATO and the US DOD have allowed for the development of cutting-edge AI algorithms with ethical considerations baked into their code (for example, erring on the side of caution when collateral damage is a potential outcome of a proposed action).

“The success of these ethical principles has also been demonstrated through their ability to adopt and embed AI mindfully, taking into account AI’s potential dangers, which the Pentagon is determined to avoid.”

Public opinion in the US seems confounding, with some people viewing the autonomous military decision-making of AI systems favorably, while others do not reconcile with the idea of AI initiating combat action without human authorization.

“These results could be due to the perceived lack of accountability, which is considered something that could entirely negate the value of AI, as a fully autonomous system that makes its own decisions distances military operators or clinicians from the responsibility of the system’s actions.”

The GREAT PLEA for AI in clinical settings

In the present work, the authors develop and propose a novel system for the ethical use of generative AI in medical environments. Christened the “GREAT PLEA,” the system encapsulates and builds upon ethical documents published by the US DOD, the American Medical Association (AMA), the WHO, and the Coalition for Health AI (CHAI). NATO’s principles have been notably excluded due to their overdependence on AI’s adherence to International Military Law, which does not apply to the clinical setting. In all else, NATO’s recommendations are nearly identical to those prescribed by the US DOD principles.

The GREAT PLEA is an abbreviation for Governability, Reliability, Equity, Accountability, and Traceability, all of which are common between military and medical domains. Privacy, Lawfulness, Empathy, and Autonomy are concerns emphasized in the healthcare spectrum and form a novel set of considerations hitherto ignored in medical literature.

Since generative AI necessitates guidelines that consider misinformation, potential data bias, and generalized evaluation metrics, these nine principles form the foundation for future AI systems to exist and function with minimal human intervention.

These principles can be enforced through cooperation with lawmakers and the establishment of standards for developers and users, as well as a partnership with recognized governing bodies within the healthcare sector, such as the WHO or AMA.

Journal reference:
Hugo Francisco de Souza

Written by

Hugo Francisco de Souza

Hugo Francisco de Souza is a scientific writer based in Bangalore, Karnataka, India. His academic passions lie in biogeography, evolutionary biology, and herpetology. He is currently pursuing his Ph.D. from the Centre for Ecological Sciences, Indian Institute of Science, where he studies the origins, dispersal, and speciation of wetland-associated snakes. Hugo has received, amongst others, the DST-INSPIRE fellowship for his doctoral research and the Gold Medal from Pondicherry University for academic excellence during his Masters. His research has been published in high-impact peer-reviewed journals, including PLOS Neglected Tropical Diseases and Systematic Biology. When not working or writing, Hugo can be found consuming copious amounts of anime and manga, composing and making music with his bass guitar, shredding trails on his MTB, playing video games (he prefers the term ‘gaming’), or tinkering with all things tech.

Citations

Please use one of the following formats to cite this article in your essay, paper or report:

  • APA

    Francisco de Souza, Hugo. (2023, December 05). 'GREAT PLEA' system proposed for responsible use of generative AI in healthcare. News-Medical. Retrieved on April 27, 2024 from https://www.news-medical.net/news/20231205/GREAT-PLEA-system-proposed-for-responsible-use-of-generative-AI-in-healthcare.aspx.

  • MLA

    Francisco de Souza, Hugo. "'GREAT PLEA' system proposed for responsible use of generative AI in healthcare". News-Medical. 27 April 2024. <https://www.news-medical.net/news/20231205/GREAT-PLEA-system-proposed-for-responsible-use-of-generative-AI-in-healthcare.aspx>.

  • Chicago

    Francisco de Souza, Hugo. "'GREAT PLEA' system proposed for responsible use of generative AI in healthcare". News-Medical. https://www.news-medical.net/news/20231205/GREAT-PLEA-system-proposed-for-responsible-use-of-generative-AI-in-healthcare.aspx. (accessed April 27, 2024).

  • Harvard

    Francisco de Souza, Hugo. 2023. 'GREAT PLEA' system proposed for responsible use of generative AI in healthcare. News-Medical, viewed 27 April 2024, https://www.news-medical.net/news/20231205/GREAT-PLEA-system-proposed-for-responsible-use-of-generative-AI-in-healthcare.aspx.

Comments

The opinions expressed here are the views of the writer and do not necessarily reflect the views and opinions of News Medical.
Post a new comment
Post

While we only use edited and approved content for Azthena answers, it may on occasions provide incorrect responses. Please confirm any data provided with the related suppliers or authors. We do not provide medical advice, if you search for medical information you must always consult a medical professional before acting on any information provided.

Your questions, but not your email details will be shared with OpenAI and retained for 30 days in accordance with their privacy principles.

Please do not ask questions that use sensitive or confidential information.

Read the full Terms & Conditions.

You might also like...
Empowering Change: How Hologic is Shaping the Future of Women's Healthcare