Building flexible governance for biological data powering AI systems

In a Policy Forum, Doni Bloomfield and colleagues discuss the need for expanded – yet tailored and flexible – governance for the biological data used to develop powerful artificial intelligence (AI) models.

Rapidly advancing AI systems trained on biological data have enabled researchers to design new molecules, predict protein structure and function, and probe vast and highly complex biological datasets for novel insights that could greatly expand our understanding of nature and human health. However, these same tools could also be misused for dangerous purposes, such as designing harmful pathogens or generating genetic sequences that bypass safety checks. Despite these widely recognized risks, current governance is severely lacking, and increasingly powerful models are often released without safety evaluation.

Here, Bloomfield et al. discuss how biological data governance could be achieved to both mitigate potential risks of biological AI systems without impeding their research potential. Just as researchers accept limits on access to personal information in genetic datasets in order to protect privacy without halting research, say the authors, similar frameworks could restrict only a narrow class of especially sensitive pathogen data while leaving most scientific data openly available. Such targeted controls would make it harder for malicious actors to obtain the rare and costly datasets needed to train dangerous models, without significantly impeding legitimate research, especially if paired with secure digital research environments.

Bloomfield et al. argue that this oversight should, however, remain limited, targeted, and flexible so that governance frameworks can adapt as required to keep up with both technological and scientific advancements. Moreover, to prevent abuse or excessive bureaucratic control, the research community should have the ability to appeal data classifications, and governing agencies should promise to ensure fast, transparent review processes so that much-needed safety measures do not become obstacles to legitimate scientific processes. "Formalizing a system of data access would allow researchers to scrutinize and develop these controls and would give scientists and companies clarity in an environment that is currently somewhat unpredictable," write the authors. "Starting this work will also allow scientists and governments to learn more about the nature of AI risk and revise data-access controls in light of tangible evidence, rather than guesswork."

Source:
Journal reference:

Bloomfield, D., et al. (2026). Biological data governance in an age of AI. Science. DOI: 10.1126/science.aeb2689. https://www.science.org/doi/10.1126/science.aeb2689

Comments

The opinions expressed here are the views of the writer and do not necessarily reflect the views and opinions of News Medical.
Post a new comment
Post

While we only use edited and approved content for Azthena answers, it may on occasions provide incorrect responses. Please confirm any data provided with the related suppliers or authors. We do not provide medical advice, if you search for medical information you must always consult a medical professional before acting on any information provided.

Your questions, but not your email details will be shared with OpenAI and retained for 30 days in accordance with their privacy principles.

Please do not ask questions that use sensitive or confidential information.

Read the full Terms & Conditions.

You might also like...
Study maps highest hemochromatosis risk in Celtic populations