Close Menu
  • Home
  • AI Models
    • DeepSeek
    • xAI
    • OpenAI
    • Meta AI Llama
    • Google DeepMind
    • Amazon AWS AI
    • Microsoft AI
    • Anthropic (Claude)
    • NVIDIA AI
    • IBM WatsonX Granite 3.1
    • Adobe Sensi
    • Hugging Face
    • Alibaba Cloud (Qwen)
    • Baidu (ERNIE)
    • C3 AI
    • DataRobot
    • Mistral AI
    • Moonshot AI (Kimi)
    • Google Gemma
    • xAI
    • Stability AI
    • H20.ai
  • AI Research
    • Allen Institue for AI
    • arXiv AI
    • Berkeley AI Research
    • CMU AI
    • Google Research
    • Microsoft Research
    • Meta AI Research
    • OpenAI Research
    • Stanford HAI
    • MIT CSAIL
    • Harvard AI
  • AI Funding & Startups
    • AI Funding Database
    • CBInsights AI
    • Crunchbase AI
    • Data Robot Blog
    • TechCrunch AI
    • VentureBeat AI
    • The Information AI
    • Sifted AI
    • WIRED AI
    • Fortune AI
    • PitchBook
    • TechRepublic
    • SiliconANGLE – Big Data
    • MIT News
    • Data Robot Blog
  • Expert Insights & Videos
    • Google DeepMind
    • Lex Fridman
    • Matt Wolfe AI
    • Yannic Kilcher
    • Two Minute Papers
    • AI Explained
    • TheAIEdge
    • Matt Wolfe AI
    • The TechLead
    • Andrew Ng
    • OpenAI
  • Expert Blogs
    • François Chollet
    • Gary Marcus
    • IBM
    • Jack Clark
    • Jeremy Howard
    • Melanie Mitchell
    • Andrew Ng
    • Andrej Karpathy
    • Sebastian Ruder
    • Rachel Thomas
    • IBM
  • AI Policy & Ethics
    • ACLU AI
    • AI Now Institute
    • Center for AI Safety
    • EFF AI
    • European Commission AI
    • Partnership on AI
    • Stanford HAI Policy
    • Mozilla Foundation AI
    • Future of Life Institute
    • Center for AI Safety
    • World Economic Forum AI
  • AI Tools & Product Releases
    • AI Assistants
    • AI for Recruitment
    • AI Search
    • Coding Assistants
    • Customer Service AI
    • Image Generation
    • Video Generation
    • Writing Tools
    • AI for Recruitment
    • Voice/Audio Generation
  • Industry Applications
    • Finance AI
    • Healthcare AI
    • Legal AI
    • Manufacturing AI
    • Media & Entertainment
    • Transportation AI
    • Education AI
    • Retail AI
    • Agriculture AI
    • Energy AI
  • AI Art & Entertainment
    • AI Art News Blog
    • Artvy Blog » AI Art Blog
    • Weird Wonderful AI Art Blog
    • The Chainsaw » AI Art
    • Artvy Blog » AI Art Blog
What's Hot

IBM Launches Power11 Servers | Inside HPC & AI News

Alex Kendall of Wayve brings the future of autonomous AI to Disrupt 2025

Simulating Breaking Glass | Two Minute Papers #4

Facebook X (Twitter) Instagram
Advanced AI News
  • Home
  • AI Models
    • OpenAI (GPT-4 / GPT-4o)
    • Anthropic (Claude 3)
    • Google DeepMind (Gemini)
    • Meta (LLaMA)
    • Cohere (Command R)
    • Amazon (Titan)
    • IBM (Watsonx)
    • Inflection AI (Pi)
  • AI Research
    • Allen Institue for AI
    • arXiv AI
    • Berkeley AI Research
    • CMU AI
    • Google Research
    • Meta AI Research
    • Microsoft Research
    • OpenAI Research
    • Stanford HAI
    • MIT CSAIL
    • Harvard AI
  • AI Funding
    • AI Funding Database
    • CBInsights AI
    • Crunchbase AI
    • Data Robot Blog
    • TechCrunch AI
    • VentureBeat AI
    • The Information AI
    • Sifted AI
    • WIRED AI
    • Fortune AI
    • PitchBook
    • TechRepublic
    • SiliconANGLE – Big Data
    • MIT News
    • Data Robot Blog
  • AI Experts
    • Google DeepMind
    • Lex Fridman
    • Meta AI Llama
    • Yannic Kilcher
    • Two Minute Papers
    • AI Explained
    • TheAIEdge
    • The TechLead
    • Matt Wolfe AI
    • Andrew Ng
    • OpenAI
    • Expert Blogs
      • François Chollet
      • Gary Marcus
      • IBM
      • Jack Clark
      • Jeremy Howard
      • Melanie Mitchell
      • Andrew Ng
      • Andrej Karpathy
      • Sebastian Ruder
      • Rachel Thomas
      • IBM
  • AI Tools
    • AI Assistants
    • AI for Recruitment
    • AI Search
    • Coding Assistants
    • Customer Service AI
  • AI Policy
    • ACLU AI
    • AI Now Institute
    • Center for AI Safety
  • Industry AI
    • Finance AI
    • Healthcare AI
    • Education AI
    • Energy AI
    • Legal AI
LinkedIn Instagram YouTube Threads X (Twitter)
Advanced AI News
Center for AI Safety

Study: AI-Powered Research Prowess Now Outstrips Human Experts, Raising Bioweapon Risks

By Advanced AI EditorMay 14, 2025No Comments5 Mins Read
Share Facebook Twitter Pinterest Copy Link Telegram LinkedIn Tumblr Email
Share
Facebook Twitter LinkedIn Pinterest Email


Researchers from the Center for AI Safety (CAIS), MIT’s Media Lab, the Brazilian university UFABC, and the pandemic prevention non-profit SecureBio have found that leading artificial intelligence models can outperform experienced, PhD-level virologists in troubleshooting complex laboratory procedures.

The findings, detailed in a new study introducing the Virology Capabilities Test (VCT), demonstrate AI’s proficiency in specialized scientific tasks but also highlight serious dual-use concerns, suggesting these tools could lower the barrier for creating dangerous biological agents.

The VCT benchmark, consisting of 322 questions and detailed further in its research paper, was designed specifically to measure an AI’s ability to assist with intricate ‘wet lab’ virology protocols, assessing fundamental, visual, and tacit understanding – the kind of practical know-how often gained through hands-on lab experience.

The results showed OpenAI’s o3 model achieved 43.8% accuracy, substantially exceeding the 22.1% average scored by specialized human virologists answering questions within their fields. Google’s Gemini 2.5 Pro also performed strongly, scoring 37.6%. According to the VCT analysis, o3’s performance surpassed 94% of the human experts on tailored question subsets.

AI Virologist Chatbots Pose Dual-Use Dilemma

This emergent AI capability – providing expert-level guidance for sensitive lab work – presents a clear dual-use scenario: useful for accelerating legitimate research but potentially dangerous if misused. Seth Donoughe, a SecureBio research scientist and study co-author, conveyed his apprehension to TIME, stating the findings made him “little nervous.”

He elaborated on the historical context: “Throughout history, there are a fair number of cases where someone attempted to make a bioweapon—and one of the major reasons why they didn’t succeed is because they didn’t have access to the right level of expertise… So it seems worthwhile to be cautious about how these capabilities are being distributed.”

Reflecting this, the VCT researchers propose that this AI skill warrants inclusion within governance frameworks designed for dual-use life science technologies.

The VCT findings spurred immediate calls for action from safety advocates. Dan Hendrycks, director of the Center for AI Safety, stressed the need for immediate action, urging AI companies to implement robust safeguards within six months, calling inaction “reckless.”

He advocated for tiered or gated access controls as a potential mitigation strategy. “We want to give the people who have a legitimate use for asking how to manipulate deadly viruses—like a researcher at the MIT biology department—the ability to do so,” Hendrycks explained to TIME. “But random people who made an account a second ago don’t get those capabilities.”

Industry Responses and Calls for Oversight

Having been briefed on the VCT results months ago, AI developers have reacted differently. xAI, Elon Musk’s company, in February, published a risk management framework acknowledging the paper and mentioning potential virology safeguards for its Grok model, such as training it to decline harmful requests.

OpenAI stated it “deployed new system-level mitigations for biological risks” for its recently released o3 and o4-mini models, including specific measures like “blocking harmful outputs.”

This measure reportedly resulted from a “thousand-hour red-teaming campaign in which 98.7% of unsafe bio-related conversations were successfully flagged and blocked.” Red-teaming is a common security practice involving simulated attacks to find vulnerabilities. Anthropic, another leading AI lab, acknowledged the VCT results in its system documentation but offered no specific mitigation plans, while Google declined to comment on the matter to TIME.

However, some experts believe self-policing by the industry isn’t sufficient. Tom Inglesby from the Johns Hopkins Center for Health Security advocated for governmental policy and regulation. “The current situation is that the companies that are most virtuous are taking time and money to do this work, which is good for all of us, but other companies don’t have to do it,” he told TIME, adding, “That doesn’t make sense.” Inglesby proposed mandatory evaluations for new large language models before their release “to make sure it will not produce pandemic-level outcomes.”

AI’s Expanding Footprint in Scientific Research

The VCT results are not an isolated incident but rather a stark data point within a broader landscape where AI is rapidly integrating into specialized scientific fields. OpenAI, creator of the top-performing o3 model, was already known to be exploring biological applications; Winbuzzer reported in January on its collaboration with Retro Biosciences using a model named GPT-4b Micro to optimize proteins involved in stem cell creation.

Similarly, Google DeepMind has been highly active. Besides the Gemini model family, its widely used AlphaFold program predicts protein structures, while an “AI Co-Scientist” project, detailed in February, aims to generate novel scientific hypotheses, sometimes mirroring unpublished human research.

Microsoft entered the fray in February with BioEmu-1, a model focused on predicting the dynamic movement of proteins, complementing AlphaFold’s static predictions. These tools, focusing on protein engineering, hypothesis generation, and molecular simulation, illustrate AI’s expanding role, moving beyond data analysis toward complex scientific reasoning and procedural assistance – amplifying both the potential scientific gains and the safety challenges highlighted by the VCT.



Source link

Follow on Google News Follow on Flipboard
Share. Facebook Twitter Pinterest LinkedIn Tumblr Email Copy Link
Previous ArticleCohere Health Secures $90M Series C to Expand AI-Powered Platform Transforming Health Plan Clinical Decision-Making
Next Article OM1’s PhenOM® Foundation AI Surpasses One Billion Years of Health History in Model Training
Advanced AI Editor
  • Website

Related Posts

Researchers alarmed as AI begins to lie, scheme and threaten

July 7, 2025

Honeywell Launches AI-Powered Cybersecurity Solutions for Industrial Autonomy, ET CIO

June 26, 2025

How autonomous truck developers are pushing forward AI safety research boundaries

June 24, 2025
Leave A Reply

Latest Posts

Prospect New Orleans Will Not Mount Next Edition in 2027

Rising Painter Dies at 43

Confederate Group Sues Stone Mountain Over Show on Racism and Slavery

UK MPs to Debate Banning Advertising by Oil Companies

Latest Posts

IBM Launches Power11 Servers | Inside HPC & AI News

July 8, 2025

Alex Kendall of Wayve brings the future of autonomous AI to Disrupt 2025

July 8, 2025

Simulating Breaking Glass | Two Minute Papers #4

July 8, 2025

Subscribe to News

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

Recent Posts

  • IBM Launches Power11 Servers | Inside HPC & AI News
  • Alex Kendall of Wayve brings the future of autonomous AI to Disrupt 2025
  • Simulating Breaking Glass | Two Minute Papers #4
  • Fear Creates Unfinished Work | Recruiting News Network
  • Former OpenAI board member questions Mark Zuckerberg AI hiring spree – East Bay Times

Recent Comments

No comments to show.

Welcome to Advanced AI News—your ultimate destination for the latest advancements, insights, and breakthroughs in artificial intelligence.

At Advanced AI News, we are passionate about keeping you informed on the cutting edge of AI technology, from groundbreaking research to emerging startups, expert insights, and real-world applications. Our mission is to deliver high-quality, up-to-date, and insightful content that empowers AI enthusiasts, professionals, and businesses to stay ahead in this fast-evolving field.

Subscribe to Updates

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

LinkedIn Instagram YouTube Threads X (Twitter)
  • Home
  • About Us
  • Advertise With Us
  • Contact Us
  • DMCA
  • Privacy Policy
  • Terms & Conditions
© 2025 advancedainews. Designed by advancedainews.

Type above and press Enter to search. Press Esc to cancel.