Close Menu
  • Home
  • AI Models
    • DeepSeek
    • xAI
    • OpenAI
    • Meta AI Llama
    • Google DeepMind
    • Amazon AWS AI
    • Microsoft AI
    • Anthropic (Claude)
    • NVIDIA AI
    • IBM WatsonX Granite 3.1
    • Adobe Sensi
    • Hugging Face
    • Alibaba Cloud (Qwen)
    • Baidu (ERNIE)
    • C3 AI
    • DataRobot
    • Mistral AI
    • Moonshot AI (Kimi)
    • Google Gemma
    • xAI
    • Stability AI
    • H20.ai
  • AI Research
    • Allen Institue for AI
    • arXiv AI
    • Berkeley AI Research
    • CMU AI
    • Google Research
    • Microsoft Research
    • Meta AI Research
    • OpenAI Research
    • Stanford HAI
    • MIT CSAIL
    • Harvard AI
  • AI Funding & Startups
    • AI Funding Database
    • CBInsights AI
    • Crunchbase AI
    • Data Robot Blog
    • TechCrunch AI
    • VentureBeat AI
    • The Information AI
    • Sifted AI
    • WIRED AI
    • Fortune AI
    • PitchBook
    • TechRepublic
    • SiliconANGLE – Big Data
    • MIT News
    • Data Robot Blog
  • Expert Insights & Videos
    • Google DeepMind
    • Lex Fridman
    • Matt Wolfe AI
    • Yannic Kilcher
    • Two Minute Papers
    • AI Explained
    • TheAIEdge
    • Matt Wolfe AI
    • The TechLead
    • Andrew Ng
    • OpenAI
  • Expert Blogs
    • François Chollet
    • Gary Marcus
    • IBM
    • Jack Clark
    • Jeremy Howard
    • Melanie Mitchell
    • Andrew Ng
    • Andrej Karpathy
    • Sebastian Ruder
    • Rachel Thomas
    • IBM
  • AI Policy & Ethics
    • ACLU AI
    • AI Now Institute
    • Center for AI Safety
    • EFF AI
    • European Commission AI
    • Partnership on AI
    • Stanford HAI Policy
    • Mozilla Foundation AI
    • Future of Life Institute
    • Center for AI Safety
    • World Economic Forum AI
  • AI Tools & Product Releases
    • AI Assistants
    • AI for Recruitment
    • AI Search
    • Coding Assistants
    • Customer Service AI
    • Image Generation
    • Video Generation
    • Writing Tools
    • AI for Recruitment
    • Voice/Audio Generation
  • Industry Applications
    • Finance AI
    • Healthcare AI
    • Legal AI
    • Manufacturing AI
    • Media & Entertainment
    • Transportation AI
    • Education AI
    • Retail AI
    • Agriculture AI
    • Energy AI
  • AI Art & Entertainment
    • AI Art News Blog
    • Artvy Blog » AI Art Blog
    • Weird Wonderful AI Art Blog
    • The Chainsaw » AI Art
    • Artvy Blog » AI Art Blog
What's Hot

Abstracts: Heat Transfer and Deep Learning with Hongxia Hao and Bing Lv

Stanford HAI’s annual report highlights rapid adoption and growing accessibility of powerful AI systems

IBM launches new LinuxONE server with Telum II chips

Facebook X (Twitter) Instagram
Advanced AI News
  • Home
  • AI Models
    • Adobe Sensi
    • Aleph Alpha
    • Alibaba Cloud (Qwen)
    • Amazon AWS AI
    • Anthropic (Claude)
    • Apple Core ML
    • Baidu (ERNIE)
    • ByteDance Doubao
    • C3 AI
    • Cohere
    • DataRobot
    • DeepSeek
  • AI Research & Breakthroughs
    • Allen Institue for AI
    • arXiv AI
    • Berkeley AI Research
    • CMU AI
    • Google Research
    • Meta AI Research
    • Microsoft Research
    • OpenAI Research
    • Stanford HAI
    • MIT CSAIL
    • Harvard AI
  • AI Funding & Startups
    • AI Funding Database
    • CBInsights AI
    • Crunchbase AI
    • Data Robot Blog
    • TechCrunch AI
    • VentureBeat AI
    • The Information AI
    • Sifted AI
    • WIRED AI
    • Fortune AI
    • PitchBook
    • TechRepublic
    • SiliconANGLE – Big Data
    • MIT News
    • Data Robot Blog
  • Expert Insights & Videos
    • Google DeepMind
    • Lex Fridman
    • Meta AI Llama
    • Yannic Kilcher
    • Two Minute Papers
    • AI Explained
    • TheAIEdge
    • Matt Wolfe AI
    • The TechLead
    • Andrew Ng
    • OpenAI
  • Expert Blogs
    • François Chollet
    • Gary Marcus
    • IBM
    • Jack Clark
    • Jeremy Howard
    • Melanie Mitchell
    • Andrew Ng
    • Andrej Karpathy
    • Sebastian Ruder
    • Rachel Thomas
    • IBM
  • AI Policy & Ethics
    • ACLU AI
    • AI Now Institute
    • Center for AI Safety
    • EFF AI
    • European Commission AI
    • Partnership on AI
    • Stanford HAI Policy
    • Mozilla Foundation AI
    • Future of Life Institute
    • Center for AI Safety
    • World Economic Forum AI
  • AI Tools & Product Releases
    • AI Assistants
    • AI for Recruitment
    • AI Search
    • Coding Assistants
    • Customer Service AI
    • Image Generation
    • Video Generation
    • Writing Tools
    • AI for Recruitment
    • Voice/Audio Generation
  • Industry Applications
    • Education AI
    • Energy AI
    • Finance AI
    • Healthcare AI
    • Legal AI
    • Media & Entertainment
    • Transportation AI
    • Manufacturing AI
    • Retail AI
    • Agriculture AI
  • AI Art & Entertainment
    • AI Art News Blog
    • Artvy Blog » AI Art Blog
    • Weird Wonderful AI Art Blog
    • The Chainsaw » AI Art
    • Artvy Blog » AI Art Blog
Advanced AI News
Home » Study: AI-Powered Research Prowess Now Outstrips Human Experts, Raising Bioweapon Risks
Center for AI Safety

Study: AI-Powered Research Prowess Now Outstrips Human Experts, Raising Bioweapon Risks

Advanced AI BotBy Advanced AI BotMay 8, 2025No Comments5 Mins Read
Share Facebook Twitter Pinterest Copy Link Telegram LinkedIn Tumblr Email
Share
Facebook Twitter LinkedIn Pinterest Email


Researchers from the Center for AI Safety (CAIS), MIT’s Media Lab, the Brazilian university UFABC, and the pandemic prevention non-profit SecureBio have found that leading artificial intelligence models can outperform experienced, PhD-level virologists in troubleshooting complex laboratory procedures.

The findings, detailed in a new study introducing the Virology Capabilities Test (VCT), demonstrate AI’s proficiency in specialized scientific tasks but also highlight serious dual-use concerns, suggesting these tools could lower the barrier for creating dangerous biological agents.

The VCT benchmark, consisting of 322 questions and detailed further in its research paper, was designed specifically to measure an AI’s ability to assist with intricate ‘wet lab’ virology protocols, assessing fundamental, visual, and tacit understanding – the kind of practical know-how often gained through hands-on lab experience.

The results showed OpenAI’s o3 model achieved 43.8% accuracy, substantially exceeding the 22.1% average scored by specialized human virologists answering questions within their fields. Google’s Gemini 2.5 Pro also performed strongly, scoring 37.6%. According to the VCT analysis, o3’s performance surpassed 94% of the human experts on tailored question subsets.

AI Virologist Chatbots Pose Dual-Use Dilemma

This emergent AI capability – providing expert-level guidance for sensitive lab work – presents a clear dual-use scenario: useful for accelerating legitimate research but potentially dangerous if misused. Seth Donoughe, a SecureBio research scientist and study co-author, conveyed his apprehension to TIME, stating the findings made him “little nervous.”

He elaborated on the historical context: “Throughout history, there are a fair number of cases where someone attempted to make a bioweapon—and one of the major reasons why they didn’t succeed is because they didn’t have access to the right level of expertise… So it seems worthwhile to be cautious about how these capabilities are being distributed.”

Reflecting this, the VCT researchers propose that this AI skill warrants inclusion within governance frameworks designed for dual-use life science technologies.

The VCT findings spurred immediate calls for action from safety advocates. Dan Hendrycks, director of the Center for AI Safety, stressed the need for immediate action, urging AI companies to implement robust safeguards within six months, calling inaction “reckless.”

He advocated for tiered or gated access controls as a potential mitigation strategy. “We want to give the people who have a legitimate use for asking how to manipulate deadly viruses—like a researcher at the MIT biology department—the ability to do so,” Hendrycks explained to TIME. “But random people who made an account a second ago don’t get those capabilities.”

Industry Responses and Calls for Oversight

Having been briefed on the VCT results months ago, AI developers have reacted differently. xAI, Elon Musk’s company, in February, published a risk management framework acknowledging the paper and mentioning potential virology safeguards for its Grok model, such as training it to decline harmful requests.

OpenAI stated it “deployed new system-level mitigations for biological risks” for its recently released o3 and o4-mini models, including specific measures like “blocking harmful outputs.”

This measure reportedly resulted from a “thousand-hour red-teaming campaign in which 98.7% of unsafe bio-related conversations were successfully flagged and blocked.” Red-teaming is a common security practice involving simulated attacks to find vulnerabilities. Anthropic, another leading AI lab, acknowledged the VCT results in its system documentation but offered no specific mitigation plans, while Google declined to comment on the matter to TIME.

However, some experts believe self-policing by the industry isn’t sufficient. Tom Inglesby from the Johns Hopkins Center for Health Security advocated for governmental policy and regulation. “The current situation is that the companies that are most virtuous are taking time and money to do this work, which is good for all of us, but other companies don’t have to do it,” he told TIME, adding, “That doesn’t make sense.” Inglesby proposed mandatory evaluations for new large language models before their release “to make sure it will not produce pandemic-level outcomes.”

AI’s Expanding Footprint in Scientific Research

The VCT results are not an isolated incident but rather a stark data point within a broader landscape where AI is rapidly integrating into specialized scientific fields. OpenAI, creator of the top-performing o3 model, was already known to be exploring biological applications; Winbuzzer reported in January on its collaboration with Retro Biosciences using a model named GPT-4b Micro to optimize proteins involved in stem cell creation.

Similarly, Google DeepMind has been highly active. Besides the Gemini model family, its widely used AlphaFold program predicts protein structures, while an “AI Co-Scientist” project, detailed in February, aims to generate novel scientific hypotheses, sometimes mirroring unpublished human research.

Microsoft entered the fray in February with BioEmu-1, a model focused on predicting the dynamic movement of proteins, complementing AlphaFold’s static predictions. These tools, focusing on protein engineering, hypothesis generation, and molecular simulation, illustrate AI’s expanding role, moving beyond data analysis toward complex scientific reasoning and procedural assistance – amplifying both the potential scientific gains and the safety challenges highlighted by the VCT.



Source link

Follow on Google News Follow on Flipboard
Share. Facebook Twitter Pinterest LinkedIn Tumblr Email Copy Link
Previous ArticleX Money Payment System hits roadblock in New York
Next Article RSAC 2025: Cisco and Meta put open-source AI at the heart of threat defense
Advanced AI Bot
  • Website

Related Posts

Study: AI-Powered Research Prowess Now Outstrips Human Experts, Raising Bioweapon Risks

May 8, 2025

Study: AI-Powered Research Prowess Now Outstrips Human Experts, Raising Bioweapon Risks

May 8, 2025

Study: AI-Powered Research Prowess Now Outstrips Human Experts, Raising Bioweapon Risks

May 8, 2025
Leave A Reply Cancel Reply

Latest Posts

Beyond ‘Love,’ The Enduring Legacy Of Robert Indiana Resonates Deeply Through Pace Gallery Representation

Ancient Greek Author and Title of Charred Herculaneum Scroll Revealed

Bonhams To Auction Museum Quality Work from The Holly Solomon Collection.

Justin Bateman Turns Stones Into Ephemeral Art

Latest Posts

Abstracts: Heat Transfer and Deep Learning with Hongxia Hao and Bing Lv

May 8, 2025

Stanford HAI’s annual report highlights rapid adoption and growing accessibility of powerful AI systems

May 8, 2025

IBM launches new LinuxONE server with Telum II chips

May 8, 2025

Subscribe to News

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

Welcome to Advanced AI News—your ultimate destination for the latest advancements, insights, and breakthroughs in artificial intelligence.

At Advanced AI News, we are passionate about keeping you informed on the cutting edge of AI technology, from groundbreaking research to emerging startups, expert insights, and real-world applications. Our mission is to deliver high-quality, up-to-date, and insightful content that empowers AI enthusiasts, professionals, and businesses to stay ahead in this fast-evolving field.

Subscribe to Updates

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

YouTube LinkedIn
  • Home
  • About Us
  • Advertise With Us
  • Contact Us
  • DMCA
  • Privacy Policy
  • Terms & Conditions
© 2025 advancedainews. Designed by advancedainews.

Type above and press Enter to search. Press Esc to cancel.