Close Menu
  • Home
  • AI Models
    • DeepSeek
    • xAI
    • OpenAI
    • Meta AI Llama
    • Google DeepMind
    • Amazon AWS AI
    • Microsoft AI
    • Anthropic (Claude)
    • NVIDIA AI
    • IBM WatsonX Granite 3.1
    • Adobe Sensi
    • Hugging Face
    • Alibaba Cloud (Qwen)
    • Baidu (ERNIE)
    • C3 AI
    • DataRobot
    • Mistral AI
    • Moonshot AI (Kimi)
    • Google Gemma
    • xAI
    • Stability AI
    • H20.ai
  • AI Research
    • Allen Institue for AI
    • arXiv AI
    • Berkeley AI Research
    • CMU AI
    • Google Research
    • Microsoft Research
    • Meta AI Research
    • OpenAI Research
    • Stanford HAI
    • MIT CSAIL
    • Harvard AI
  • AI Funding & Startups
    • AI Funding Database
    • CBInsights AI
    • Crunchbase AI
    • Data Robot Blog
    • TechCrunch AI
    • VentureBeat AI
    • The Information AI
    • Sifted AI
    • WIRED AI
    • Fortune AI
    • PitchBook
    • TechRepublic
    • SiliconANGLE – Big Data
    • MIT News
    • Data Robot Blog
  • Expert Insights & Videos
    • Google DeepMind
    • Lex Fridman
    • Matt Wolfe AI
    • Yannic Kilcher
    • Two Minute Papers
    • AI Explained
    • TheAIEdge
    • Matt Wolfe AI
    • The TechLead
    • Andrew Ng
    • OpenAI
  • Expert Blogs
    • François Chollet
    • Gary Marcus
    • IBM
    • Jack Clark
    • Jeremy Howard
    • Melanie Mitchell
    • Andrew Ng
    • Andrej Karpathy
    • Sebastian Ruder
    • Rachel Thomas
    • IBM
  • AI Policy & Ethics
    • ACLU AI
    • AI Now Institute
    • Center for AI Safety
    • EFF AI
    • European Commission AI
    • Partnership on AI
    • Stanford HAI Policy
    • Mozilla Foundation AI
    • Future of Life Institute
    • Center for AI Safety
    • World Economic Forum AI
  • AI Tools & Product Releases
    • AI Assistants
    • AI for Recruitment
    • AI Search
    • Coding Assistants
    • Customer Service AI
    • Image Generation
    • Video Generation
    • Writing Tools
    • AI for Recruitment
    • Voice/Audio Generation
  • Industry Applications
    • Finance AI
    • Healthcare AI
    • Legal AI
    • Manufacturing AI
    • Media & Entertainment
    • Transportation AI
    • Education AI
    • Retail AI
    • Agriculture AI
    • Energy AI
  • AI Art & Entertainment
    • AI Art News Blog
    • Artvy Blog » AI Art Blog
    • Weird Wonderful AI Art Blog
    • The Chainsaw » AI Art
    • Artvy Blog » AI Art Blog
What's Hot

How to Use Perplexity AI for Business: Simple Tricks

Paper page – DrafterBench: Benchmarking Large Language Models for Tasks Automation in Civil Engineering

Mistral AI brings deep research to Le Chat, alongside image editing, voice mode and more

Facebook X (Twitter) Instagram
Advanced AI News
  • Home
  • AI Models
    • OpenAI (GPT-4 / GPT-4o)
    • Anthropic (Claude 3)
    • Google DeepMind (Gemini)
    • Meta (LLaMA)
    • Cohere (Command R)
    • Amazon (Titan)
    • IBM (Watsonx)
    • Inflection AI (Pi)
  • AI Research
    • Allen Institue for AI
    • arXiv AI
    • Berkeley AI Research
    • CMU AI
    • Google Research
    • Meta AI Research
    • Microsoft Research
    • OpenAI Research
    • Stanford HAI
    • MIT CSAIL
    • Harvard AI
  • AI Funding
    • AI Funding Database
    • CBInsights AI
    • Crunchbase AI
    • Data Robot Blog
    • TechCrunch AI
    • VentureBeat AI
    • The Information AI
    • Sifted AI
    • WIRED AI
    • Fortune AI
    • PitchBook
    • TechRepublic
    • SiliconANGLE – Big Data
    • MIT News
    • Data Robot Blog
  • AI Experts
    • Google DeepMind
    • Lex Fridman
    • Meta AI Llama
    • Yannic Kilcher
    • Two Minute Papers
    • AI Explained
    • TheAIEdge
    • The TechLead
    • Matt Wolfe AI
    • Andrew Ng
    • OpenAI
    • Expert Blogs
      • François Chollet
      • Gary Marcus
      • IBM
      • Jack Clark
      • Jeremy Howard
      • Melanie Mitchell
      • Andrew Ng
      • Andrej Karpathy
      • Sebastian Ruder
      • Rachel Thomas
      • IBM
  • AI Tools
    • AI Assistants
    • AI for Recruitment
    • AI Search
    • Coding Assistants
    • Customer Service AI
  • AI Policy
    • ACLU AI
    • AI Now Institute
    • Center for AI Safety
  • Industry AI
    • Finance AI
    • Healthcare AI
    • Education AI
    • Energy AI
    • Legal AI
LinkedIn Instagram YouTube Threads X (Twitter)
Advanced AI News
Center for AI Safety

Study: AI-Powered Research Prowess Now Outstrips Human Experts, Raising Bioweapon Risks

By Advanced AI EditorApril 27, 2025No Comments5 Mins Read
Share Facebook Twitter Pinterest Copy Link Telegram LinkedIn Tumblr Email
Share
Facebook Twitter LinkedIn Pinterest Email


Researchers from the Center for AI Safety (CAIS), MIT’s Media Lab, the Brazilian university UFABC, and the pandemic prevention non-profit SecureBio have found that leading artificial intelligence models can outperform experienced, PhD-level virologists in troubleshooting complex laboratory procedures.

The findings, detailed in a new study introducing the Virology Capabilities Test (VCT), demonstrate AI’s proficiency in specialized scientific tasks but also highlight serious dual-use concerns, suggesting these tools could lower the barrier for creating dangerous biological agents.

The VCT benchmark, consisting of 322 questions and detailed further in its research paper, was designed specifically to measure an AI’s ability to assist with intricate ‘wet lab’ virology protocols, assessing fundamental, visual, and tacit understanding – the kind of practical know-how often gained through hands-on lab experience.

The results showed OpenAI’s o3 model achieved 43.8% accuracy, substantially exceeding the 22.1% average scored by specialized human virologists answering questions within their fields. Google’s Gemini 2.5 Pro also performed strongly, scoring 37.6%. According to the VCT analysis, o3’s performance surpassed 94% of the human experts on tailored question subsets.

AI Virologist Chatbots Pose Dual-Use Dilemma

This emergent AI capability – providing expert-level guidance for sensitive lab work – presents a clear dual-use scenario: useful for accelerating legitimate research but potentially dangerous if misused. Seth Donoughe, a SecureBio research scientist and study co-author, conveyed his apprehension to TIME, stating the findings made him “little nervous.”

He elaborated on the historical context: “Throughout history, there are a fair number of cases where someone attempted to make a bioweapon—and one of the major reasons why they didn’t succeed is because they didn’t have access to the right level of expertise… So it seems worthwhile to be cautious about how these capabilities are being distributed.”

Reflecting this, the VCT researchers propose that this AI skill warrants inclusion within governance frameworks designed for dual-use life science technologies.

The VCT findings spurred immediate calls for action from safety advocates. Dan Hendrycks, director of the Center for AI Safety, stressed the need for immediate action, urging AI companies to implement robust safeguards within six months, calling inaction “reckless.”

He advocated for tiered or gated access controls as a potential mitigation strategy. “We want to give the people who have a legitimate use for asking how to manipulate deadly viruses—like a researcher at the MIT biology department—the ability to do so,” Hendrycks explained to TIME. “But random people who made an account a second ago don’t get those capabilities.”

Industry Responses and Calls for Oversight

Having been briefed on the VCT results months ago, AI developers have reacted differently. xAI, Elon Musk’s company, in February, published a risk management framework acknowledging the paper and mentioning potential virology safeguards for its Grok model, such as training it to decline harmful requests.

OpenAI stated it “deployed new system-level mitigations for biological risks” for its recently released o3 and o4-mini models, including specific measures like “blocking harmful outputs.”

This measure reportedly resulted from a “thousand-hour red-teaming campaign in which 98.7% of unsafe bio-related conversations were successfully flagged and blocked.” Red-teaming is a common security practice involving simulated attacks to find vulnerabilities. Anthropic, another leading AI lab, acknowledged the VCT results in its system documentation but offered no specific mitigation plans, while Google declined to comment on the matter to TIME.

However, some experts believe self-policing by the industry isn’t sufficient. Tom Inglesby from the Johns Hopkins Center for Health Security advocated for governmental policy and regulation. “The current situation is that the companies that are most virtuous are taking time and money to do this work, which is good for all of us, but other companies don’t have to do it,” he told TIME, adding, “That doesn’t make sense.” Inglesby proposed mandatory evaluations for new large language models before their release “to make sure it will not produce pandemic-level outcomes.”

AI’s Expanding Footprint in Scientific Research

The VCT results are not an isolated incident but rather a stark data point within a broader landscape where AI is rapidly integrating into specialized scientific fields. OpenAI, creator of the top-performing o3 model, was already known to be exploring biological applications; Winbuzzer reported in January on its collaboration with Retro Biosciences using a model named GPT-4b Micro to optimize proteins involved in stem cell creation.

Similarly, Google DeepMind has been highly active. Besides the Gemini model family, its widely used AlphaFold program predicts protein structures, while an “AI Co-Scientist” project, detailed in February, aims to generate novel scientific hypotheses, sometimes mirroring unpublished human research.

Microsoft entered the fray in February with BioEmu-1, a model focused on predicting the dynamic movement of proteins, complementing AlphaFold’s static predictions. These tools, focusing on protein engineering, hypothesis generation, and molecular simulation, illustrate AI’s expanding role, moving beyond data analysis toward complex scientific reasoning and procedural assistance – amplifying both the potential scientific gains and the safety challenges highlighted by the VCT.



Source link

Follow on Google News Follow on Flipboard
Share. Facebook Twitter Pinterest LinkedIn Tumblr Email Copy Link
Previous ArticleAlibaba’s (BABA) Quark Crushes AI Rivals to Become Top App in China
Next Article Subdivisions.com Builds the Foundation for AI-Ready Real Estate Search and Hyperlocal Discovery
Advanced AI Editor
  • Website

Related Posts

California’s Draft AI Law Would Protect More than Just People

July 16, 2025

Northwest Seaport Alliance launches zero-emission drayage truck incentive program

July 12, 2025

Dan Hendrycks: The 100 Most Influential People in AI 2023

July 8, 2025
Leave A Reply

Latest Posts

Chanel Will Return to New York City with Métiers d’Art Collection

Rashid Johnson Painting Spotted in Trump Official’s Home

Christie’s Reports $2.1 B. Sales Total for H1 2024

Morning Links for July 16, 2025

Latest Posts

How to Use Perplexity AI for Business: Simple Tricks

July 17, 2025

Paper page – DrafterBench: Benchmarking Large Language Models for Tasks Automation in Civil Engineering

July 17, 2025

Mistral AI brings deep research to Le Chat, alongside image editing, voice mode and more

July 17, 2025

Subscribe to News

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

Recent Posts

  • How to Use Perplexity AI for Business: Simple Tricks
  • Paper page – DrafterBench: Benchmarking Large Language Models for Tasks Automation in Civil Engineering
  • Mistral AI brings deep research to Le Chat, alongside image editing, voice mode and more
  • Anthropic Hit With Class Action Certification in High-Stakes AI Copyright Battle
  • An OpenAI Employee’s Farewell Letter Offers Rare Glimpse Into the Company

Recent Comments

  1. melhor código de indicac~ao binance on Google DeepMind develops AlphaEvolve AI agent optimized for coding and math
  2. aviator official website on Former Tesla AI czar Andrej Karpathy coins ‘vibe coding’: Here’s what it means
  3. BitStarz on Former Tesla AI czar Andrej Karpathy coins ‘vibe coding’: Here’s what it means
  4. bit starz best game on Former Tesla AI czar Andrej Karpathy coins ‘vibe coding’: Here’s what it means
  5. binance referral on Qwen 2.5 Coder and Qwen 3 Lead in Open Source LLM Over DeepSeek and Meta

Welcome to Advanced AI News—your ultimate destination for the latest advancements, insights, and breakthroughs in artificial intelligence.

At Advanced AI News, we are passionate about keeping you informed on the cutting edge of AI technology, from groundbreaking research to emerging startups, expert insights, and real-world applications. Our mission is to deliver high-quality, up-to-date, and insightful content that empowers AI enthusiasts, professionals, and businesses to stay ahead in this fast-evolving field.

Subscribe to Updates

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

LinkedIn Instagram YouTube Threads X (Twitter)
  • Home
  • About Us
  • Advertise With Us
  • Contact Us
  • DMCA
  • Privacy Policy
  • Terms & Conditions
© 2025 advancedainews. Designed by advancedainews.

Type above and press Enter to search. Press Esc to cancel.