Close Menu
  • Home
  • AI Models
    • DeepSeek
    • xAI
    • OpenAI
    • Meta AI Llama
    • Google DeepMind
    • Amazon AWS AI
    • Microsoft AI
    • Anthropic (Claude)
    • NVIDIA AI
    • IBM WatsonX Granite 3.1
    • Adobe Sensi
    • Hugging Face
    • Alibaba Cloud (Qwen)
    • Baidu (ERNIE)
    • C3 AI
    • DataRobot
    • Mistral AI
    • Moonshot AI (Kimi)
    • Google Gemma
    • xAI
    • Stability AI
    • H20.ai
  • AI Research
    • Allen Institue for AI
    • arXiv AI
    • Berkeley AI Research
    • CMU AI
    • Google Research
    • Microsoft Research
    • Meta AI Research
    • OpenAI Research
    • Stanford HAI
    • MIT CSAIL
    • Harvard AI
  • AI Funding & Startups
    • AI Funding Database
    • CBInsights AI
    • Crunchbase AI
    • Data Robot Blog
    • TechCrunch AI
    • VentureBeat AI
    • The Information AI
    • Sifted AI
    • WIRED AI
    • Fortune AI
    • PitchBook
    • TechRepublic
    • SiliconANGLE – Big Data
    • MIT News
    • Data Robot Blog
  • Expert Insights & Videos
    • Google DeepMind
    • Lex Fridman
    • Matt Wolfe AI
    • Yannic Kilcher
    • Two Minute Papers
    • AI Explained
    • TheAIEdge
    • Matt Wolfe AI
    • The TechLead
    • Andrew Ng
    • OpenAI
  • Expert Blogs
    • François Chollet
    • Gary Marcus
    • IBM
    • Jack Clark
    • Jeremy Howard
    • Melanie Mitchell
    • Andrew Ng
    • Andrej Karpathy
    • Sebastian Ruder
    • Rachel Thomas
    • IBM
  • AI Policy & Ethics
    • ACLU AI
    • AI Now Institute
    • Center for AI Safety
    • EFF AI
    • European Commission AI
    • Partnership on AI
    • Stanford HAI Policy
    • Mozilla Foundation AI
    • Future of Life Institute
    • Center for AI Safety
    • World Economic Forum AI
  • AI Tools & Product Releases
    • AI Assistants
    • AI for Recruitment
    • AI Search
    • Coding Assistants
    • Customer Service AI
    • Image Generation
    • Video Generation
    • Writing Tools
    • AI for Recruitment
    • Voice/Audio Generation
  • Industry Applications
    • Finance AI
    • Healthcare AI
    • Legal AI
    • Manufacturing AI
    • Media & Entertainment
    • Transportation AI
    • Education AI
    • Retail AI
    • Agriculture AI
    • Energy AI
  • AI Art & Entertainment
    • AI Art News Blog
    • Artvy Blog » AI Art Blog
    • Weird Wonderful AI Art Blog
    • The Chainsaw » AI Art
    • Artvy Blog » AI Art Blog
What's Hot

Claude Free features: Voice mode and search

Google DeepMind Boss Says He’d Pay Thousands To ‘Get Rid’ Of His Email — So He’s Building An AI To Do It – Alphabet (NASDAQ:GOOG), Alphabet (NASDAQ:GOOGL)

[2502.00657] LLM Safety Alignment is Divergence Estimation in Disguise

Facebook X (Twitter) Instagram
Advanced AI News
  • Home
  • AI Models
    • Adobe Sensi
    • Aleph Alpha
    • Alibaba Cloud (Qwen)
    • Amazon AWS AI
    • Anthropic (Claude)
    • Apple Core ML
    • Baidu (ERNIE)
    • ByteDance Doubao
    • C3 AI
    • Cohere
    • DataRobot
    • DeepSeek
  • AI Research & Breakthroughs
    • Allen Institue for AI
    • arXiv AI
    • Berkeley AI Research
    • CMU AI
    • Google Research
    • Meta AI Research
    • Microsoft Research
    • OpenAI Research
    • Stanford HAI
    • MIT CSAIL
    • Harvard AI
  • AI Funding & Startups
    • AI Funding Database
    • CBInsights AI
    • Crunchbase AI
    • Data Robot Blog
    • TechCrunch AI
    • VentureBeat AI
    • The Information AI
    • Sifted AI
    • WIRED AI
    • Fortune AI
    • PitchBook
    • TechRepublic
    • SiliconANGLE – Big Data
    • MIT News
    • Data Robot Blog
  • Expert Insights & Videos
    • Google DeepMind
    • Lex Fridman
    • Meta AI Llama
    • Yannic Kilcher
    • Two Minute Papers
    • AI Explained
    • TheAIEdge
    • Matt Wolfe AI
    • The TechLead
    • Andrew Ng
    • OpenAI
  • Expert Blogs
    • François Chollet
    • Gary Marcus
    • IBM
    • Jack Clark
    • Jeremy Howard
    • Melanie Mitchell
    • Andrew Ng
    • Andrej Karpathy
    • Sebastian Ruder
    • Rachel Thomas
    • IBM
  • AI Policy & Ethics
    • ACLU AI
    • AI Now Institute
    • Center for AI Safety
    • EFF AI
    • European Commission AI
    • Partnership on AI
    • Stanford HAI Policy
    • Mozilla Foundation AI
    • Future of Life Institute
    • Center for AI Safety
    • World Economic Forum AI
  • AI Tools & Product Releases
    • AI Assistants
    • AI for Recruitment
    • AI Search
    • Coding Assistants
    • Customer Service AI
    • Image Generation
    • Video Generation
    • Writing Tools
    • AI for Recruitment
    • Voice/Audio Generation
  • Industry Applications
    • Education AI
    • Energy AI
    • Finance AI
    • Healthcare AI
    • Legal AI
    • Media & Entertainment
    • Transportation AI
    • Manufacturing AI
    • Retail AI
    • Agriculture AI
  • AI Art & Entertainment
    • AI Art News Blog
    • Artvy Blog » AI Art Blog
    • Weird Wonderful AI Art Blog
    • The Chainsaw » AI Art
    • Artvy Blog » AI Art Blog
Advanced AI News
Home » AI Safety Index Released – Future of Life Institute
Future of Life Institute

AI Safety Index Released – Future of Life Institute

Advanced AI BotBy Advanced AI BotDecember 11, 2024No Comments4 Mins Read
Share Facebook Twitter Pinterest Copy Link Telegram LinkedIn Tumblr Email
Share
Facebook Twitter LinkedIn Pinterest Email


For Immediate Release
December 11th, 2024
Media Contact: Chase Hardin, chase@futureoflife.org
+1 (623)986-0161

Major AI Companies Have ‘Significant Gaps’ in Safety Measures Say Leading AI Experts in External Safety Review

CAMPBELL, CA  – Today, the Future of Life Institute (FLI) released its 2024 AI Safety Index, in which several of the world’s foremost AI and governance experts assessed the safety standards of six prominent companies developing AI, including Anthropic, Google DeepMind, Meta, OpenAI, x.AI, and Zhipu AI. The independent panel evaluated each company in six categories: Risk Assessment, Current Harms, Safety Frameworks, Existential Safety Strategy, Governance & Accountability, and Transparency & Communication.

The review panel found that, while some companies demonstrated commendable practices in select domains, there are significant risk management disparities between them. All of the flagship models were found to be vulnerable to adversarial attacks, and despite their explicit ambitions to develop systems that rival or exceed human intelligence, firms have no adequate strategy for ensuring such systems remain beneficial and under human control.

“It’s horrifying that the very companies whose leaders predict AI could end humanity have no strategy to avert such a fate,” said panelist David Krueger, Assistant Professor at Université de Montreal and a core member of Mila.

“The findings of the AI Safety Index project suggest that although there is a lot of activity at AI companies that goes under the heading of ‘safety,’ it is not yet very effective,” said panelist Stuart Russell, a Professor of Computer Science at UC Berkeley. “In particular, none of the current activity provides any kind of quantitative guarantee of safety; nor does it seem possible to provide such guarantees given the current approach to AI via giant black boxes trained on unimaginably vast quantities of data. And it’s only going to get harder as these AI systems get bigger. In other words, it’s possible that the current technology direction can never support the necessary safety guarantees, in which case it’s really a dead end.”

Grading: Uses the US GPA system for grade boundaries: A+, A, A-, B+, […], F letter values corresponding to numerical values 4.3, 4.0, 3.7, 3.3, […], 0.

The final report can be viewed here.

“Evaluation initiatives like this Index are very important because they can provide valuable insights into the safety practices of leading AI companies. They are an essential step in holding firms accountable for their safety commitments and can help highlight emerging best practices and encourage competitors to adopt more responsible approaches,” said Professor Yoshua Bengio, Full Professor at Université de Montréal, Founder and Scientific Director of Mila – Quebec AI Institute and 2018 A.M. Turing Award co-winner.

Grades were assessed based on publicly available information as well as the companies’ responses to a survey conducted by FLI.  The review raised concerns that ongoing competitive pressures are encouraging companies to ignore or sidestep questions around the risks posed by developing this technology, resulting in significant gaps in safety measures and a serious need for improved accountability.

“We launched the Safety Index to give the public a clear picture of where these AI labs stand on safety issues,” said FLI president Max Tegmark, a professor doing AI research at MIT. “The reviewers have decades of combined experience in AI and risk assessment, so when they speak up about AI safety, we should pay close attention to what they say.”

Review panelists: 

Yoshua Bengio, Professor at Université de Montreal and Founder of Mila – Quebec Artificial Intelligence Institute. He is the recipient of the 2018 A.M. Turing Award.

Atoosa Kasirzadeh, an Assistant Professor at Carnegie Mellon University and a 2024 Schmidt Sciences AI2050 Fellow.

David Krueger, Assistant Professor at Université de Montreal and a core member of Mila and the Center for Human-compatible AI.

Tegan Maharaj, Assistant Professor at HEC Montréal and core faculty of Mila. She leads the ERRATA lab on Responsible AI.

Jessica Newman, Director of the AI Security Initiative at UC Berkeley and the Co-Director of the UC Berkeley AI Policy Hub.

Sneha Revanur, founder of youth AI advocacy organization Encode Justice and a Forbes 30 Under 30 honoree.

Stuart Russell is a Professor of Computer Science at UC Berkeley, where he leads the Center for Human-compatible AI. He co-authored the standard AI textbook used in more than 1,500 universities in 135 countries.

The Future of Life Institute is a global non-profit organization working to steer the development of transformative technologies towards benefiting life and away from extreme large-scale risks. To find out more about our mission or explore our work, visit www.futureoflife.org.



Source link

Follow on Google News Follow on Flipboard
Share. Facebook Twitter Pinterest LinkedIn Tumblr Email Copy Link
Previous ArticleOpenAI Launches Sora Turbo for Video Generation
Next Article AI Art Shop Team AI Art Shop Founder Featured on UnrealInternet’s “…
Advanced AI Bot
  • Website

Related Posts

Are we close to an intelligence explosion?

March 21, 2025

The Impact of AI in Education: Navigating the Imminent Future

February 13, 2025

A Buddhist Perspective on AI: Cultivating freedom of attention and true diversity in an AI future

January 20, 2025
Leave A Reply Cancel Reply

Latest Posts

Edition Hotels’ Latest Residences Offer Sweeping Views Of Nashville

How The ‘Dine With Dez’ Series Fosters Community For Fashion Creatives

2025 Guide to The Newest, The Coolest And The Craziest Music Festivals

‘Squid Game’ And Other K-Culture Moments

Latest Posts

Claude Free features: Voice mode and search

June 3, 2025

Google DeepMind Boss Says He’d Pay Thousands To ‘Get Rid’ Of His Email — So He’s Building An AI To Do It – Alphabet (NASDAQ:GOOG), Alphabet (NASDAQ:GOOGL)

June 3, 2025

[2502.00657] LLM Safety Alignment is Divergence Estimation in Disguise

June 3, 2025

Subscribe to News

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

Welcome to Advanced AI News—your ultimate destination for the latest advancements, insights, and breakthroughs in artificial intelligence.

At Advanced AI News, we are passionate about keeping you informed on the cutting edge of AI technology, from groundbreaking research to emerging startups, expert insights, and real-world applications. Our mission is to deliver high-quality, up-to-date, and insightful content that empowers AI enthusiasts, professionals, and businesses to stay ahead in this fast-evolving field.

Subscribe to Updates

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

YouTube LinkedIn
  • Home
  • About Us
  • Advertise With Us
  • Contact Us
  • DMCA
  • Privacy Policy
  • Terms & Conditions
© 2025 advancedainews. Designed by advancedainews.

Type above and press Enter to search. Press Esc to cancel.