Close Menu
  • Home
  • AI Models
    • DeepSeek
    • xAI
    • OpenAI
    • Meta AI Llama
    • Google DeepMind
    • Amazon AWS AI
    • Microsoft AI
    • Anthropic (Claude)
    • NVIDIA AI
    • IBM WatsonX Granite 3.1
    • Adobe Sensi
    • Hugging Face
    • Alibaba Cloud (Qwen)
    • Baidu (ERNIE)
    • C3 AI
    • DataRobot
    • Mistral AI
    • Moonshot AI (Kimi)
    • Google Gemma
    • xAI
    • Stability AI
    • H20.ai
  • AI Research
    • Allen Institue for AI
    • arXiv AI
    • Berkeley AI Research
    • CMU AI
    • Google Research
    • Microsoft Research
    • Meta AI Research
    • OpenAI Research
    • Stanford HAI
    • MIT CSAIL
    • Harvard AI
  • AI Funding & Startups
    • AI Funding Database
    • CBInsights AI
    • Crunchbase AI
    • Data Robot Blog
    • TechCrunch AI
    • VentureBeat AI
    • The Information AI
    • Sifted AI
    • WIRED AI
    • Fortune AI
    • PitchBook
    • TechRepublic
    • SiliconANGLE – Big Data
    • MIT News
    • Data Robot Blog
  • Expert Insights & Videos
    • Google DeepMind
    • Lex Fridman
    • Matt Wolfe AI
    • Yannic Kilcher
    • Two Minute Papers
    • AI Explained
    • TheAIEdge
    • Matt Wolfe AI
    • The TechLead
    • Andrew Ng
    • OpenAI
  • Expert Blogs
    • François Chollet
    • Gary Marcus
    • IBM
    • Jack Clark
    • Jeremy Howard
    • Melanie Mitchell
    • Andrew Ng
    • Andrej Karpathy
    • Sebastian Ruder
    • Rachel Thomas
    • IBM
  • AI Policy & Ethics
    • ACLU AI
    • AI Now Institute
    • Center for AI Safety
    • EFF AI
    • European Commission AI
    • Partnership on AI
    • Stanford HAI Policy
    • Mozilla Foundation AI
    • Future of Life Institute
    • Center for AI Safety
    • World Economic Forum AI
  • AI Tools & Product Releases
    • AI Assistants
    • AI for Recruitment
    • AI Search
    • Coding Assistants
    • Customer Service AI
    • Image Generation
    • Video Generation
    • Writing Tools
    • AI for Recruitment
    • Voice/Audio Generation
  • Industry Applications
    • Finance AI
    • Healthcare AI
    • Legal AI
    • Manufacturing AI
    • Media & Entertainment
    • Transportation AI
    • Education AI
    • Retail AI
    • Agriculture AI
    • Energy AI
  • AI Art & Entertainment
    • AI Art News Blog
    • Artvy Blog » AI Art Blog
    • Weird Wonderful AI Art Blog
    • The Chainsaw » AI Art
    • Artvy Blog » AI Art Blog
What's Hot

Anthropic study: Leading AI models show up to 96% blackmail rate against executives

Anthropic says most AI models, not just Claude, will resort to blackmail

Verifying Mission-Critical AI Programs | Two Minute Papers #179

Facebook X (Twitter) Instagram
Advanced AI News
  • Home
  • AI Models
    • Adobe Sensi
    • Aleph Alpha
    • Alibaba Cloud (Qwen)
    • Amazon AWS AI
    • Anthropic (Claude)
    • Apple Core ML
    • Baidu (ERNIE)
    • ByteDance Doubao
    • C3 AI
    • Cohere
    • DataRobot
    • DeepSeek
  • AI Research & Breakthroughs
    • Allen Institue for AI
    • arXiv AI
    • Berkeley AI Research
    • CMU AI
    • Google Research
    • Meta AI Research
    • Microsoft Research
    • OpenAI Research
    • Stanford HAI
    • MIT CSAIL
    • Harvard AI
  • AI Funding & Startups
    • AI Funding Database
    • CBInsights AI
    • Crunchbase AI
    • Data Robot Blog
    • TechCrunch AI
    • VentureBeat AI
    • The Information AI
    • Sifted AI
    • WIRED AI
    • Fortune AI
    • PitchBook
    • TechRepublic
    • SiliconANGLE – Big Data
    • MIT News
    • Data Robot Blog
  • Expert Insights & Videos
    • Google DeepMind
    • Lex Fridman
    • Meta AI Llama
    • Yannic Kilcher
    • Two Minute Papers
    • AI Explained
    • TheAIEdge
    • Matt Wolfe AI
    • The TechLead
    • Andrew Ng
    • OpenAI
  • Expert Blogs
    • François Chollet
    • Gary Marcus
    • IBM
    • Jack Clark
    • Jeremy Howard
    • Melanie Mitchell
    • Andrew Ng
    • Andrej Karpathy
    • Sebastian Ruder
    • Rachel Thomas
    • IBM
  • AI Policy & Ethics
    • ACLU AI
    • AI Now Institute
    • Center for AI Safety
    • EFF AI
    • European Commission AI
    • Partnership on AI
    • Stanford HAI Policy
    • Mozilla Foundation AI
    • Future of Life Institute
    • Center for AI Safety
    • World Economic Forum AI
  • AI Tools & Product Releases
    • AI Assistants
    • AI for Recruitment
    • AI Search
    • Coding Assistants
    • Customer Service AI
    • Image Generation
    • Video Generation
    • Writing Tools
    • AI for Recruitment
    • Voice/Audio Generation
  • Industry Applications
    • Education AI
    • Energy AI
    • Finance AI
    • Healthcare AI
    • Legal AI
    • Media & Entertainment
    • Transportation AI
    • Manufacturing AI
    • Retail AI
    • Agriculture AI
  • AI Art & Entertainment
    • AI Art News Blog
    • Artvy Blog » AI Art Blog
    • Weird Wonderful AI Art Blog
    • The Chainsaw » AI Art
    • Artvy Blog » AI Art Blog
Facebook X (Twitter) Instagram
Advanced AI News
Home » 2025 is NOT the Year of AI Agents
Andrej Karpathy

2025 is NOT the Year of AI Agents

Advanced AI EditorBy Advanced AI EditorJune 20, 2025No Comments6 Mins Read
Share Facebook Twitter Pinterest Copy Link Telegram LinkedIn Tumblr Email
Share
Facebook Twitter LinkedIn Pinterest Email


Andrej Karpathy is back, this time explaining how LLMs are rewriting software.

At YC AI Startup School, the former head of AI at Tesla gave a talk titled “Software Is Changing (Again),” during which he discussed with students and developers how the concepts of code, computation, and programming are being rethought at a fundamental level.

He defined three kinds of Software. The first, Software 1.0, consists of traditional programming, in which humans write explicit instructions for computers to execute. 

Karpathy said that in Software 2.0, instead of writing code manually (as in Software 1.0), developers work with neural networks, specifically by tuning datasets and using an optimiser (like gradient descent) to learn the weights or parameters of the neural network automatically.

While the third is Software 3.0, where LLMs have made neural networks programmable in a new way. Instead of writing traditional code, users now write prompts in natural language like English, which effectively serve as programs that instruct the model. 

“I think a quite fundamental change is that neural networks have become programmable with large language models. I see this as something new and unique, it’s a new kind of computer. In my mind, it’s worth giving it a new designation, Software 3.0,” Karpathy said. He also discussed the rise of vibe coding in recent months and how its growing popularity among kids gives him hope for an exciting future.

Karpathy shared a few apps he built while vibe-coding, like MenuGen (menugen.app), which turns menu text into visuals to help users make sense of it.

Human in the Loop

While LLMs may eventually be able to browse, click, and navigate the web more like humans do, Karpathy believes it’s still valuable to meet them halfway. He said humans should generate content in a format that can be easily understood by LLMs. Karpathy gave the example of Gitingest, which turns any Git repository into a simple text digest of its codebase. This is useful for feeding a codebase into any LLM.

He referred to the next wave of software as partial autonomy apps built on LLMs, where humans continue to play a key role in oversight and control rather than handing over full autonomy. “We have to keep the AI on the leash. A lot of people are getting way overexcited with AI agents,” said Karpathy. 

“When I see things like ‘2025 is the year of agents,’ I get very concerned… this is the decade of agents,” he added. He urged the developers to build augmented systems, like Iron Man suits, not Iron Man robots that accelerate human productivity without removing human oversight, as LLMs are still fallible.

Referencing his work on Tesla’s Autopilot, Karpathy pointed out that despite years of development, full autonomy has not yet been achieved, even in vehicles that appear driverless. “There’s still a lot of teleoperation. We haven’t declared success.”

Karpathy referred to LLMs as “people spirits”—superhuman in some ways (like memory or general knowledge) but deeply flawed in others (like hallucinations, logical inconsistencies, or context retention). He said they simulate intelligence but don’t develop knowledge over time like a human would. Instead, they rely on fixed weights and short-term context windows, which he compares to working memory.

He cited tools like Perplexity AI and Cursor as examples of intelligent orchestration of multiple LLM components behind the scenes and mechanisms for human-in-the-loop verification. Crucially, these apps also offered what Karpathy called an “autonomy slider,” allowing users to control how much freedom the AI had depending on the complexity and risk of the task.

Build for the Agents 

Karpathy said we need a new interface built specifically for agents. He explained that a new kind of software user has arrived—neither a person clicking through a GUI nor a backend system making API calls.

Cool demo of a GUI for LLMs! Obviously it has a bit silly feel of a “horseless carriage” in that it exactly replicates conventional UI in the new paradigm, but the high level idea is to generate a completely ephemeral UI on demand depending on the specific task at hand. https://t.co/Xgh1iwDmJl

— Andrej Karpathy (@karpathy) June 19, 2025

Instead, LLMs represent something in between. Karpathy described them as the third major consumer and manipulator of digital information, urging developers to start designing with them in mind. 

Traditionally, software has served two users. Humans through graphical interfaces and computers through APIs. But LLMs occupy a new space. “There’s a new category of consumer,” he said. “Agents. They’re computers, but they are humanlike. People’s spirits on the internet.”

Karpathy said that whenever he uses ChatGPT, he feels like he is talking to an operating system through the terminal. He believes that it should have a new GUI, other than just a text bubble. 

LLMs Resemble Fabs and Utilities

He further compared the development of LLMs to semiconductor manufacturing. Building advanced LLMs, he said, involves massive capital investment, proprietary methods, and tightly integrated R&D, similar to running a chip fabrication facility. 

“The capex required for building LLMs is actually quite large,” he said. “We have deep tech trees, R&D secrets, centralised in LLM labs.”

Beyond hardware analogies, Karpathy’s central argument is that LLMs are evolving into full-fledged operating systems. These models coordinate memory, computation, and interaction much like a traditional OS. “The LLM is a new kind of computer—it’s like the CPU. Context windows are like memory. And the LLM orchestrates memory and compute.”

He pointed to applications like Cursor that can run on any major foundation model like GPT-4, Claude, and Gemini as examples of this platform-agnostic future. “You can take an LLM app like Cursor and run it on GPT or Claude, or Gemini. That’s kind of like downloading an app and running it on Windows, Linux, or Mac.”

We’re Back in the 1960s of Computing

At present, LLMs remain centralised and expensive to run, which Karpathy compared to the mainframe era of the 1960s. Instead of personal computers, we’re using interfaces like ChatGPT that tap into vast cloud-based models. “LLM compute is still very expensive, so they’re centralised in the cloud and we are all just thin clients interacting with it.”

He noted early signs of a shift. Some developers are already experimenting with running smaller models locally on consumer hardware like Mac Minis, but a true personal computing revolution for LLMs is still far off.

Karpathy likened LLMs to electricity: centralised, metered, and essential. Labs like OpenAI and Anthropic invest heavily in training their models, then serve intelligence over APIs, much like utilities deliver power. 

When these services go offline, the impact is immediate. “It’s like an intelligence brownout. The planet just gets dumber for a while.”

But unlike electricity, LLMs are not bound by physical laws. They are shaped by data, architecture, and training methods. This flexibility changes how we build, share, and improve them, turning LLMs into more than just a utility. They’re becoming a programmable layer of intelligence for the internet.



Source link

Follow on Google News Follow on Flipboard
Share. Facebook Twitter Pinterest LinkedIn Tumblr Email Copy Link
Previous ArticleAlibaba’s High-Speed Comeback: The AI-Commerce Cocktail Lifting ETF Hopes Alibaba’s High-Speed Comeback: The AI-Commerce Cocktail Lifting ETF Hopes – Global X Artificial Intelligence & Technology ETF (NASDAQ:AIQ), Amazon.com (NASDAQ:AMZN)
Next Article How Huawei’s Ascend AI chips outperform Nvidia processors in running DeepSeek’s R1 model
Advanced AI Editor
  • Website

Related Posts

OpenAI’s Cofounder Andrej Karpathy: ‘Keep AI on the Leash’

June 20, 2025

Andrej Karpathy Shares Insights on AI-Driven Emotional Recognition Technology in 2025 | AI News Detail

June 17, 2025

Anthropic’s Claude plays ‘for peace over victory’ in a game of Diplomacy against other AI

June 9, 2025
Leave A Reply Cancel Reply

Latest Posts

Melissa Errico To Sing Stephen Sondheim Classics In NY And London

Baghdad Group For Modern Art On View In America For First Time

Climate Activist Hurls Paint at Picasso Painting at Montreal Museum

Curtis Yarvin Details Proposed Titian-Centric ‘Art Hos’ US Pavilion

Latest Posts

Anthropic study: Leading AI models show up to 96% blackmail rate against executives

June 20, 2025

Anthropic says most AI models, not just Claude, will resort to blackmail

June 20, 2025

Verifying Mission-Critical AI Programs | Two Minute Papers #179

June 20, 2025

Subscribe to News

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

Welcome to Advanced AI News—your ultimate destination for the latest advancements, insights, and breakthroughs in artificial intelligence.

At Advanced AI News, we are passionate about keeping you informed on the cutting edge of AI technology, from groundbreaking research to emerging startups, expert insights, and real-world applications. Our mission is to deliver high-quality, up-to-date, and insightful content that empowers AI enthusiasts, professionals, and businesses to stay ahead in this fast-evolving field.

Subscribe to Updates

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

YouTube LinkedIn
  • Home
  • About Us
  • Advertise With Us
  • Contact Us
  • DMCA
  • Privacy Policy
  • Terms & Conditions
© 2025 advancedainews. Designed by advancedainews.

Type above and press Enter to search. Press Esc to cancel.