Close Menu
  • Home
  • AI Models
    • DeepSeek
    • xAI
    • OpenAI
    • Meta AI Llama
    • Google DeepMind
    • Amazon AWS AI
    • Microsoft AI
    • Anthropic (Claude)
    • NVIDIA AI
    • IBM WatsonX Granite 3.1
    • Adobe Sensi
    • Hugging Face
    • Alibaba Cloud (Qwen)
    • Baidu (ERNIE)
    • C3 AI
    • DataRobot
    • Mistral AI
    • Moonshot AI (Kimi)
    • Google Gemma
    • xAI
    • Stability AI
    • H20.ai
  • AI Research
    • Allen Institue for AI
    • arXiv AI
    • Berkeley AI Research
    • CMU AI
    • Google Research
    • Microsoft Research
    • Meta AI Research
    • OpenAI Research
    • Stanford HAI
    • MIT CSAIL
    • Harvard AI
  • AI Funding & Startups
    • AI Funding Database
    • CBInsights AI
    • Crunchbase AI
    • Data Robot Blog
    • TechCrunch AI
    • VentureBeat AI
    • The Information AI
    • Sifted AI
    • WIRED AI
    • Fortune AI
    • PitchBook
    • TechRepublic
    • SiliconANGLE – Big Data
    • MIT News
    • Data Robot Blog
  • Expert Insights & Videos
    • Google DeepMind
    • Lex Fridman
    • Matt Wolfe AI
    • Yannic Kilcher
    • Two Minute Papers
    • AI Explained
    • TheAIEdge
    • Matt Wolfe AI
    • The TechLead
    • Andrew Ng
    • OpenAI
  • Expert Blogs
    • François Chollet
    • Gary Marcus
    • IBM
    • Jack Clark
    • Jeremy Howard
    • Melanie Mitchell
    • Andrew Ng
    • Andrej Karpathy
    • Sebastian Ruder
    • Rachel Thomas
    • IBM
  • AI Policy & Ethics
    • ACLU AI
    • AI Now Institute
    • Center for AI Safety
    • EFF AI
    • European Commission AI
    • Partnership on AI
    • Stanford HAI Policy
    • Mozilla Foundation AI
    • Future of Life Institute
    • Center for AI Safety
    • World Economic Forum AI
  • AI Tools & Product Releases
    • AI Assistants
    • AI for Recruitment
    • AI Search
    • Coding Assistants
    • Customer Service AI
    • Image Generation
    • Video Generation
    • Writing Tools
    • AI for Recruitment
    • Voice/Audio Generation
  • Industry Applications
    • Finance AI
    • Healthcare AI
    • Legal AI
    • Manufacturing AI
    • Media & Entertainment
    • Transportation AI
    • Education AI
    • Retail AI
    • Agriculture AI
    • Energy AI
  • AI Art & Entertainment
    • AI Art News Blog
    • Artvy Blog » AI Art Blog
    • Weird Wonderful AI Art Blog
    • The Chainsaw » AI Art
    • Artvy Blog » AI Art Blog
What's Hot

EU Commission: “AI Gigafactories” to strengthen Europe as a business location

United States, China, and United Kingdom Lead the Global AI Ranking According to Stanford HAI’s Global AI Vibrancy Tool

Foundation AI: Cisco launches AI model for integration in security applications

Facebook X (Twitter) Instagram
Advanced AI News
  • Home
  • AI Models
    • Adobe Sensi
    • Aleph Alpha
    • Alibaba Cloud (Qwen)
    • Amazon AWS AI
    • Anthropic (Claude)
    • Apple Core ML
    • Baidu (ERNIE)
    • ByteDance Doubao
    • C3 AI
    • Cohere
    • DataRobot
    • DeepSeek
  • AI Research & Breakthroughs
    • Allen Institue for AI
    • arXiv AI
    • Berkeley AI Research
    • CMU AI
    • Google Research
    • Meta AI Research
    • Microsoft Research
    • OpenAI Research
    • Stanford HAI
    • MIT CSAIL
    • Harvard AI
  • AI Funding & Startups
    • AI Funding Database
    • CBInsights AI
    • Crunchbase AI
    • Data Robot Blog
    • TechCrunch AI
    • VentureBeat AI
    • The Information AI
    • Sifted AI
    • WIRED AI
    • Fortune AI
    • PitchBook
    • TechRepublic
    • SiliconANGLE – Big Data
    • MIT News
    • Data Robot Blog
  • Expert Insights & Videos
    • Google DeepMind
    • Lex Fridman
    • Meta AI Llama
    • Yannic Kilcher
    • Two Minute Papers
    • AI Explained
    • TheAIEdge
    • Matt Wolfe AI
    • The TechLead
    • Andrew Ng
    • OpenAI
  • Expert Blogs
    • François Chollet
    • Gary Marcus
    • IBM
    • Jack Clark
    • Jeremy Howard
    • Melanie Mitchell
    • Andrew Ng
    • Andrej Karpathy
    • Sebastian Ruder
    • Rachel Thomas
    • IBM
  • AI Policy & Ethics
    • ACLU AI
    • AI Now Institute
    • Center for AI Safety
    • EFF AI
    • European Commission AI
    • Partnership on AI
    • Stanford HAI Policy
    • Mozilla Foundation AI
    • Future of Life Institute
    • Center for AI Safety
    • World Economic Forum AI
  • AI Tools & Product Releases
    • AI Assistants
    • AI for Recruitment
    • AI Search
    • Coding Assistants
    • Customer Service AI
    • Image Generation
    • Video Generation
    • Writing Tools
    • AI for Recruitment
    • Voice/Audio Generation
  • Industry Applications
    • Education AI
    • Energy AI
    • Finance AI
    • Healthcare AI
    • Legal AI
    • Media & Entertainment
    • Transportation AI
    • Manufacturing AI
    • Retail AI
    • Agriculture AI
  • AI Art & Entertainment
    • AI Art News Blog
    • Artvy Blog » AI Art Blog
    • Weird Wonderful AI Art Blog
    • The Chainsaw » AI Art
    • Artvy Blog » AI Art Blog
Facebook X (Twitter) Instagram
Advanced AI News
Home » 10 strategies OpenAI uses to create powerful AI agents – that you should use too
OpenAI

10 strategies OpenAI uses to create powerful AI agents – that you should use too

Advanced AI EditorBy Advanced AI EditorJune 20, 2025No Comments11 Mins Read
Share Facebook Twitter Pinterest Copy Link Telegram LinkedIn Tumblr Email
Share
Facebook Twitter LinkedIn Pinterest Email


ai concept

Just_Super/Getty Images

AI integration is moving at an astonishing pace. Just a few months ago, we were coming to terms with the idea of AI agents, or what the buzzword mavens call “agentic AI.” Now, we’re starting to look at issues of practical deployment.

If you’re not fully up to speed on agents, that’s okay. Few people are. OpenAI defines agents as “Systems that independently accomplish tasks on your behalf,” with an emphasis on “independently.” ZDNET has a full guide on the topic, which is essential reading. 

Also: 8 ways to write better ChatGPT prompts – and get the results you want faster

OpenAI recently released a 34-page PDF entitled “A practical guide to building agents,” which includes notes and guidelines from the company’s experience deploying agents for its clients. After reading the guide, I’ve distilled OpenAI’s recommendations into 10 best practices, detailed below.

(Disclosure: Ziff Davis, ZDNET’s parent company, filed an April 2025 lawsuit against OpenAI, alleging it infringed Ziff Davis copyrights in training and operating its AI systems.)

1. Prioritize stubborn workflows

AI is so dominating the conversation that we sometimes forget to ask an obvious question: when should it be used? We ask that for all our tools. Even if a tool is very general purpose (like a computer or an AI), we still want to know what it’s best at and how it can help.

Also: Is ChatGPT Plus really worth $20 when the free version offers so many premium features?

This is particularly true of agents, which are designed to carry out sets of tasks. When I read OpenAI’s guide, something clicked in terms of my understanding. It was this statement: “As you evaluate where agents can add value, prioritize workflows that have previously resisted automation, especially where traditional methods encounter friction.”

This makes total sense. If you, like me, have been automating workflows for years, you’ve probably run into some that just seem to fight back. They’re either too undefined, too variable, or require too many judgment calls to efficiently automate, even if such automation could save hours or days of time.

Don’t throw agents at every problem you have. If a traditional “deterministic” or algorithmic solution would work, use it. Only move into AI when the old school techniques aren’t up to the task.

2. Understand models, tools, and instructions

Agents consist of three main elements: models, tools, and instructions. The model is the AI itself, the large language model that does the reasoning. Tools are the mechanisms (usually APIs) that agents use to get things done. And instructions are the prompts that are fed to the AIs that instruct the AI agents about what you want and how to get the job done.

Let’s say you run a footwear e-commerce site that lets users upload pictures of their shoes. You want an agent to accept relevant shoe pictures, but filter out all the images that aren’t relevant. The tools are an image reader API and the database API. The model is some LLM that’s used to make a determination.

And the instructions specify that you want footwear, and that you only want to allow footwear that matches the model of shoe being shown. You might also add some guardrails, telling the AI to only accept images from the knee down, only accept images that meet certain standards and practices, only accept images with a certain level of focus and contrast, only accept images without any text, and so on.

Also: Can you build a billion-dollar business with only AI agents (yet)? This author thinks so

By combining models, tools, and instructions, you can automate a task that might have otherwise been difficult to automate. It’s also a task that might have been difficult to assign to human workers because attention spans waver, and human inspection might be slower and less accurate.

3. Build smart, then dumb down

When you see AI models referenced, they often have fairly cryptic names like gpt-4o, gpt-4o mini, gpt-4, gpt-4 turbo, gpt-4.1, gpt-4.1 mini, gpt-4.1 nano, gpt-4.5, o1, o1-mini, o1-pro, o3, o3-mini, o3-mini-high, o3-pro, o4-mini, and o4-mini-high.

Some models are newer than others, and some are less resource-intensive. It turns out that the smarter models tend to require more resources and time. That means they cost more to operate. Models that cost more to operate cost more to rent.

Also: 60% of AI agents work in IT departments – here’s what they do every day

You might think, therefore, that it makes sense to use the least expensive model when you initially field your agents. But OpenAI suggests otherwise. It recommends you assign each task the smartest, most capable model, just to get the task working properly and establish a performance baseline.

Once the task is working, then consider replacing the model with a less capable model. See if the performance stays the same or degrades. I don’t think OpenAI is making this recommendation just to convince you to spend more on AI fees. I think it genuinely wants to make sure your AI solution works before tinkering with cost reduction. That’s the same process that many manufactured products take. Build it, then cost reduce it until you’ve reached an optimal solution.

4. Max out single agents

One of the more intriguing aspects of agents is that you can build teams of agents, much like you would build human teams.

Also: Crawl, then walk, before you run with AI agents, experts recommend

Back when I built the AI Newsroom in 2011, I implemented a fleet of agents, each of which scanned the news for one topic. Then I implemented an editor agent, which accepted the news submissions and validated them as relevant. Another agent grouped related news items together. And yet another agent prepared them for publication on my experimental websites.

Back then, we didn’t have the generative AI tools we have now, so each agent took an entire server, resulting in a couple of racks, to get the job done.

Building the agents was one thing. Getting them all to work together was incredibly difficult. That’s where OpenAI’s recommendation to maximize a single agent’s capabilities first makes so much sense.

Build out slowly. Get one agent to do as much as possible. As soon as you start adding agents, you multiply your complexity.

5. Use prompt templates

One technique OpenAI recommends that can help you max out a single agent is to use prompt templates. These are basically full AI prompts implemented as fill-in-the-blank templates.

This way, rather than switching to a new agent, you just modify the prompt by substituting variables. This technique can prove very helpful not only to reduce agent sprawl but also to make it easier to debug agent behavior.

6. More tools, more agents

OpenAI recommends that, as you add tools, you also add agents. I kind of think of this like building a house. There are different trades involved, including plumbers, electricians, carpenters, roofers, landscapers, and so on. Each of the trades knows their specific skills.

Also: How AI coding agents could destroy open source software

If you’re building up a team of agents, think of each agent as knowing a trade. Assign each agent the tools related to its particular task. As you add new and potentially unrelated tasks, add agents.

7. Tool similarity matters

Apparently, AIs also struggle when there are multiple tools that overlap or have similar functionality. The AIs have difficulty distinguishing when to use which tools.

So, if your tools are very different, a single agent might be able to handle them because it can easily determine what gets assigned to what.

But if you start to assign very similar tools, the agent may skip between the tools. OpenAI recommends, “Use multiple agents if improving tool clarity by providing descriptive names, clear parameters, and detailed descriptions doesn’t improve performance.”

8. Split when agents struggle

This next recommendation is so human, so managerial, it’s almost surreal. Yet it makes sense.

OpenAI says, “When your agents fail to follow complicated instructions or consistently select incorrect tools, you may need to further divide your system and introduce more distinct agents.”

In other words, if your agents become overwhelmed, get them some help. You can tell if they’re overwhelmed if they screw up by not following directions or using the wrong tools.

Also: How to turn ChatGPT into your AI coding power tool – and double your output

If you have an employee or a team who is not following directions or consistently using resources incorrectly, you’d initiate some mediation. Sometimes, it’s time to replace that team member. Other times, it’s simply the fact that your team members have way too much on their plates and need help.

In the case of agents, either split the work among more agents or simplify the process.

9. Build guardrails as layered defense

On a mountain road, guardrails are designed to keep cars that lose control from flying off the edge of a cliff. In AI, guardrails are designed to keep agents that lose control from doing dangerous things.

OpenAI defines seven types of guardrails.

Relevance classifier: Flags off-topic user queries to keep the agent within intended scopeSafety classifier: Detects unsafe inputs like jailbreaks or prompt injection attemptsPersonal information filter: Screens model output for personally identifiable informationModeration: Identifies and blocks harmful content like hate speech and harassmentTool safeguards: Assesses tools based on risk level (for example, what kind of access they have)Rules-based protections: Implements blocklists, regular expression filters, and input limits to prevent known threatsOutput validation: Ensures the model’s responses align with brand or content policies

OpenAI suggests thinking of guardrails as a layered defense, understanding that no one guardrail will provide full protection, but a combined set might help keep things on track.

As with all the other OpenAI recommendations, guardrails should be implemented incrementally. Build in guardrails for risks you’ve identified prior to deployment, then add in and layer on more guardrails when new vulnerabilities are discovered.

10. Plan for human intervention

Face it. Something is going to break. When it does, your guardrails may not catch it. You can’t just set your AIs in action and leave them unattended. AIs are workflow process force multipliers, but they need human supervision.

Also: AI agents make great teammates, but don’t let them code alone – here’s why

Design your system to report to real people. Send anomalous behavior reports to people. Design your AIs to reach out to people when something doesn’t seem right.

OpenAI says that there are two triggers that require human intervention. The first is when an agent or workflow exceeds failure limits. Essentially, when the AI tries and tries and tries and keeps failing. The second is when the AI is about to take a high-risk action, whether that’s doing something that’s not reversible, is sensitive, or could cost a lot of money.

The last thing you want is an AI giving out ginormous refunds without some level of human oversight. AIs are perfect targets for malicious intervention. We humans aren’t perfect, but we can often spot things that our AIs can’t.

Best practices

OpenAI recommends that no matter how you’re planning on building an agentic AI solution, keep the components flexible and drive them by clear, well-structured prompts. Start small, validate with real users, and grow capabilities over time.

If you think about it, these are the same guidelines IT installations have used since the early days. The more things change, the more they stay the same.

Have you started experimenting with AI agents in your workflows? What kinds of tasks have proven hardest to automate? Do you think agents could finally make them manageable? Which of OpenAI’s best practices, like prompt templates or layered guardrails, resonate most with your own experience? Are there areas where you think human oversight is still absolutely essential? Let us know in the comments below.

You can follow my day-to-day project updates on social media. Be sure to subscribe to my weekly update newsletter, and follow me on Twitter/X at @DavidGewirtz, on Facebook at Facebook.com/DavidGewirtz, on Instagram at Instagram.com/DavidGewirtz, on Bluesky at @DavidGewirtz.com, and on YouTube at YouTube.com/DavidGewirtzTV.

Get the morning’s top stories in your inbox each day with our Tech Today newsletter.





Source link

Follow on Google News Follow on Flipboard
Share. Facebook Twitter Pinterest LinkedIn Tumblr Email Copy Link
Previous ArticleStanford HAI’s annual report highlights rapid adoption and growing accessibility of powerful AI systems
Next Article Ancestra says a lot about the current state of AI-generated videos
Advanced AI Editor
  • Website

Related Posts

Week in Review – Trump’s mobile venture, online data brokers used to target MN victims, and an OpenAI-Microsoft breakup

June 20, 2025

Week in Review – Trump’s mobile venture, online data brokers used to target MN victims, and an OpenAI-Microsoft breakup

June 20, 2025

Inside Microsoft’s complicated relationship with OpenAI

June 20, 2025
Leave A Reply Cancel Reply

Latest Posts

Melissa Errico To Sing Stephen Sondheim Classics In NY And London

Baghdad Group For Modern Art On View In America For First Time

Climate Activist Hurls Paint at Picasso Painting at Montreal Museum

Curtis Yarvin Details Proposed Titian-Centric ‘Art Hos’ US Pavilion

Latest Posts

EU Commission: “AI Gigafactories” to strengthen Europe as a business location

June 20, 2025

United States, China, and United Kingdom Lead the Global AI Ranking According to Stanford HAI’s Global AI Vibrancy Tool

June 20, 2025

Foundation AI: Cisco launches AI model for integration in security applications

June 20, 2025

Subscribe to News

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

Welcome to Advanced AI News—your ultimate destination for the latest advancements, insights, and breakthroughs in artificial intelligence.

At Advanced AI News, we are passionate about keeping you informed on the cutting edge of AI technology, from groundbreaking research to emerging startups, expert insights, and real-world applications. Our mission is to deliver high-quality, up-to-date, and insightful content that empowers AI enthusiasts, professionals, and businesses to stay ahead in this fast-evolving field.

Subscribe to Updates

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

YouTube LinkedIn
  • Home
  • About Us
  • Advertise With Us
  • Contact Us
  • DMCA
  • Privacy Policy
  • Terms & Conditions
© 2025 advancedainews. Designed by advancedainews.

Type above and press Enter to search. Press Esc to cancel.