OpenAI’s next AI agent is a self-testing software engineer that does what humans won’t

OpenAI is set to launch A-SWE, an AI agent that can perform software engineering tasks along with quality assurance and bug testing. However, previous claims about AI capabilities have often been overstated, leading to scepticism about the true potential of these models.

Aman Gupta
Published13 Apr 2025, 07:26 AM IST
OpenAI is working on developing a new AI agent
OpenAI is working on developing a new AI agent(Represenational image generated using AI)

OpenAI is working on a new AI agent that is sure to cause quite a stir if the description given by OpenAI CFO Sarah Friar is anything to go by. Friar recently confirmed that the ChatGPT maker will soon be releasing its third AI agent, “A-SWE”, or Agentic Software Engineer. This agent will not only be able to perform the tasks that a normal software engineer can do, but will also be able to take on other additional tasks such as quality assurance, bug testing, and bug bashing.

Also Read | How to create free action figure AI images using ChatGPT: Check our guide

In an interaction with Goldman Sachs, Friar said, “The third (AI Agent) that is coming is what we call A-SWE… agentic software engineer. And this is not just augmenting the current software engineers in the workforce, which is kind of what we can do today through Copilot, but instead, it's literally an agentic software engineer that can build an app for you.”

Friar added, “It can take a PR that you would give to any other engineer and go build it. But not only does it build it, it does all the things that software engineers hate to do. It does its own QA, its own quality assurance, its own bug testing and bug bashing, and it does documentation, things you can never get software engineers to do so, suddenly, you can force, multiply your software engineering workforce.”

OpenAI launched its first AI agent, Operator, in January, soon followed by Deep Research in February. Both AI offerings are currently only available to ChatGPT's paying customers.

Why you shouldn't freak out right now?

OpenAI has a history of making tall claims about its products, some of which never materialise. Take Deep Research, for example. At the time of its launch, OpenAI said that the new tool will be able to replace a research assistant, a claim that Friar repeated on Friday.

Also Read | Job seeker forgets to delete ChatGPT prompt, sends awkward email to recruiter

While many of OpenAI’s peers, including xAI and Perplexity, have rolled out similar tools, it’s still unclear how much of a research assistant’s role these AI models can genuinely take over. The reason? They're still prone to hallucinations—confidently generating information that simply isn’t true.

The problem isn't that these large language models are inaccurate because, after all, humans make mistakes, too. What’s more concerning is that these models present false information with an air of absolute confidence, making it harder to separate fact from fabrication. And that hasn’t changed much since ChatGPT first rolled out to the public in late 2022.

So when OpenAI says that its upcoming AI agents will be able to essentially do all of what the current software engineers do and more, take those claims with a pinch of salt.

Catch all the Technology News and Updates on Live Mint. Download The Mint News App to get Daily Market Updates & Live Business News.

Business NewsTechnologyNewsOpenAI’s next AI agent is a self-testing software engineer that does what humans won’t
MoreLess
First Published:13 Apr 2025, 07:26 AM IST
Most Active Stocks
Market Snapshot
  • Top Gainers
  • Top Losers
  • 52 Week High
Recommended For You
    More Recommendations
    Gold Prices
    • 24K
    • 22K
    Fuel Price
    • Petrol
    • Diesel
    Popular in Technology