No one knows what the hell an AI agent is
AI agents are vaguely defined, leading to confusion in tech.

The term 'AI agents' has been gaining traction in the tech industry, with CEOs like Sam Altman, Satya Nadella, and Marc Benioff making bold predictions about their role in the workforce. However, there is a lack of consensus on what constitutes an AI agent, leading to confusion and frustration among companies and customers. OpenAI, Microsoft, and Salesforce each have their interpretations, with definitions ranging from automated systems to LLMs equipped with tools. Ryan Salva, formerly of GitHub Copilot and now with Google, criticizes the industry's overuse of the term to the point of meaninglessness.
OpenAI recently tried to clarify their stance, referring to agents as automated systems that can independently accomplish tasks. However, their own documentation varied, describing agents as LLMs (Large Language Models) equipped with instructions and tools. Lehr Pathak of OpenAI added to the confusion by suggesting that 'agents' and 'assistants' were interchangeable, contrasting Microsoft's attempt to separate them, with agents tailored for specific expertise and assistants handling general tasks.
Anthropic provides a broader perspective, noting that agents can be fully autonomous or prescriptive, potentially leading to misaligned expectations and ROI challenges. Meanwhile, Salesforce's definition encompasses six categories of agents, from simple reflex types to utility-based ones. Jim Rowan from Deloitte points out that the lack of a standardized understanding makes it difficult to ensure consistent outcomes in AI projects.
Marketing's role in diluting the technical meaning of 'agents' is notable, according to Andrew Ng, who attributes this to companies wielding marketing influence without strict adherence to technical definitions. This flexibility might enable innovation but also creates ambiguities, making it challenging to evaluate the effectiveness and value of AI agents.
As companies like OpenAI, Google, and Perplexity experiment with their initial agents, such as OpenAI's Operator and Google's Project Mariner, their capabilities vary widely. Rich Villars of IDC observes that tech companies often prioritize their technical objectives over rigid definitions, further complicating the landscape. The ever-evolving and nebulous nature of AI agents makes it unlikely that a unified definition will emerge soon.
Sources: TechCrunch, OpenAI, Microsoft, Salesforce, Anthropic, DeepLearning.ai, Deloitte