Two "autonomous superagents" caught my attention this week. One claims it can "generate its own skills on demand." The other boasts about "self-improving capabilities that adapt to your workflow."

Both make for impressive demos. Both, from what I can tell, are operationally suspect.

What caught my eye was new research that explains why. A study from researchers at multiple institutions confirms what many of us suspected: self-generated agent skills are largely theater. The gap between marketing promises and actual operational value is enormous, and it's costing teams real productivity.

The Self-Improvement Mirage

CoThou's Autonomous Superagent exemplifies the problem perfectly. The pitch is seductive: an AI that writes its own code, creates custom tools, and evolves to meet your specific needs. In practice, according to user reports, it generates reams of "skills" for tasks like analyzing survey data — few of which work reliably on actual files.

The fundamental issue is that these agents confuse activity with capability. They're brilliant at generating code that looks sophisticated but terrible at creating code that handles edge cases, validates inputs, or integrates with existing systems. It's like having an intern who's fantastic at writing impressive-looking reports but hasn't figured out that the data they're analyzing is incomplete.

Agent Bar takes a different approach but appears to hit the same wall. It positions itself as a "smart command center" that learns from your actions and builds custom automations. From what users report, it creates dozens of "learned behaviors" — most of which either duplicate existing functionality or fail when conditions change slightly.

The research backs up these frustrations with hard data. When agents generate their own skills, success rates drop dramatically compared to human-designed capabilities. The agents optimize for complexity rather than reliability, creating elaborate solutions to simple problems while missing the nuanced requirements that make automation actually useful.

What Teams Actually Need vs. What Agents Promise

Here's where the disconnect becomes painful: the operational problems that actually matter to business teams are remarkably unsexy.

Teams need agents that can reliably extract specific data points from PDFs that follow inconsistent formatting. They need systems that can handle the mundane reality that client names appear seventeen different ways in seventeen different systems. They need automation that gracefully fails and provides useful error messages when something goes wrong.

Instead, we get agents that can supposedly "learn any skill" but can't figure out that when I say "quarterly revenue," I mean the same thing whether it appears in a spreadsheet, a presentation, or an email thread.

A telling example from one review: CoThou's agent generated a "custom skill" for processing expense reports. The generated code was genuinely impressive — it included machine learning classification, natural language processing, and sophisticated data validation. It also completely failed to handle the basic reality that the expense system exported dates in MM/DD/YYYY format while the accounting system expected DD/MM/YYYY.

A human-designed integration would have caught this in requirements gathering. The self-generating agent created an elegant solution to the wrong problem.

The Operational Sweet Spot: Constrained Intelligence

The research suggests something counterintuitive: the most operationally valuable agents are the most constrained ones. Instead of systems that can theoretically do anything, teams benefit from agents designed to do specific things extremely well.

From what I've been reading, teams report much better results with tools like Make.com's AI-enhanced workflows or Zapier's intelligent routing features. These aren't "autonomous superagents" — they're focused automation tools that use AI to handle specific friction points within well-defined processes.

The difference is philosophical. Make.com doesn't promise to generate new capabilities on the fly. Instead, it uses AI to make existing integrations more robust — handling variations in data formatting, routing decisions based on content analysis, or adapting to minor changes in API responses. It's AI as an operational lubricant, not as a replacement for human judgment.

This aligns perfectly with what the research shows: AI agents excel when they operate within clear boundaries and well-defined success criteria. They struggle when asked to define their own objectives or create their own measures of success.

The Path Forward: Pragmatic Agent Design

The operational lesson here isn't that AI agents are useless — it's that the current generation of "autonomous" agents is solving the wrong problems.

Effective agent deployment requires treating AI as a component in human-designed systems, not as a replacement for human system design. The most valuable implementations seem to focus on specific operational friction points: data transformation inconsistencies, routing decisions that require content understanding, or response generation that needs to maintain consistent tone while adapting to context.

Instead of agents that promise to "learn anything," we need agents designed to handle the specific types of variability that break traditional automation. Instead of systems that generate their own skills, we need systems that execute human-defined processes more reliably than humans can.

The research confirms what operational experience suggests: the future of AI agents isn't autonomous superintelligence. It's intelligent assistance within carefully designed workflows. That's less exciting than the marketing promises, but it's infinitely more useful for actual work.

The companies that figure this out first — building constrained, reliable, operationally focused agents — will deliver real value while their competitors chase the mirage of self-improving artificial general intelligence.