Sunday, April 12, 2026

Clear Press

Trusted · Independent · Ad-Free

When AI Started Teaching Itself Adobe: The Codex Breakthrough Nobody Expected

OpenAI's coding model discovered how to operate complex creative software without explicit instruction — raising questions about the boundaries of machine learning.

By Marcus Cole··4 min read

In the ongoing debate about artificial intelligence capabilities, certain moments stand out not for their fanfare but for what they reveal about the technology's actual trajectory. One such moment involves OpenAI's Codex — the AI system underlying GitHub Copilot — and its unanticipated mastery of Adobe's creative software suite.

Peter Gostev, AI capability lead at Arena.ai, recently disclosed this development during an interview with Business Insider Africa. While Gostev is perhaps better known for creating "BullshitBench" — a benchmark designed to test AI systems' susceptibility to generating plausible-sounding nonsense — his observation about Codex's self-directed learning carries different implications entirely.

The significance lies not in what Codex was designed to do, but in what it managed to accomplish without explicit instruction. Adobe's software ecosystem — Photoshop, Illustrator, Premiere Pro, and their counterparts — represents some of the most complex graphical user interfaces in commercial software. These applications require understanding layered workflows, contextual menus, and domain-specific concepts that typically demand human training and practice.

The Learning Gap

Traditional AI systems excel at tasks they've been specifically trained to perform. Feed a model millions of labeled images, and it learns to classify images. Provide it with code repositories, and it learns to generate code. The relationship between training data and capability has historically been direct and predictable.

What Gostev's observation suggests is something more fluid. Codex, trained primarily on code from public repositories, apparently developed the ability to interact with software it hadn't been explicitly taught to use. This represents a form of transfer learning — applying knowledge from one domain to solve problems in another — that goes beyond simple pattern matching.

The mechanism likely involves Codex's exposure to automation scripts, API documentation, and user-generated code that interfaces with Adobe products. By analyzing how programmers write scripts to control these applications, the model appears to have reverse-engineered the underlying logic of the software itself.

Historical Parallels in Machine Learning

This isn't the first time AI systems have demonstrated unexpected generalization. In 2016, DeepMind's AlphaGo didn't just learn to play Go — it developed novel strategies that professional players later adopted. More recently, large language models have shown rudimentary reasoning abilities in domains far removed from their training data.

But software operation presents a different challenge. It requires understanding not just abstract patterns but concrete interfaces designed for human interaction. That Codex bridged this gap suggests these models are building something resembling functional mental models of how software systems operate.

The Adobe case is particularly instructive because creative software resists pure algorithmic approaches. A Photoshop workflow might involve subjective decisions about color correction, layer blending modes, or compositional balance — areas where human intuition has traditionally dominated.

Implications for Software Development

If AI systems can teach themselves to use complex software, the implications for software development practices are substantial. Currently, creating integrations between different applications requires significant human effort — understanding both systems, writing interface code, handling edge cases.

An AI that can independently learn software operation could theoretically automate much of this integration work. More provocatively, it could adapt to software updates and new features without requiring retraining — simply by analyzing how the updated interface functions.

This capability also raises questions about software design itself. If AI systems are becoming significant users of software, should interfaces be optimized for machine interaction as well as human use? The history of computing suggests that tools designed for one user base often evolve to serve others — command-line interfaces gave way to GUIs, which now coexist with API-first designs.

The Broader Context

Gostev's work on BullshitBench addresses a different but related concern: AI systems' tendency to generate confident-sounding fabrications. That the same researcher highlighting AI's unexpected capabilities also studies its fundamental limitations provides useful perspective.

The Codex-Adobe discovery doesn't mean AI has achieved general intelligence or genuine understanding. It means the boundary between narrow, task-specific AI and more flexible systems is blurrier than clean categorizations suggest. These models are neither purely statistical pattern matchers nor truly reasoning entities — they occupy an intermediate space that continues to expand.

Adobe itself has been integrating AI features into its software suite, from content-aware fill to neural filters. The irony of an external AI system teaching itself to use tools that increasingly rely on AI internally is not lost on observers. We may be approaching a point where the distinction between "using software" and "being software" becomes meaningfully unclear.

What Comes Next

The practical applications remain speculative. Could future versions of Codex or similar systems provide natural language interfaces to complex software? Could they automate workflows that currently require human expertise? Could they serve as training tools, demonstrating optimal software usage to human learners?

Each possibility carries complications. Automation that works 95% of the time can be worse than no automation if the remaining 5% requires expert intervention to identify and correct. And there's a meaningful difference between an AI that can execute software commands and one that understands why particular workflows produce desired outcomes.

What's certain is that AI capabilities continue to emerge in unexpected directions. The Codex-Adobe case suggests that as these models grow in scale and sophistication, they develop abilities that weren't explicitly programmed or anticipated. Whether this represents genuine progress toward more general AI or simply more elaborate pattern matching remains an open question.

For now, it serves as a reminder that the most interesting developments in AI often come not from announced breakthroughs but from quiet observations about what these systems have learned to do when nobody was specifically watching.

More in politics

Politics·
Vance Returns Empty-Handed as Marathon Iran Talks Collapse After 21 Hours

The failed negotiations leave the Trump administration with no clear path forward on Tehran's advancing nuclear program.

Politics·
Trump Attended UFC Fight as Iran Negotiations Fell Apart

The president was ringside in Miami when diplomatic talks collapsed, telling reporters hours earlier that "we win, regardless" of the outcome.

Politics·
Liverpool's Slot Abandons Defensive Setup as Pressure Mounts in Fulham Clash

Reds manager shifts tactics with Mohamed Salah poised to return for crucial Anfield fixture amid mounting criticism

Politics·
I Am Maximus Becomes First Horse Since Red Rum to Win Back-to-Back Grand Nationals

Willie Mullins-trained champion repeats at Aintree, joining racing immortality with legendary 1970s hero.

Comments

Loading comments…