Let’s dive into Andrej Karpathy’s X post from June 4, 2025, and break it down in detail. Karpathy, a prominent figure in AI and machine learning (formerly at Tesla and OpenAI, now with xAI), is commenting on the future of software products in the context of human-AI collaboration. His post highlights a critical shift in how software interfaces and architectures need to evolve to remain relevant as AI, particularly large language models (LLMs), becomes integral to workflows. We will explain the post, elaborate on its implications, and provide detailed context for the concepts he’s addressing.
The Core Argument
Karpathy’s central thesis is that software products with complex, opaque user interfaces (UIs) and data formats that aren’t easily accessible or manipulable by AI are at risk of becoming obsolete ("ngmi" = "not gonna make it") in an era where human-AI collaboration is becoming the norm. He argues that for software to thrive, it must:
Expose its underlying representations and settings in a way that AI can understand and manipulate (e.g., through text-based formats or scripting).
Enable seamless integration with AI systems so that LLMs can act as co-pilots for professionals and empower "prosumers" (aspiring professionals or hobbyists) to engage in "vibe coding"—a more intuitive, less rigid form of programming.
He then categorizes software products into risk levels based on how AI-friendly their architectures are and warns that products failing to adapt will struggle as AI capabilities advance.
Breaking Down the Post
1. The Problem with Traditional Software UIs
Karpathy starts by critiquing software with "extensive/rich UIs" that rely heavily on graphical elements like sliders, switches, and menus, but lack scripting support and use "opaque, custom, binary formats." Examples of such software include Adobe products (like Photoshop), digital audio workstations (DAWs), and CAD/3D modeling tools.
What are "opaque, custom, binary formats"?
Binary formats are non-human-readable data representations (e.g., a .psd file in Photoshop or a proprietary CAD file). Unlike text-based formats (e.g., JSON, XML, or Python scripts), binary formats are encoded in a way that’s efficient for machines to process but difficult for humans or AI to interpret without specialized tools.
For example, a Photoshop .psd file contains layers, effects, and settings, but you can’t easily open it in a text editor to see or modify its structure. This makes it hard for an AI to understand or manipulate the file programmatically.
Why is this a problem for human-AI collaboration?
LLMs, like the ones Karpathy is referring to, excel at processing and generating text. They can understand and manipulate code, configurations, or data if it’s in a text-based format (e.g., a script that defines a 3D model in Blender). However, they struggle with binary formats because they can’t "read" them directly.
In a collaborative workflow, an AI co-pilot needs to understand the state of a project (e.g., the settings of a Photoshop filter or the parameters of a 3D model) and modify it based on user prompts (e.g., "increase the brightness by 10%"). If the software’s data is locked in a binary format, the AI can’t access or manipulate it without cumbersome workarounds like screen scraping or UI automation.
2. The Role of Scripting and Text-Based Representations
Karpathy emphasizes that software needs to be scriptable and use text-based domain-specific languages (DSLs) to be AI-friendly.
3. Human-AI Collaboration: Professionals and Prosumers
Karpathy highlights two groups that benefit from AI-friendly software:
4. Risk Spectrum of Software Products
Karpathy categorizes software into four risk levels based on how AI-friendly they are:
High Risk (Binary Objects, No Text DSL):
Examples: Adobe products (Photoshop, Premiere), DAWs (Ableton Live, FL Studio), CAD/3D tools (SolidWorks, AutoCAD).
These products rely heavily on binary formats and lack comprehensive scripting support. For example, Adobe Photoshop has some scripting capabilities (via JavaScript or Actions), but its core file format (.psd) is binary, making it hard for AI to manipulate directly.
Karpathy predicts these products are at the highest risk of becoming obsolete because they can’t integrate seamlessly with AI workflows.
Medium-High Risk (Partially Text Scriptable):
Examples: Blender, Unity.
Blender and Unity are more AI-friendly because they support scripting (Blender via Python, Unity via C#). For instance, the Stack Overflow result shows how to instantiate a Blender model in Unity, highlighting that these tools can be manipulated programmatically to some extent.
However, they still have limitations. Not all features in Blender or Unity are fully scriptable, and some data (e.g., Unity’s scene hierarchy) may still be opaque to AI without additional parsing.
Medium-Low Risk (Mostly Text, Some Automation):
Examples: Excel.
Excel is largely text-based (spreadsheets are essentially tables of data), and it supports automation via VBA (Visual Basic for Applications) and a plugin ecosystem. An AI can easily read an Excel file (e.g., as a CSV) and manipulate its contents.
However, Excel still has some graphical elements (e.g., its charting tools) that aren’t fully scriptable, which is why it’s not "low risk."
Low Risk (All Text, AI-Friendly):
Examples: IDEs (VS Code), Figma, Jupyter, Obsidian.
These tools are already text-native. For example, VS Code uses JSON for its settings and extensions, Figma’s files can be manipulated via its API (as noted in the UX Pilot search result), and Jupyter notebooks are essentially JSON files with code and markdown.
These products are "lucky" because their architecture aligns naturally with how LLMs operate, making them ideal for human-AI collaboration.
5. The Future of AI and UI/UX
Karpathy acknowledges that AI will improve at interacting with traditional UIs (e.g., through tools like "Operator," likely referring to AI systems that can navigate graphical interfaces by interpreting screenshots and simulating clicks). However, he argues that relying solely on this approach is risky.
Elaborating on the Implications
For Software Companies
Karpathy’s post is a wake-up call for software companies, especially those with legacy products. Companies like Adobe, which dominate creative industries, face a significant challenge:
Adapting Legacy Systems:
The Adobe news result mentions Adobe’s efforts to integrate AI into its Experience Cloud (e.g., AI-powered content creation with Adobe Firefly). However, Karpathy’s critique suggests that superficial AI features (like generating images) aren’t enough. Adobe needs to fundamentally rethink how its products expose data and functionality to AI.
For example, Adobe could develop a text-based DSL for Photoshop that allows users (and AI) to define layers, filters, and adjustments as code. This would make Photoshop more scriptable and AI-friendly.
Competitive Pressure:
Karpathy’s reply to Godwin Osama ("Figma to buy Adobe 2035?") is a provocative suggestion that text-native, AI-friendly tools like Figma could eventually overtake legacy giants like Adobe. Figma’s API and collaborative, web-based nature make it more adaptable to AI workflows, as noted in the UX Pilot search result.
For Developers and Designers
For professionals and prosumers, Karpathy’s vision of human-AI collaboration is empowering but requires a shift in mindset:
For AI Development
Karpathy’s post also has implications for the AI community:
For the Broader Tech Ecosystem
Karpathy’s post ties into broader trends in technology:
Connection to Other Posts in the Thread
Detailed Example: Adobe vs. Figma
Let’s compare Adobe Photoshop (high risk) and Figma (low risk) to illustrate Karpathy’s point:
Photoshop:
File Format: .psd is binary, making it hard for AI to read or modify directly.
Scripting: Photoshop supports some scripting via JavaScript, but it’s not comprehensive. Many features are only accessible through the UI.
AI Integration: Adobe is adding AI features (e.g., Firefly for generative content), but these are surface-level. An AI can’t deeply understand or manipulate a .psd file’s structure.
Risk: High. Photoshop’s architecture makes it difficult for AI to act as a co-pilot.
Figma:
File Format: Figma’s data is accessible via its API, which uses JSON (a text-based format). This makes it easy for AI to read and manipulate designs.
Scripting: Figma supports plugins and has a REST API, allowing programmatic control.
AI Integration: As the UX Pilot result shows, Figma is already integrating with AI tools to generate designs, wireframes, and sitemaps.
Risk: Low. Figma’s text-native, API-driven approach makes it ideal for human-AI collaboration.
Conclusion
Karpathy’s post is a forward-looking critique of software design in the age of AI. He argues that software must evolve to be AI-friendly by adopting text-based representations, scripting support, and DSLs, or risk becoming obsolete as human-AI collaboration becomes the norm. His risk spectrum provides a clear framework for understanding which products are well-positioned (e.g., Figma, VS Code) and which are at risk (e.g., Adobe products, DAWs). The broader implications touch on software development, AI research, and the future of creative workflows, urging companies to adapt now rather than wait for AI to solve the problem through UI automation.
For professionals and prosumers, this shift promises a future where AI co-pilots can enhance productivity and creativity, but it requires software to meet AI halfway. For companies, it’s a call to action: rethink your architecture, or risk being left behind in the AI-driven future.
Here’s a list of 10 key words and phrases from the detailed explanation of Andrej Karpathy’s X post, capturing the core concepts and themes discussed:
Human-AI Collaboration
Central to Karpathy’s argument, this phrase refers to the growing partnership between humans and AI systems, particularly LLMs, in creative and professional workflows.
Opaque Binary Formats
Text-Based DSL (Domain-Specific Language)
Scripting Support
Vibe Coding
Karpathy’s term for an intuitive, exploratory coding approach enabled by AI, where users describe goals in natural language, and AI generates the necessary code.
Prosumers
Risk Spectrum
Karpathy’s categorization of software into high, medium-high, medium-low, and low risk based on their AI compatibility (e.g., Adobe at high risk, Figma at low risk).
AI Co-Pilot
UI Automation
Text-Native Software
These terms encapsulate the main ideas, challenges, and proposed solutions in the context of evolving software design for the AI era.