Software Is Changing (Again) – The Dawn of Software 3.0

When Andrej Karpathy titled his recent keynote “Software Is Changing (Again),” it wasn’t just a nice slogan. It marks what he argues is a fundamental shift in the way we build, think about, and interact with software. Based on his talk at AI Startup School (June 2025), here’s what “Software 3.0” means, why it matters, and how you can prepare.

What Is Changing: The Three Eras

Karpathy frames software evolution in three eras:

EraWhat it meantKey Characteristics
Software 1.0Traditional code-first era: developers write explicit instructions, rules, algorithms. Think C++, Java, manual logic.Highly deterministic, rule-based; heavy human specification; hard to scale certain tasks especially with unstructured or subtle data.
Software 2.0Rise of machine learning / neural nets: train models on data vs hand-coding every condition. The model “learns” patterns.Better handling of unstructured data (images, text), but still needs labeled data, training, testing, deployment. Not always interpretable.
Software 3.0The new shift: large language models (LLMs) + natural language / prompt-driven interfaces become first-class means of programming. “English as code,” vibe coding, agents, prompt/context engineering.You describe what you want; software (via LLMs) helps shape it. More autonomy, more natural interfaces. Faster prototyping. But also new risks (hallucinations, brittleness, security, lack of memory) and need for human oversight.

Let’s break down what he means.

The “Old” World: Software 1.0

For the last 50+ years, we’ve lived in the era of what Karpathy calls “Software 1.0.” This is the software we all know and love (or love to hate). It’s built on a simple, deterministic principle:

  1. A human programmer writes explicit rules in a programming language like Python, C++, or Java.
  2. The computer compiles these rules into binary instructions.
  3. The CPU executes these instructions, producing a predictable output for a given input.

Think of a tax calculation function. The programmer defines the logic: if income > X, then tax = Y. It’s precise, debuggable, and entirely human-written. The programmer’s intellect is directly encoded into the logic. The problem? Its capabilities are limited by the programmer’s ability to foresee and explicitly code for every possible scenario. Teaching a computer to recognize a cat using Software 1.0 would require writing millions of lines of code describing edges, textures, and shapes—a nearly impossible task.

The Emerging World: Software 2.0

The new paradigm, “Software 2.0,” is a complete inversion of this process. Instead of writing the rules, we curate data and specify a goal.

  1. A human programmer (or, increasingly, an “AI Engineer”) gathers a large dataset (e.g., millions of images of cats and “not-cats”).
  2. They define a flexible, neural network architecture—a blank slate capable of learning complex patterns.
  3. They specify a goal or a “loss function” (e.g., “minimize the number of incorrect cat identifications”).
  4. Using massive computational power (GPUs/TPUs), an optimization algorithm (like backpropagation) searches the vast space of possible neural network configurations to find one that best maps the inputs to the desired outputs.

The “code” of Software 2.0 isn’t a set of human-readable if/else statements. It’s the learned weights and parameters of the neural network—a massive matrix of numbers that is completely inscrutable to a human. We didn’t write it; we grew it from data.

As Karpathy famously put it, think of the neural network as the source code, and the process of training as “compiling” the data into an executable model.

The Rise of the “AI Engineer” and the LLM Operating System: Software 3.0

Karpathy’s most recent observations take this a step further with the explosion of Large Language Models (LLMs) like GPT-4. He describes the modern AI stack as a new kind of operating system.

In this analogy:

  • The LLM is the CPU—the core processor, but for cognitive tasks.
  • The context window is the RAM—the working memory.
  • Prompting is the programming—the primary way we “instruct” this new computer.
  • Tools and APIs (web search, code execution, calculators) are the peripherals and I/O.

This reframes the role of the “AI Engineer.” They are now orchestrating these powerful, pre-trained models, “programming” them through sophisticated prompting, retrieval-augmented generation (RAG), and fine-tuning to build complex applications. This is the practical, applied side of the Software 2.0 revolution that is currently creating a gold rush in the tech industry.

Core Themes from Karpathy’s Keynote

Here are some of the biggest insights:

  • Natural Language as Programming Interface: Instead of writing verbose code, developers (and increasingly non-developers) can prompt LLMs in English (or human language) to generate code, UI, workflows. Karpathy demos “MenuGen,” a vibe-coding app prototype, as example of how quickly one can build via prompts.
  • LLMs as the New Platform / OS: Karpathy likens current LLMs to utilities or operating systems: infrastructure layers that provide default capabilities and can be built upon. Labs like OpenAI, Anthropic become “model fabs” producing foundational layers; people will build on top of them.
  • Vibe Coding & Prompt Engineering: He introduces / popularizes the idea of “vibe coding” — where the code itself feels less visible, you interact via prompts, edits, possibly via higher levels of abstraction. With that comes the need for better prompt or context engineering to reduce errors.
  • Jagged Intelligence: LLMs are powerful in some domains, weak in others. They may hallucinate, err at basic math, or make logically inconsistent decisions. Part of working well with this new paradigm is designing for those imperfections. Human-in-the-loop, verification, testing.
  • Building Infrastructure for Agents: Karpathy argues that software needs to be architected so LLMs / agents can interact with it, consume documentation, knowledge bases, have memory/context, manage feedback loops. Things like llms.txt, agent-friendly docs, and file & knowledge storage that is easy for agents to read/interpret.

Final Thoughts: Is This Another Shift – Or the Same One Again?

Karpathy would say this is not just incremental – “Software is changing again” implies something qualitatively different. In many ways, Software 3.0 composes both the lessons of Software 1.0 (like performance, correctness, architectural rigor) and Software 2.0 (learning from data, dealing with unstructured inputs), but adds a layer where language, agents, and human-AI collaboration become central.

In a nutshell: we’re not just upgrading the tools; we’re redefining what software means.

Federated Learning, Reinforcement Learning, and Imitation Learning: AI Paradigms Powering the Next Generation of Intelligent Systems

Artificial Intelligence (AI) has evolved beyond traditional models that simply learn from centralized datasets. Today, organizations are leveraging Federated Learning, Reinforcement Learning, and Imitation Learning to create more intelligent, scalable, and privacy-preserving systems. In this article, we decode these paradigms and explore how they’re being used in the real world across industries.

Federated Learning (FL)

What It Is:

Federated Learning is a decentralized machine learning approach where the model is trained across multiple devices or servers holding local data samples, without exchanging them. Instead of sending data to a central server, only model updates are shared, preserving data privacy.

Key Features:

  • Data stays on-device
  • Ensures data privacy and security
  • Reduces latency and bandwidth requirements

Real-Life Use Cases:

  1. Healthcare:
    • Example: Hospitals collaboratively train diagnostic models (e.g., for brain tumor detection from MRIs) without sharing sensitive patient data.
    • Players: NVIDIA Clara, Owkin
  2. Financial Services:
    • Example: Banks train fraud detection models across different branches or countries, avoiding cross-border data sharing.
  3. Smartphones / IoT:
    • Example: Google uses FL in Gboard to improve next-word prediction based on typing habits, without uploading keystroke data to its servers.

Reinforcement Learning (RL)

What It Is:

Reinforcement Learning is a paradigm where an agent learns to make sequential decisions by interacting with an environment, receiving rewards or penalties based on its actions.

Key Features:

  • Focused on learning optimal policies
  • Works best in dynamic, interactive environments
  • Learns from trial-and-error

Real-Life Use Cases:

  1. Retail & E-commerce:
    • Example: Optimizing product recommendations and personalized pricing strategies by learning customer behavior.
    • Player: Amazon uses RL in their retail engine.
  2. Robotics & Manufacturing:
    • Example: A robot arm learning to sort or assemble components by maximizing efficiency and precision.
    • Players: Boston Dynamics, FANUC.
  3. Energy:
    • Example: Google DeepMind applied RL to reduce cooling energy consumption in Google data centers by up to 40%.
  4. Airlines / Logistics:
    • Example: Dynamic route planning for aircrafts or delivery trucks to minimize fuel consumption and delays.

Imitation Learning (IL)

What It Is:

Imitation Learning is a form of supervised learning where the model learns to mimic expert behavior by observing demonstrations, rather than learning from scratch via trial-and-error.

Key Features:

  • Ideal for situations where safe exploration is needed
  • Requires a high-quality expert dataset
  • Often used as a starting point before fine-tuning with RL

Real-Life Use Cases:

  1. Autonomous Vehicles:
    • Example: Self-driving cars learn to navigate complex traffic by observing professional driver behavior.
    • Players: Waymo, Tesla (for some autopilot capabilities).
  2. Aviation Training Simulators:
    • Example: Simulators that mimic experienced pilots’ actions for training purposes.
  3. Gaming AI:
    • Example: AI bots learning to play video games like Dota 2 or StarCraft by mimicking professional human players.
  4. Warehouse Automation:
    • Example: Robots that imitate human pickers to optimize picking routes and behavior.

How They Complement Each Other

These paradigms aren’t mutually exclusive:

  • Federated RL is being explored for multi-agent decentralized systems (e.g., fleets of autonomous drones).
  • Imitation Learning + RL: IL can provide a strong initial policy which RL then optimizes further through exploration.

Closing Thoughts

From privacy-centric learning to autonomous decision-making and human-like imitation, Federated Learning, Reinforcement Learning, and Imitation Learning are shaping the AI landscape across industries. Businesses embracing these paradigms are not only improving efficiency but also future-proofing their operations in a world increasingly defined by intelligent, adaptive systems.

From Bots to Brains: Why AI Is Outpacing RPA in the Automation Race

In the early 2010s, Robotic Process Automation (RPA) became the darling of digital transformation. It promised businesses a way to automate repetitive, rule-based tasks – fast, scalable, and with minimal disruption.

But fast forward to 2025, and the automation landscape looks very different. The rise of Artificial Intelligence (AI), especially Generative AI (GenAI) and Agentic AI, is redefining what automation means.

So, what’s the difference between RPA and AI? Why are enterprises increasingly favoring AI over traditional RPA?

Let’s break it down.

What Is Robotic Process Automation (RPA)?

RPA is software that mimics human actions to execute structured, rule-based tasks across systems. It works well for:

  • Data entry and validation
  • Invoice processing
  • Copy-paste jobs between applications
  • Simple workflow automation

RPA bots follow pre-defined scripts, and if something changes (like a UI tweak), they often break. They’re fast but not intelligent.

What Is Artificial Intelligence (AI)?

AI enables systems to simulate human intelligence – from recognizing images and understanding language to making decisions. It includes:

  • Machine Learning (pattern recognition, forecasting)
  • Natural Language Processing (NLP) (chatbots, document reading)
  • Generative AI (content creation, summarization, ideation)
  • Agentic AI (autonomous systems that can plan, act, and adapt)

AI systems learn from data, evolve over time, and can handle unstructured, ambiguous scenarios – something RPA cannot do.

RPA vs. AI: A Quick Comparison

FeatureRPAAI / GenAI / Agentic AI
NatureRule-basedData-driven, adaptive
Task TypeRepetitive, structuredUnstructured, dynamic
Learning AbilityNoYes (ML)
ScalabilityLimited by scriptsScales with data models
Cognitive CapabilitiesNoneNatural language, vision, decision-making
MaintenanceHigh (fragile bots)Low-to-medium (models learn and adjust)

Why Enterprises Are Shifting to AI/GenAI/Agentic AI

  1. Handling Complex Use Cases
    AI can interpret documents, summarize legal contracts, analyze sentiment, and make predictive decisions – things RPA was never built for.
  2. Scalability Without Fragility
    GenAI-based assistants don’t break when the UI changes. They can adapt and even reason contextually, reducing the brittle nature of traditional automation.
  3. Contextual Understanding
    Agentic AI systems can take on tasks like a virtual analyst or associate – autonomously interacting with APIs, querying data, and even making decisions in real-time.
  4. Better ROI
    While RPA was often a stopgap solution, AI brings strategic transformation – automating not just tasks, but insights and decision-making.
  5. Human-like Interaction
    With conversational AI and GenAI copilots, enterprises now prefer solutions that work with humans, not just automate behind the scenes.
  6. Integration with Modern Tech Stacks
    AI integrates seamlessly with cloud-native ecosystems, APIs, and data lakes – ideal for digital-first businesses.

Example Use-Cases Driving the Shift

IndustryRPA Use-CaseAI/GenAI Use-Case
BankingLoan document sortingAI extracting insights, summarizing risk
HealthcarePatient appointment schedulingAI interpreting EHRs, triaging cases
RetailOrder reconciliationGenAI creating personalized product offers
TravelInvoice validationAI assistant managing full travel itineraries
ManufacturingInventory updatesAgentic AI optimizing supply chain flows

Final Thoughts: From Automation to Autonomy

RPA was a critical first step in the automation journey – but today, businesses want more than faster copy-paste. They want smart, self-learning systems that can understand, generate, decide, and act.

That’s why the spotlight is now firmly on AI – and its GenAI and Agentic variants.

If you’re still relying on RPA-only architectures, it’s time to rethink your automation roadmap. Because in the age of AI, it’s not just about doing things faster – it’s about doing things smarter.

Rather than a complete replacement, it’s believed that the future lies in combining RPA with AI (a trend called “Hyperautomation”). RPA handles structured tasks, while AI manages cognitive functions, creating a seamless automation ecosystem.

Additional resource for reference: https://www.techtarget.com/searchenterpriseai/tip/Compare-AI-agents-vs-RPA-Key-differences-and-overlap