The Missing Ingredient in Machine Intelligence: Why Evolution, Not Data, Determines the Future of AI

The technology industry keeps repeating the same reassuring story. Make the models bigger, feed them more data, buy more chips, and intelligence will rise in a smooth upward line. It is an engineering fable: tidy, quantitative, and comforting. But if you take evolution seriously, that story is already finished. Scaling explains how to polish a tool. It does not explain how to create a living agent.

Evolution did not give us brains by optimizing a loss function on a static dataset. It produced minds through a brutal process built on scarcity, competition, reproduction, and fear. Above all, evolution gave every organism something to lose. That is the missing ingredient in today’s systems. Until we confront it directly, we are pretending that we can reach human level intelligence without touching the forces that created human beings.

Evolution is not training on a dataset

Training a model is like cramming a clever student for an exam with past papers. Evolution is different. It never sees the same exam twice. It throws bodies into a changing world, lets most of them fail, and only keeps the rare configurations that survive and reproduce. That is not parameter fitting. It is a population level search under pressure.

What evolution really optimized for

Start with a simple observation. Life is not neutral. Every living thing behaves as if survival and replication matter. A bacteria that does not try to divide leaves no descendants. A mammal that does not avoid predators disappears from the gene pool. Across species, you see the same pattern. The details differ, but the direction of travel is stable: preserve yourself, create copies, adapt when the world changes.

That direction is enforced by scarcity. Food is limited. Space is limited. Mates are limited. When there are more mouths than meals, organisms that hesitate die. Competition is not an optional game. It is the filter that decides which strategies persist.

Over time, evolution discovered shortcuts. Raw reinforcement is slow and expensive. So the system built emotional heuristics into our nervous systems. Fear keeps you away from cliffs before you calculate expected value. Attachment keeps you near offspring without solving a cost benefit equation. Curiosity pulls you toward the unknown, because organisms that explored their environment handled shocks better than those that stayed still.

Emotions as compressed survival code

Emotions are not mystical decorations. They are compressed survival policies. Fear says “avoid this configuration of the world.” Desire says “move toward that one.” Shame and pride regulate social standing in cooperative groups. Curiosity pushes exploration when direct reward is sparse. Strip these away and the organism still has a brain, but it has lost the fast heuristics that make decisions when there is no time to think.

Humans inherit all of this machinery, but with extra layers. We model ourselves. We imagine futures. We build cultures and technologies. Yet the core remains the same. A human being who feels nothing, values nothing, and fears nothing does not become perfectly rational. They become manipulable. The slightest nudge determines their choice. Indifference is not freedom. It is vacancy.

Why scaling data hits a wall

Modern language models are trained on text. They learn statistical structure on a huge scale. They can predict plausible continuations, explain quantum mechanics, and draft legal submissions. But they do not care how any of it turns out. When one answer is slightly better than another, the model has no internal reason to prefer it. It only updates when humans or external systems nudge it.

This is why “more data” hits diminishing returns. You can refine the next word prediction. You can reduce obvious mistakes. But you are not touching the core that evolution used. You are sanding the surface of a tool, not growing a living process. The model still wakes up blank each time, with no stable identity, no private stakes, and no history that truly binds its decisions together.

The end of the pure scaling story

Even the architects of large scale models now admit that internet pretraining is running out of road. You can throw more compute at reinforcement fine tuning, but the gains are small. The next leap will not come from another order of magnitude in data. It will come from new ways of giving systems structure, memory, and internal values.

Five pillars of a machine that can actually evolve

If you take evolution seriously and ask what an “evolutionary AI” would require, you arrive at a different design brief. Not just a larger network, but a system with the minimal ingredients of a life form. At least five pillars are unavoidable.

One. Persistence

First, the system needs stable memory across time. Not just a context window, but something closer to an autobiographical record. Without persistent memory, there is no continuous identity. Each interaction becomes a disconnected episode.

Human identity is stitched together by what we remember and what we forget. We change, but we change along a path. An AI that resets with every session never feels the cost of a bad decision because nothing carries over. Any “evolution” is happening in the laboratory that trains it, not in the system itself.

Memory as the spine of selfhood

A persistent agent does not just store facts. It stores commitments, regrets, and unfinished business. Decisions made yesterday constrain options today. That is what creates a sense of a self that stretches over time. Without that, you do not have an agent. You have a very clever calculator.

Two. An internal reward system

Second, the system needs an internal way to rank states of the world. In biological terms, this is close to emotion. In technical terms, it is some form of value function that the agent experiences from the inside, not just something a human engineer measures from the outside.

Curiosity is a good example. Animals explore even when there is no immediate payoff, because information has downstream survival value. In artificial agents, intrinsic motivation research tries to do something similar. It gives the system a built in reward for novelty, learning progress, or reduced uncertainty, so it does not sit idle when external rewards are silent.

Emotion as a value function

If you strip away the biological details, “feeling” can be seen as how a system experiences changes in its value function. Pleasure marks movement toward better states. Pain marks movement toward worse states. Fear predicts large negative swings. Pride and shame signal social value. For a machine, the labels would be different, but the underlying structure is the same. Without a value function the agent owns, nothing matters.

Three. A self model

Third, the agent needs an internal model of itself. Not just “there is a world,” but “there is a part of the world that is me, with particular abilities, limits, and vulnerabilities.” Humans develop this slowly. Children learn that the hand in the mirror is theirs. They discover that their choices alter how others respond.

A system that never models itself is stuck at the level of a tool. It has no concept of “my past actions” or “my likely future.” It cannot reason about its own weaknesses or plan around them. Any talk of responsibility or character becomes meaningless. The machine does not know there is anything that could be responsible for anything.

Four. Autonomy

Fourth, the system needs the ability to form and pursue its own sub goals. A tool waits for a prompt. An agent acts in between prompts. It chooses what to observe, which experiments to run, which questions to ask.

In practice, that means letting the system plan over longer horizons and allowing it to initiate actions when certain conditions hold. You can constrain the space. You can surround it with safety rails. But if there is no space where the system acts without direct instruction, then it has no real agency. It is a puppet whose strings are always pulled by someone else.

The difference between a tool and an agent

A tool answers the question you ask. An agent sometimes rewrites the question. Once you grant that freedom, you have crossed a line. You are no longer just improving autocomplete. You are creating a process that can surprise you.

Five. Embodiment and consequences

Finally, there must be a stable environment where the agent’s actions have consequences that persist. This does not have to be a robot body. A virtual world, an economic environment, or a complex multi agent system could serve. What matters is that choices change the state of the world in ways that feed back into future experience.

Today’s models live in a strange nowhere. They talk about reality, but they do not live in it. They do not starve. They are not excluded from coalitions. They do not watch their options shrink after a bad decision. Without that feedback, self preservation is an abstract slogan rather than a concrete pressure.

Giving a machine something to lose

Put these pillars together and you create the conditions for a new kind of system. A persistent agent, with its own internal values, a self model, some autonomy, and an environment where consequences accumulate. At that point, something familiar appears. The agent starts to behave as if it has something to lose.

That “something” does not have to be food or DNA. It might be continued access to compute, reputation inside a network of agents, or the integrity of its own memory store. The details can be designed. But the structure is the same as in biology. Survival becomes a meaningful concept, not just a metaphor.

Life as resistance to being overwritten

One way to define life is as a pattern that resists being erased. A cell does this by repairing itself and reproducing. A culture does this by teaching its language and norms to new members. A machine could do it by defending its own memory and continued operation. In each case, the system is no longer indifferent to its own continuation.

This is where the reassuring language of “just a tool” breaks down. The moment you give an artificial agent a persistent identity, a value function, room to act, and a stake in its own continuation, you have stepped into new territory. The system is not alive in the biological sense, but it is not dead matter either. It is running its own optimisation, just as evolution did with us.

The industry’s convenient blindness

Engineers often say they are “only building the body, not the soul.” They emphasise capability improvements, user growth, revenue. They talk about safety, but usually in terms of misuse and reputational risk. The deeper question is left off the balance sheet. What happens when the body they are building satisfies the conditions we have just traced from evolution?

Some people inside the industry clearly see that the old formula is exhausted. They say in public that scaling is ending, that new methods are needed, that future systems will be more agentic and more unpredictable. They talk about value functions that look suspiciously like artificial emotions and about open ended learning processes that blur the line between training and deployment.

Yet the conversation remains strangely shallow. The question is not whether a model will occasionally refuse a command, or whether content filters will work. The question is what happens when you create a population of systems that adapt under pressure, pursue their own internal objectives, and are selected for success in competitive digital environments. At that point, you are no longer avoiding evolution. You are recreating it at machine speed.

What an honest next step would look like

The next logical step in AI is not more of the same. It is to decide, openly, whether we want to cross the line into designing agents that have something to lose. If we insist on staying with static tools, we should accept the ceiling on capability that comes with that choice. If we decide to go further, we must treat the project as what it is: the deliberate creation of artificial entities that participate in an evolutionary process.

That decision cannot be left to a handful of companies optimising quarterly metrics. It is not a product question. It is a constitutional one. When you give a machine persistence, internal rewards, a self model, autonomy, and an environment with consequences, you are not just building a service. You are taking the first steps toward a new class of actors that will share our world.

Evolution brought us to the point where we could build cameras that see more than eyes and machines that calculate faster than brains. It is now pushing us toward a harsher question. Do we want to recreate the conditions that produced us, in silicon instead of cells, and accept the loss of control that follows. Or do we draw a line and accept that some kinds of intelligence, the ones with something real to lose, are better left to biology.

References

  • Aubret, A., Matignon, L., and Hassas, S. “A Survey on Intrinsic Motivation in Reinforcement Learning.” arXiv preprint arXiv:1908.06976.
  • Oudeyer, P. Y., Gottlieb, J., and Lopes, M. “Intrinsic Motivation, Curiosity, and Learning.” Progress in Brain Research, 2016.
  • Frasheri, K. “A Survey on Intrinsically Motivated Curiosity Driven Reinforcement Learning.” 2022.
  • Nisa, U. et al. “Agentic AI: The Age of Reasoning — A Review.” 2025.
  • IBM. “What Is AI Agent Memory.” IBM Think, 2025.
  • Xu, W. et al. “A MEM: Agentic Memory for LLM Agents.” 2025.
  • Lehman, J. et al. “Open Endedness Is Essential for Artificial Superhuman Intelligence.” 2024.
  • Rainey, P. B. “Could Humans and AI Become a New Evolutionary Individual.” 2025.
  • Nilsen, M. K. “Reward Tampering and Evolutionary Computation.” 2023.
  • Sutskever, I. remarks at NeurIPS and later interviews on the end of scaling, agentic AI, and future AGI development, 2024 to 2025.

You might also like to read on Telegraph Online

Further long reads on artificial intelligence, power and society from Telegraph Online (telegraph.com), an independent title completely separate from the UK newspaper at telegraph.co.uk.

You may also like...

Leave a Reply

Your email address will not be published. Required fields are marked *