There’s a mathematical concept called the ‘kissing number.’ Somewhat disappointingly, it’s got nothing to do with actual kissing; It enumerates how many spheres can touch (or ‘kiss’) a single sphere of equal size without crossing it. In one dimension, the kissing number is two. In two dimensions it’s 6 (think the New York Times’spelling bee puzzle configuration). As the number of dimensions grows, the answer becomes less obvious: For most dimensionalities over 4, only upper and lower bounds on the kissing number are known. Now, an AI agent developed by Google DeepMind called AlphaEvolve has made its contribution to the problem, increasing the lower bound on the kissing number in 11 dimensions from 592 to 593.
This may seem like an incremental improvement on the problem, especially given that the upper bound on the kissing number in 11 dimensions is 868, so the unknown range is still quite large. But it represents a novel mathematical discovery by an AI agent, and challenges the idea that large language models are not capable of original scientific contributions.

And this is just one example of what AlphaEvolve has accomplished. “We applied AlphaEvolve across a range of open problems in research mathematics, and we deliberately picked problems from different parts of math: analysis, combinatorics, geometry,” says Matej Baloga research scientist at DeepMind that worked on the project. They found that for 75 percent of the problems, the AI model replicated the already known optimal solution. In 20 percent of cases, it found a new optimum that surpassed any known solution. “Every single such case is a new discovery,” Balog says. (In the other 5 percent of cases, the AI converged on a solution that was worse than the known optimal one.)
The model also developed a new algorithm for matrix multiplication—the operation that underlies much of machine learning. A previous version of DeepMind’s AI model, called AlphaTensor, had already beat the previous best known algorithm, discovered in 1969, for multiplying 4 by 4 matrices. AlphaEvolve found a more general version of that improved algorithm.
DeepMind’s AlphaEvolve made improvements to several practical problems at Google. Google DeepMind
In addition to abstract math, the team also applied their model to practical problems Google as a company faces every day. The AI was also used to optimize data center orchestration to gain 1 percent improvement, to optimize the design of the next Google tensor processing unit, and to discover an improvement to a kernel used in Gemini training leading to a 1 percent reduction in training time.
“It’s very surprising that you can do so many different things with a single system,” says Alexander Novikova senior research scientist at DeepMind who also worked on AlphaEvolve.
How AlphaEvolve Works
AlphaEvolve is able to be so general because it can be applied to almost any problem that can be expressed as code, and which can be checked by another piece of code. The user supplies an initial stab at the problem—a program that solves the problem at hand, however suboptimally—and a verifier program that checks how well a piece of code meets the required criteria.
Then, a large language model, in this case Gemini, comes up with other candidate programs to solve the same problem, and each one is tested by the verifier. From there, AlphaEvolve uses a genetic algorithm such that the ‘fittest’ of the proposed solutions survive and evolve to the next generation. This process repeats until the solutions stop improving.
AlphaEvolve uses an ensemble of Gemini large language models (LLMs) in conjunction with an evaluation code, all orchestrated by a genetic algorithm to optimize a piece of code. Google DeepMind
“Large language models came around, and we started asking ourselves, is it the case that they are only going to add what’s in the training data, or can we actually use them to discover something completely new, new algorithms or new knowledge?” Balog says. This research, Balog claims, shows that “if you use the large language models in the right way, then you can, in a very precise sense, get something that’s provably new and provably correct in the form of an algorithm.”
AlphaEvolve comes from a long lineage of DeepMind’s models, going back to AlphaZero, which stunned the world by learning to play chess, Go, and other games better than any human player without using any human knowledge—just by playing the game and using reinforcement learning to master it. Another math-solving AI based on reinforcement learning, AlphaProof, performed at the silver-medalist level on the 2024 International Math Olympiad.
For AlphaEvolve, however, the team broke from the reinforcement learning tradition in favor of the genetic algorithm. “The system is much simpler,” Balog says. “And that actually has consequences, that it’s much easier to set up on a wide range of problems.”
The (Totally Not Scary) Future
The team behind AlphaEvolve hopes to evolve their system in two ways.
First, they want to apply it to a broader range of problems, including those in the natural sciences. To pursue this goal, they are planning to open up an early access program for interested academics to use AlphaEvolve in their research. It may be harder to adapt the system to the natural sciences, as verification of proposed solutions may be less straightforward. But, Balog says, “we know that in the natural sciences, there are plenty of simulators for different types of problems, and then those can be used within AlphaEvolve as well. And we are, in the future, very much interested in broadening the scope in this direction.”
Second, they want to improve the system itself, perhaps by coupling it with another DeepMind project: the AI co-scientist. This AI also uses an LLM and a genetic algorithm, but it focuses on hypothesis generation in natural language. “They develop these higher-level ideas and hypotheses,” Balog says. “Incorporating this component into AlphaEvolve-like systems, I believe, will allow us to go to higher levels of abstraction.”
These prospects are exciting, but for some they may also sound menacing—for example, AlphaEvolve’s optimization of Gemini training may be seen as the beginning of recursively self-improving AI, which some worry would lead to a runaway intelligence explosion referred to as the singularity. The DeepMind team maintains that that is not their goal, of course. “We are excited to contribute to advancing AI that benefits humanity,” Novikov says.
From Your Site Articles
Related Articles Around the Web
GIPHY App Key not set. Please check settings