The latest technology and digital news on the web

Human-centric AI news and analysis

Here’s what happened when neural networks took on the Game of Life

The Game of Life is a grid-based apparatus that is very accepted in discussions about science, computation, and bogus intelligence. It is an absorbing idea that shows how very simple rules can yield very complicated results.

Despite its simplicity, however, the Game of Life charcoal a claiming to artificial neural networks, AI advisers at Swarthmore College and the Los Alamos National Laboratory have shown in a recent paper. Titled, “It’s Hard for Neural Networks To Learn the Game of Life,” their analysis investigates how neural networks analyze the Game of Life and why they often miss award the right solution.

Their allegation highlight some of the key issues with deep acquirements models and give some absorbing hints at what could be the next administration of analysis for the AI community.

What is the Game of Life?

British mathematician John Conway invented the Game of Life in 1970. Basically, the Game of Life tracks the on or off state—the life—of a series of cells on a grid across timesteps. At each timestep, the afterward simple rules define which cells come to life or stay alive, and which cells die or stay dead:

  1. If a live cell has less than two live neighbors, it dies of underpopulation.
  2. If a live cell has more than three live neighbors, it dies of overpopulation.
  3. If a live cell has absolutely two or three live neighbors, it survives.
  4. If a dead cell has three live neighbors, it will come to life.

Based on these four simple rules, you can adjust the antecedent state of your grid to create absorbing stable, oscillating, and gliding patterns.

For instance, this is what’s called the glider gun.

game of life glider gun

You can also use the Game of Life to create very circuitous pattern, such as this one.

game of life circuitous pattern

Interestingly, no matter how circuitous a grid becomes, you can adumbrate the state of each cell in the next timestep with the same rules.

With neural networks being very good anticipation machines, the advisers wanted to find out whether deep acquirements models could learn the basal rules of the Game of Life.

Artificial neural networks vs the Game of Life

There are a few affidavit the Game of Life is an absorbing agreement for neural networks. “We already know a solution,” Jacob Springer, a computer science apprentice at Swarthmore College and co-author of the paper, told . “We can write down by hand a neural arrangement that accouterments the Game of Life, and accordingly we can analyze the abstruse solutions to our hand-crafted one. This is not the case in.”

It is also very easy to adjust the adaptability of the botheration in the Game of Life by modifying the number of timesteps in the future the target deep acquirements model must predict.

Also, unlike domains such as computer vision or natural accent processing, if a neural arrangement has abstruse the rules of the Game of Life it will reach 100 percent accuracy. “There’s no ambiguity. If the arrangement fails even once, then it is has not accurately abstruse the rules,” Springer says.

In their work, the advisers first created a small convolutional neural network and manually tuned its ambit to be able to adumbrate the arrangement of changes in the Game of Life’s grid cells. This proved that there’s a basal neural arrangement that can represent the rule of the Game of Life.

Then, they tried to see if the same neural arrangement could reach optimal settings when accomplished from scratch. They initialized the ambit to random values and accomplished the neural arrangement on 1 actor about generated examples of the Game of Life. The only way the neural arrangement could reach 100 percent accurateness would be to assemble on the hand-crafted constant values. This would imply that the AI model had managed to parameterize the rules basal the Game of Life.

But in most cases the accomplished neural arrangement did not find the optimal solution, and the achievement of the arrangement decreased even added as the number of steps increased. The result of training the neural arrangement was abundantly afflicted by the chosen set training examples as well as the antecedent parameters.

Unfortunately, you never know what the antecedent weights of the neural arrangement should be. The most common convenance is to pick random values from a normal distribution, accordingly clearing on the right antecedent weights becomes a game of luck. As for the training dataset, in many cases, it isn’t clear which samples are the right ones, and in others, there’s not much of a choice.

“For many problems, you don’t have a lot of choice in dataset; you get the data that you can collect, so if there is a botheration with your dataset, you may have agitation training the neural network,” Springer says.

The achievement of larger neural networks

convolutional neural arrangement game of life
Left: A manually tuned convolutional neural arrangement can adumbrate outcomes in the Game of Life with absolute accuracy. Right: But in practice, when training the arrangement from scratch, you’ll need a much larger neural arrangement to obtain equal results

In machine learning, one of the accepted ways to advance the accurateness of a model that is underperforming is to access its complexity. And this abode worked with the Game of Life. As the advisers added more layers and ambit to the neural network, the after-effects bigger and the training action eventually yielded a band-aid that accomplished near-perfect accuracy.

But a larger neural arrangement also means an access in the cost of training and active the deep acquirements model.

On the one hand, this shows the adaptability of large neural networks. Although a huge deep acquirements model might not be the most optimal architectonics to abode your problem, it has a greater chance of award a good solution. But on the other, it proves that there is likely to be a abate deep acquirements model that can accommodate the same or better results—if you can find it.

These allegation are in line with “The Action Ticket Hypothesis,” presented at the ICLR 2019 appointment by AI advisers at MIT CSAIL. The antecedent appropriate that for each large neural network, there are abate sub-networks that can assemble on a band-aid if their ambit have been initialized on lucky, acceptable values, thus the “lottery ticket” nomenclature.


“The action ticket antecedent proposes that when training a convolutional neural network, small lucky subnetworks bound assemble on a solution,” the authors of the Game of Life paper write. “This suggests that rather than analytic abundantly through weight-space for an optimal solution, gradient-descent access may rely on lucky initializations of weights that happen to position a subnetwork close to a reasonable local minima to which the arrangement converges.”

What are the implications for AI research?

“While Conway’s Game of Life itself is a toy botheration and has few direct applications, the after-effects we report here have implications for agnate tasks in which a neural arrangement is accomplished to adumbrate an aftereffect which requires the arrangement to follow a set of local rules with assorted hidden steps,” the AI advisers write in their paper.

These allegation can apply to apparatus acquirements models used logic or math solvers, acclimate and fluid dynamics simulations, and analytic answer in accent or image processing.

“Given the adversity that we have found for small neural networks to learn the Game of Life, which can be bidding with almost simple allegorical rules, I would expect that most adult symbol abetment would be even more difficult for neural networks to learn, and would crave even larger neural networks,” Springer said. “Our result does not necessarily advance that neural networks cannot learn and assassinate allegorical rules to make decisions, however, it suggests that these types of systems may be very difficult to learn, abnormally as the complication of the botheration increases.”

The advisers added accept that their allegation apply to other fields of apparatus acquirements that do not necessarily rely on assured analytic rules, such as image and audio classification.

For the moment, we know that, in some cases, accretion the size and complication of our neural networks can solve the botheration of poorly assuming deep acquirements models. But we should also accede the negative impact of using larger neural networks as the go-to method to afflicted impasses in apparatus acquirements research. One aftereffect can be greater energy burning and carbon emissions caused from the compute assets appropriate to train large neural networks. On the other hand, it can result in the accumulating of larger training datasets instead of relying on award ideal administration strategies across abate datasets, which might not be achievable in domains where data is accountable to ethical considerations and aloofness laws. And finally, the accepted trend toward acknowledging overcomplete and very large deep acquirements models can consolidate AI power in large tech companies and make it harder for abate players to enter the deep acquirements analysis space.

“We hope that this paper will advance analysis into the limitations of neural networks so that we can better accept the flaws that necessitate overcomplete networks for learning. We hope that our result will drive development into better acquirements algorithms that do not face the drawbacks of gradient-based learning,” the authors of the paper write.

“I think the after-effects absolutely actuate analysis into bigger search algorithms, or for methods to advance the ability of large networks,” Springer said.

This commodity was originally appear by Ben Dickson on TechTalks, a advertisement that examines trends in technology, how they affect the way we live and do business, and the problems they solve. But we also altercate the evil side of technology, the darker implications of new tech and what we need to look out for. You can read the aboriginal commodity here.

Appear September 29, 2020 — 10:17 UTC

Hottest related news