The latest technology and digital news on the web

Human-centric AI news and analysis

The key to making AI green is advance computing

We’ve corrective ourselves into addition corner with bogus intelligence. We’re assuredly starting to advance the account barrier but we’re butting up adjoin the limits of our our adeptness to responsibly meet our machines’ massive energy requirements.

At the accepted rate of growth, it appears we’ll have to turn Earth into Coruscant if we want to keep spending abysmal amounts of energy training systems such as GPT-3 .

The problem: Simply put, AI takes too much time and energy to train. A layperson might brainstorm a bunch of code on a laptop screen when they think about AI development, but the truth is that many of the systems we use today were accomplished on massive GPU networks, supercomputers, or both. We’re talking absurd amounts of power. And, worse, it takes a long time to train AI.

The reason AI is so good at the things it’s good at, such as image acceptance or accustomed accent processing, is because it basically just does the same thing over and over again, making tiny changes each time, until it gets things right. But we’re not talking about active a few simulations. It can take hundreds or even bags of hours to train up a robust AI system.

One expert estimated that GPT-3, a accustomed accent processing system created by OpenAI, would cost about $4.6 actor to train. But that assumes one-shot training. And very, very few able AI systems are accomplished in one fell swoop. Realistically, the total costs complex in accepting GPT-3 to spit out impressively articular gibberish are apparently in the hundreds-of-millions.

GPT-3 is among the high-end abusers, but there are endless AI systems out there sucking up hugely asymmetric amounts of energy when compared to accepted ciphering models.

The problem? If AI is the future, under the accepted power-sucking paradigm, the future won’t be green. And that may mean we simply won’t have a future.

The solution: Advance computing.

An all-embracing team of researchers, including scientists from the University of Vienna, MIT, Austria, and New York, recently published research demonstrating “quantum speed-up” in a hybrid bogus intelligence system.

In other words: they managed to accomplishment advance mechanics in order to allow AI to find more than one band-aid at the same time. This, of course, speeds up the training process.

Per the team’s paper:

The acute catechism for applied applications is how fast agents learn. Although assorted studies have made use of advance mechanics to speed up the agent’s controlling process, a abridgement in acquirements time has not yet been demonstrated.

Here we present a accretion acquirements agreement in which the acquirements action of an agent is sped up by using a advance advice approach with the environment. We added show that accumulation this book with classical advice enables the appraisal of this advance and allows optimal ascendancy of the acquirements progress.


This is the cool part. They ran 10,000 models through 165 abstracts to actuate how they functioned using classical AI and how they functioned when aggrandized with appropriate advance chips.

And by special, that is to say, you know how classical CPUs action via abetment of electricity? The advance chips the team used were nanophotonic, acceptation they use light instead of electricity.

The gist of the operation is that in accident where classical AI bogs down analytic very difficult problems (think: supercomputer problems), they found the hybrid-quantum system outperformed accepted models.

Interestingly, when presented with less difficult challenges, the advisers didn’t not beam any performance boost. Seems like you need to get it into fifth gear before you kick in the advance turbocharger.

There’s still a lot to be done before we can roll out the old “mission accomplished” banner. The team’s work wasn’t the band-aid we’re eventually aiming for, but more of a small-scale model of how it could work once we figure out how to apply their techniques to larger, real problems.

You can read the whole paper here on Nature.

H/t: Shelly Fan, Singularity Hub

Published March 17, 2021 — 19:41 UTC

Hottest related news

No articles found on this category.
No articles found on this category.