Our world is ruled by 1s and 0s.
Most of us rarely think about it, but when we turn on our
smartphones and PCs, we’re giving ourselves over to machines that reduce every
single task to a series of 1s and 0s. That’s what digital means.
But
according to Doug Burger, a researcher with Microsoft’s Extreme
Computing Group, this may be coming to an end. Burger thinks we
could be entering a new era where we don’t need digital accuracy. To hear him
tell it, the age of really big data may well be an age of slightly
less-accurate computing. We could drop the digital straightjacket and write
software that’s comfortable working on hardware that sometimes makes errors.
For about half a century now, companies like Intel have made
their microprocessors faster and faster by adding more transistors — and,
lately, more and more processor “cores” that can operate in parallel. But these
regular performance boosts seem to be coming to an end. It’s inevitable,
really. Chip parts are getting so small, they can’t be shrunk much more.
Intel’s current state-of-the art chipmaking process will soon
shrink to 14 nanometers, aka 14 billionths of a meter. When transistors get
that small, it becomes devilishly hard to keep them operating in the precise
on-or-off states required for digital computing. That’s one of the reasons why
today’s chips burn so hot.
Burger calls it the digital tax. And over the next decade this
tax is going to become too big for computer makers to keep paying it.
“Transistors are getting leakier. They’re getting noisier,” he said on Monday,
speaking during a webcast of an event at Microsoft’s headquarters. “They fail
more often.”
“The challenge is that at some point along this road, as you get
down to single atoms, that tax becomes too high. You’re not going to be able to
build a stable digital transistor out of a single atom.”
But if our future performance gains aren’t going to come from
smaller transistors, how do we improve things? That’s right. We go analog.
“I think there’s an opportunity to break this digital tyranny of
this abstraction of 1s and 0s that’s served us so well for 50 for 60 years, and
that’s to embrace approximation,” he said. “Some people call it analog.”
Burger
has been working with Luis Ceze, an associate professor at the University of
Washington to create a brand new way of programming. Instead of following binary
instructions, they break up the code. Some of it — the part of your banking app
that sends retrieves your account balance from the bank, for example — has no
tolerance for errors. Other parts — the app’s check scanning software — can
handle some errors. Ceze and Burger’s programs watch how applications work and
then build neural network models that they run on special neural processing
accelerators, called NPUs. “We’re using a neural network to
approximate things that are supposed to run in a regular processor,” Ceze said
in an interview earlier this year. “What we want to do is use neural networks
for your browser, for your games, for all sorts of things.”
The researchers aim to build compete systems — processors,
storage, and software — that use this approximate computing model. They think
they’ll be able to run them at far lower voltages than conventional systems,
which will save money on power and cooling. They’ve built their first NPU’s
using programmable chips, but they’re now crafting NPUs out of analog circuits,
which will be faster and use much less power than their digital equivalents.
“Going analog is a huge efficiency gain and much more approximate too,” Ceze
said.
This approach makes some slight mistakes, so it doesn’t work for
all programming models. You wouldn’t want to build a calculator this way. But
for many types programs — image processing software, for example — it’s good
enough.
Image recognition, bioinformatics, data mining, large scale
machine learning, and speech recognition could all work with analog computing,
according to Burger. “We’re doing an enormous number of things that intersect
with the analog world in fundamental ways.”
Burger
and Ceze are not the only ones peering into an analog future. Last year, the
Defense Advanced Research Projects Agency (DARPA) kicked off a
program called UPSIDE, short for Unconventional Processing of
Signals for Intelligent Data Exploitation, seeking solve these same problems.
It will be a long time — maybe 10 to 15 years — before the
systems that Burger describes have a chance of real-world use. But this may
well be the way that the next generation of computers get their juice. “We have
no idea how far we can push this,” Burger said. “But once you’re willing to
take a little bit of error and you break again this digital tyranny, you can
start using again these devices that are noisy — and you don’t have to pay that
enormous tax to guarantee a 1 or a 0.”
via wired