As you know, Wasabi Roll’s charter includes exploring the scientific and technological breakthroughs shaping the future. Hence, submitted for your approval, the Wafer-Scale Engine (WSE), that blew the doors off today’s supercomputer, Joule.
Smaller is Better?
The adage in semiconductor physics is that smaller is better. However, is it?
This is what Cerebras Systems, the creator of the
world's biggest computer.
The Cerebras Systems' wafer-scale engine is massive no matter how you cut the chip is 8
point inches to the side that contains 1.2 trillion transistors the next
biggest chip NVIDIA A100 GPU measures just one inch at a time and has only 54
billion transistors.
The
former is new, new largely untested, and to date, is one of a kind. The latter is well-loved, mass-produced, and
has taken over the AI at supercomputing world in the last decade.
Chips beyond AI
200X!!!!
One
of those labs, the national energy technology laboratory, is looking to see
what can be done beyond AI. In a recent study, researchers pitted the chip
which is housed in an all-in-one system about the size of a dorm room mini-fridge,
called the CS-1, against a supercomputer in a fluid dynamics simulation
Simulation of fluid movements is a common supercomputer implication that is useful for solving complex problems, such as, weather forecasting and aircraft wing design.
The
trial was described in a paper written by a team led by Cerebras Systems, Michael James,
and National Energy Technology Laboratory (NETL)’s Dirk VanEssendelft, Ph.D. They
presented the paper at the SC20 supercomputing conference in January of this
year.
The team said that the CS one had completed a combustion simulation in a power plant approximately 200X faster than the joule 2.0 supercomputer had taken on a similar task.
The
CS1 was faster than real-time, i.e., Cerebras Systems posted, that the CS1 can tell you
what's going to happen in the future faster than the laws of physics produced
the same result.
Researchers
said the performance of the CS1 could not be matched by several CPUs and GPUs.
CEO and co-founder Andrew Feldman said in VentureBeat, an online publishing platform that publishes information on
technology, the result would be true no matter how big a
supercomputer is. Specifically, at one
point scaling a supercomputer like Joule doesn't produce any better results in
this kind of problem. This is why Joule's
simulations peaked at 16,384 cores which is a fraction of its total of 86,400
cores it had available.
The
comparison between the two machines drives the point home.
Where joule:
Joule, 81st Fastest Supercomputer

·
Is the 81st
fastest supercomputer in the world
·
Packs dozens of
server racks
·
Consumes up to
450 kilowatts of power
·
Costs 10s of
millions of dollars to build.
CS1, by comparison:
·
Fits into a third
of the server rack
·
Consumes 20
kilowatts of power.
·
Costs only a few
$1,000,000
Admittedly,
the applications are niche today, but useful in the mammoth calculations is chaos
systems, and still a stunning result.
Cut Down the Commute?
So, how does this work? It's all about design, cut down the commute. Computer chips begin life on a large piece of Silicon called a wafer.
Multiple
chips are etched onto the same waiver and the wafers are then cut into
individual chips. While the WSE is also edged on a Silicon
wafer, the wafer remains intact as a single operating unit this waiver scale
chip contains nearly 400,000 processing cores.
Each
core is linked to its dedicated memory and its four adjacent chords putting
that many cores on a single chip and giving them their memory is why the WSE is
bigger. That's also why, in this case, it's better.
Most
large-scale computing tasks depend on massive processing. Researchers are distributing the task to
hundreds or thousands of chips. The chips need to work in concert so they're in
constant communication shuttling information back and forth. A similar process takes place within each
chip. As the information moves between
the processor cores that make the calculations and the shared results.
It's kind of like an old-fashioned company doing all its business on paper the company uses
couriers to send and collect documents from other branches and archives across the city the couriers know the best route through the city but the journeys take a minimum amount of time determined by the distance between the branches in the archives other careers on road. In short, the distance and the traffic slowdown.
Now imagine that same company building a brand-new gleaming skyscraper. Every branch is moved to a new building and every worker has a small filing cabinet in his office to store documents now any document they need can be stored and
retrieved in the time it takes to move across the office or down the hall to their neighbor's office the information exchange has all but vanished it's all in the same house.
The mega-chip of Cerebras Systems is a bit like this skyscraper the way it shuttles information with the help of
its specially tailored compiling software is far more efficient than a
traditional supercomputer that needs a ton of traditional chips to be networked.
Simulation of the world as it unfolds
It's
worth noting that the chip can only handle problems small enough to fit on the
wafer but such problems may have quite practical applications because of the
machine's ability to high Fidelity simulation in real-time the authors note.
For example, that the machine should theoretically be able to accurately simulate the flow of
air around a helicopter trying to land on a flight deck and semi-automate the process something that is not possible with traditional chips another opportunity they note would be to use simulation as input to train network also resides on a chip in an intriguing.
In
a related example Caltech, machine learning technique recently proved to be
1000 times faster in solving the same kind of partial differential equations
that play here to simulate fluid dynamics they also note that improvements in
the chip and others like it should they arrive will push back the limits of
what can be achieved Cerberus has already teased the release of its next-generation
chip which will have 2.6 trillion transistors 85,000 cores and more than double
the memory.
This
is the first to seriously pursue it they believe that they have solved the
problem in a way that is both useful and economical other new architectures are
also being pursued in the lab memristor-based neuromorphic chips for example
mimic the brain by applying processing and memory to individual transistor-like
components and of course, once and computers are in a separate link but they
tackle similar problems it could be that this seems just as likely computing
the splinter into a bizarre quilt of radical chips all stitched together to
make the most in each of them depending on the situation.
UPDATE
Cerebras’ New Monster AI Chip Adds
1.4 Trillion Transistors
Shift
to 7-nanometer process boosts the second-generation chip’s transistor count to
a mind boggling 2.6-trillion.
Almost
from the moment Cerebras Systems announced a computer based on the largest
single computer chip ever built, the Silicon Valley startup declared its
intentions to build an even heftier processor. Today, the company announced
that its next-gen chip, the Wafer Scale Engine 2 (WSE 2), will be available in
the 3rd quarter of this year. WSE 2 is just as big physically as its
predecessor, but it has enormously increased amounts of, well, everything. The
goal is to keep ahead of the ever-increasing size of neural networks used in
machine learning.
Share and enjoy…







No comments:
Post a Comment
Thanks for your input, your ideas, critiques, suggestions are always welcome...
- Wasabi Roll Staff