Feeds

Google Brain king slashes cost of AI gear

Anything CPU can do GPU can do better

  • alert
  • submit to reddit

Providing a secure and efficient Helpdesk

The architect behind some of Google's mammoth machine-learning systems has figured out a way to dramatically reduce the cost of the infrastructure needed for weak artificial intelligence programs, making it easier for cash-strapped companies and institutions to perform research into this crucial area of technology.

In a paper published this week entitled Deep learning with COTS HPC system, Andrew Ng, the director for Stanford University's Artificial Intelligence Lab and former visiting scholar at Google's skunkwork research group "Google X", and his colleagues detail how they have created a much cheaper system for running deep-learning algorithms at scale.

Machine learning is an incredibly powerful tool that lets researchers construct models that can learn from data and therefore get better over time. It underpins many advanced technologies such as speech recognition or image categorization, and is one of Google Research's main areas of investigation due to the huge promise it holds for the development of sophisticated Chocolate Factory AI.

The secret for how Ng's Commodity Off-The-Shelf High Performance Computing (COTS HPC) system is able to cost much less than Google's system is its use of GPUs rather than relying mostly on CPUs, the researchers write.

"Our system is able to train 1 billion parameter networks on just 3 machines in a couple of days, and we show that it can scale to networks with over 11 billion parameters using just 16 machines," the researchers write. "As this infrastructure is much more easily marshaled by others, the approach enables much wider-spread research with extremely large neural networks."

The COTS HPC approach appears to supersede the system Andrew Ng played with when he was at Google, and consumes a fraction of the resources. The Chocolate Factory's "DistBelief" infrastructure used 16,000 CPU cores spread across 1,000 machines to train a machine-learning model with unlabeled images culled from YouTube into recognising high-level concepts such as cats. Google's gear cost around $1m, whereas COTS HPC costs around $20,000, according to Wired.

Using this cheap rig, the researchers trained a neural network with more than 11 billion parameters – 6.5 times larger than the model in Google's previous research – and accomplished this in only three days.

They were able to do this by building a machine-learning cluster that consisted of 16 servers, each with two quad-core CPUs and four Nvidia GTX680 GPUs, linked together by Infiniband.

The researchers paired this system with software written in C++ and built on top of the MVAPICH2 MPI implementation. By using MVAPICH2 they were able to make pointers to data in a GPU's memory be provided as an argument to an Message Passing Interface (MPI) call, letting the whole system pass information among its various GPUs via Infiniband without increasing latency. The GPU-specific code was written in Nvidia's CUDA language. Work is then divided up among separate GPUs according to a partionining scheme.

Google AI model

What the COPS HPC system's trainer program 'thinks' a face, upper body, and cat look like

To test their system, the researchers built a training data set out of 10 million YouTube video thumbnails which were scaled down to 200-by-200-pixel color images. They then fed these images into the neural network on the COTS HPC. The system developed an ability to evaluate various objects and classify them, and for the 1.8 billion parameter set demonstrated a best guess of 88.2 per cent success rate for identitying human faces, 80.2 per cent for upper bodies, and 73 per cent for cats.

It was less effective on the 11 billion parameter set (86.5 per cent, 74.5 per cent, and 69.4 per cent, respectively) but the academics hope to iron these problems out with further tuning.

"With such large networks now relatively straight-forward to train, we hope that wider adoption of this type of training machinery in deep learning will help spur rapid progress in identifying how best to make use of these expansions in scale," they write.

Though machine learning is still a very challenging area to develop tech for, the publication of papers such as COTS HPC could make it easier for organizations and universities to come to grips with the tech – and in an age in which few companies can afford to run supercomputer-grade infrastructure to fiddle with it, that's a wonderful thing. ®

Beginner's guide to SSL certificates

More from The Register

next story
ONE MILLION people already running Windows 10
A third of them are doing it in VMs, but early feedback focuses on frippery
Netscape Navigator - the browser that started it all - turns 20
It was 20 years ago today, Marc Andreeesen taught the band to play
Sign off my IT project or I’ll PHONE your MUM
Honestly, it’s a piece of piss
Sway: Microsoft's new Office app doesn't have an Undo function
Content aggregation, meet the workplace ... oh
Do Moan! MONSTER 6-day EMAIL OUTAGE hits Domain Monster
Customers freaked out by frightful service
Ploppr: The #VultureTRENDING App of the Now
This organic crowd sourced viro- social fertiliser just got REAL
Return of the Jedi – Apache reclaims web server crown
.london, .hamburg and .公司 - that's .com in Chinese - storm the web server charts
NetWare sales revive in China thanks to that man Snowden
If it ain't Microsoft, it's in fashion behind the Great Firewall
prev story

Whitepapers

Forging a new future with identity relationship management
Learn about ForgeRock's next generation IRM platform and how it is designed to empower CEOS's and enterprises to engage with consumers.
Win a year’s supply of chocolate
There is no techie angle to this competition so we're not going to pretend there is, but everyone loves chocolate so who cares.
Why cloud backup?
Combining the latest advancements in disk-based backup with secure, integrated, cloud technologies offer organizations fast and assured recovery of their critical enterprise data.
High Performance for All
While HPC is not new, it has traditionally been seen as a specialist area – is it now geared up to meet more mainstream requirements?
Saudi Petroleum chooses Tegile storage solution
A storage solution that addresses company growth and performance for business-critical applications of caseware archive and search along with other key operational systems.