Articles

Google’s Learning Software Learns To Write Learning Software

Google lists just over 1,300 people on its research website, not all of whom specialize in AI. It has many thousands more software engineers. Google parent Alphabet has 27,169 employees engaged in research and development, according to its most recent annual financial filing.

Google declined to make anyone available to discuss AutoML. Researchers outside the company say the idea of automating some work of AI experts has become a research hotspot—and is needed as AI systems become more complex.

Much work in what is called metalearning or learning to learn, including Google’s, is aimed at speeding up the process of deploying artificial neural networks. That technique involves feeding data through networks of math operations loosely inspired by studies of neurons in the brain.

That may sound highly sophisticated, but a good part of getting neural networks to perform useful tricks like processing audio comes down to well-paid grunt work. Experts must use instinct and trial and error to discover the right architecture for a neural network. “A large part of that engineer’s job is essentially a very boring task, trying multiple configurations to see which ones work better,” says Roberto Calandra, a researcher at University of California Berkeley. The challenge is getting harder, he says, because researchers are building larger networks to tackle tougher problems.

Calandra began researching metalearning after spending two frustrating weeks trying to get a robot to learn to walk during his PhD studies in 2013. He tried an experimental technique to automatically tune its software, which was based on a machine learning technique less complex than a neural network. The recalcitrant machine walked within a day.

Generating a neural-network design from scratch is harder than tweaking the settings of one that already exists. But recent research results suggest it’s getting closer to becoming practical, says Mehryar Mohri, a professor at NYU.

Mohri is working on a system called AdaNet, in a collaboration that includes researchers at Google’s New York office. When given a collection of labeled data, it builds a neural network layer by layer, testing each addition to the design to ensure it improves performance. AdaNet has shown capable of generating neural networks that can accomplish a task as well as a standard, hand-built network that’s twice as large. That’s promising, says Mohri, because many companies are trying to cram more powerful AI software onto mobile devices with limited resources.

Making it easier to generate and deploy complex AI systems might come with drawbacks. Recent research has shown that it is all too easy to accidentally make systems with a biased view of the world, for example that “Mexican” is a bad word, or have a tendency to associate women with domestic chores. Mohri argues that reducing the tedious hand-tuning required to make use of neural networks could make it easier to detect and prevent such problems. “It’s going to make people’s hands more free to tackle other aspects of the problem,” he says.

If and when Google gets AutoML working well enough to be a practical tool for programmers, its effects could be felt beyond the company itself. Pichai hinted last week that he wanted to make the tool available outside of Google. “We want to democratize this,” he said, echoing lofty language used to promote AI services offered by his cloud computing unit.