humanvsmachinecropped100633046orig
Software & Web Development

'Ex Machina,' here we come: A new algorithm helps computers learn the way we do

Machine learning is all about getting computers to "understand" new concepts, but it's still a pretty inefficient process, often requiring hundreds of examples for training. That may soon change, however, thanks to new research published on Friday.

Aiming to shorten the learning process and make it more like the way humans acquire and apply new knowledge based on just a few examples, a team of researchers has developed what they call a Bayesian Program Learning framework and then used it to teach computers to identify and reproduce handwritten characters based on just a single example.

Whereas standard pattern-recognition algorithms represent concepts as configurations of pixels or collections of features, the BPL approach learns by “explaining” the data provided to the algorithm -- in this case, the sample character. Concepts are represented as probabilistic computer programs and the algorithm essentially programs itself by constructing code to produce the letter it sees. It can also capture variations in the way different people draw a given letter.

The model also “learns to learn” by using knowledge from previous concepts to speed learning on new ones, so it can use knowledge of the Latin alphabet to learn letters in the Greek alphabet more quickly, for example.

Most compelling of all is that the algorithm allowed computers to pass a sort of "visual Turing test." Specifically, the researchers asked both humans and computers to reproduce a series of handwritten characters after being shown just a single example of each; in some cases, subjects were asked to create entirely new characters in the style of those originally shown. Bottom line: human judges couldn't tell the results apart.

The researchers have applied their model to more than 1,600 types of handwritten characters in 50 writing systems, including Sanskrit, Tibetan, Gujarati and Glagolitic. They even tried it on invented characters such as those from the television series "Futurama."

A paper describing the research was published Friday in the journal Science. Its authors were Brenden Lake, a Moore-Sloan Data Science Fellow at New York University; Ruslan Salakhutdinov, an assistant professor of Computer Science at the University of Toronto; and Joshua Tenenbaum, a professor at MIT in the Department of Brain and Cognitive Sciences and the Center for Brains, Minds and Machines.

“It has been very difficult to build machines that require as little data as humans when learning a new concept,” said Salakhutdinov. “Replicating these abilities is an exciting area of research connecting machine learning, statistics, computer vision, and cognitive science.”

IDG Insider

PREVIOUS ARTICLE

« Yes, you can overclock cheap Intel Skylake chips

NEXT ARTICLE

Bitcoin miner KnC is planning another four-week datacenter build-out »
author_image
IDG Connect

IDG Connect tackles the tech stories that matter to you

  • Mail

Recommended for You

How to (really) evaluate a developer's skillset

Adrian Bridgwater’s deconstruction & analysis of enterprise software

Unicorns are running free in the UK but Brexit poses a tough challenge

Trevor Clawson on the outlook for UK Tech startups

Cloudistics aims to trump Nutanix with 'superconvergence' play

Martin Veitch's inside track on today’s tech trends

Poll

Is your organization fully GDPR compliant?