Facebook uses artifical intelligence to analyze satellite images of the Earth and determine where people are living. Then it can build drones and satellites that can deliver Internet access to all those people in the most efficient way. Facebook A group of…
A group of Facebook engineers has spent the last two years building satellites, drones, and lasers that can bring Internet access to all those people on Earth who don’t already have it. But none of this will work unless the company can first figure out where those people are. And that requires a little artificial intelligence.
In the spring of 2014, Facebook launched its Connectivity Lab. The idea was to build all sorts of new technologies that could more efficiently stretch the Internet to the rest of the world—and thus stretch Facebook to the rest of the world. But this is no simple thing. Building a flying Internet drone—a contraption that can circle the stratosphere and beam wireless signals down to Earth—is an enormous undertaking, in terms of time, technology, and money.
Given all that effort and expense, it doesn’t really make sense for drones to beam signals into areas that don’t include real live people. You might think it’s easy to figure out where the people are. But the Earth is a mighty big place. “We realized we couldn’t answer that question—and it’s a very basic question,” says Yael Maguire, who oversees the Facebook Connectivity Lab. He describes it as a “needle-in-the-haystack problem.” Ninety-nine percent of the Earth does not include human life.
So Facebook turned to artificial intelligence. Maguire and his team used what’s called deep learning to create a global map that shows how their new Internet tech can most efficiently reach the world’s population. “We wanted to build a map for what the best technologies would be,” Maguire says.
Sending A Signal
Drawing on services provided by the company’s AI Lab, a Facebook engineer and optical physicist named Tobias Tiecke built a system that can automatically analyze satellite images of the Earth’s surface and determine where people are actually living. This insight, Maguire explains, is now guiding how the company builds those flying drones. In fact, he says, it has shown that the company’s original approach to drone-powered Internet access was all wrong.
Deep learning relies on neural networks—networks of hardware and software that approximate the web of neurons in the human brain. If you feed enough photos of a goat into a neural network, it can learn to identify a goat. If you feed enough spoken words into a neural net, it can learn to recognize the commands you speak into your smartphone. In much the same way, it can analyze satellite photos and learn to recognize where people are living.
For something like this to work, you need labeled data. In other words, humans must identify some good examples before they’re fed into a neural net. They must label a sample set of goat photos, for instance, as goat photos. The same goes for Facebook’s new project—but there’s a twist. The human trainers didn’t bother to label specific evidence of civilization. They didn’t label houses or cars or roads or farmland as a way of training the neural net. Given a photo, the humans simply noted whether or not the photo showed any sign that people were living there. “We just asked: ‘Is there a human artifact in this image or not?” says Maguire. “Binary question. Yes or no.”
Given this basic information for a relatively small number of photos—about 8,000 overhead images of India—the neural net could then identify evidence of human life across photos of about twenty other countries. In total, the system analyzed 15.6 billion images representing 21.6 million square kilometers of Earth. Using that sample of binary information—8,000 photos labeled as either containing a human artifact or not—the neural net could accurately identify other human settlements in other locations. “Just based on that information, the algorithm can then go out and find all sorts of human artifacts,” Maguire says. The error rate, he says, is less than 10 percent.
The simplicity of this approach may seem surprising. It was certainly surprising to Maguire. But deep neural nets work in sometimes surprising ways. And the aim is to build a classifier—a way of identifying photos or spoken words or other data—that’s as simple as possible. “When you’re building a classifier like Facebook’s, the more categories you ask the [neural] net to use, the harder the problem becomes, both in terms of computation and neural net tuning,” says Chris Nicholson, the CEO and founder of a deep learning startup called Skymind. “So for the sake of efficiency, you want to draw a line on how sophisticated of a classifier you need. Facebook chose to do something really simple, but if that fulfills their goal, then great.”
In they end, they can build a vast map of human artifacts with a resolution of about 5 meters. In other words, they pretty much known whether there’s evidence of human life in every 5-square-meter space across 20 countries. Then, by combining this information with census data, they can map the population density in those countries. And that’s a powerful thing.
Previously, Maguire and his team assumed that each drone would blanket an area with one large cone of wireless signal. But Facebook’s AI-enhanced map shows that this doesn’t make sense. “It changes, fundamentally, how our communications systems have to be developed,” Maguire says. He won’t yet say how the company intends to modify its drones. But the basic idea is to build wireless radios that allow the company to directly target human populations, rather than just cover an area with one giant cone of signal. “The cones should be much smaller. Ideally, they should be shaped to be just where the people are,” he says. “Actually, it’s not really a cone anymore.” So, now that it knows exactly where the people are, Facebook must build its satellites, drones, and lasers in a new way.
Click here to view full article