Using deep learning and artificial intelligence (AI), scientists are bringing to life one of about a handful of these quadruped robots in the world.
Astro is unique because he is the only one of these robots with a head, 3D printed to resemble a Doberman pinscher, that contains a (computerized) brain.
What would you get if you combined Apple’s Siri and Amazon’s Alexa with Boston Dynamic’s quadraped robots?
You’d get “Astro,” the four-legged seeing and hearing intelligent robodog.
Using deep learning and artificial intelligence (AI), scientists from Florida Atlantic University’s are bringing to life one of about a handful of these quadraped robots in the world.
Astro is unique because he is the only one of these robots with a head, 3-D printed to resemble a Doberman pinscher, that contains a (computerized) brain.
Astro not only looks like a dog; he learns like one too.
That’s because he doesn’t operate based on preprogrammed robotic automation.
Instead, Astro is being trained using inputs to a deep neural network—a computerized simulation of a brain—so that he can learn from experience to perform human-like tasks, or on his case, “doggie-like” tasks, that benefit humanity.
Equipped with sensors, high-tech radar imaging, cameras and a directional microphone, this 100-pound super robot is still a “puppy-in-training.”
Just like a regular dog, he responds to commands such as “sit,” “stand” and “lie down.”
Eventually, he will be able to understand and respond to hand signals, detect different colors, comprehend many languages, coordinate his efforts with drones, distinguish human faces, and even recognize other dogs.
As an information scout, Astro’s key missions will include detecting guns, explosives and gun residue to assist police, the military, and security personnel.
This robodog’s talents won’t just end there, he also can be programmed to assist as a service dog for the visually impaired or to provide medical diagnostic monitoring.
The MPCR team also is training Astro to serve as a first responder for search and rescue missions such as hurricane reconnaissance as well as military maneuvers.
Designed to engage and react to the world around him in real-time, this intelligent machine will be able to navigate through rough terrains and respond to dangerous situations to keep humans and animals out of harm’s way.
Astro will be outfitted with more than a dozen sensors that will consume environmental input across multiple modalities including optical, sound, gas and even radar.
To process the sensory inputs and make autonomous behavioral decisions, a set of Nvidia Jetson TX2 graphics processing units are onboard with a combined four teraflops of computing power, which amounts to about four trillion computations a second.
This robodog will be able to rapidly see and search thousands of faces in a database, smell the air to detect foreign substances, and hear and respond to distress calls that fall outside a human’s audible hearing range.
FAU’s MPCR team will program Astro to have an extensive database of experiences that he can draw upon to help him make immediate decisions on the go.
The human brains behind Astro are a team of neuroscientists, IT experts, artists, biologists, psychologists, high school students and undergraduate and graduate students at FAU.
At the helm of this project are Elan Barenholtz, William Hahn and Pedram Nimreezi.
“Our Machine Perception and Cognitive Robotics laboratory team was sought out by Drone Data’s Astro Robotics group because of their extensive expertise in cognitive neuroscience, which includes behavioral, neurophysiological and embedded computational approaches to studying the brain,” said Ata Sarajedini, dean of FAU’s Charles E. Schmidt College of Science.
“Astro is inspired by the human brain and he has come to life through machine learning and artificial intelligence, which is proving to be an invaluable resource in helping to solve some of the world’s most complex problems.”
Written by Gisele Galoustian.