Recycling Robot Learns Through System of Touch
A key step in the process happens on fast-moving conveyor belts, where workers have to sort items into categories such as paper, plastic and glass. Such jobs are dull, dirty, and often unsafe, especially in facilities where workers also have to remove normal trash from the mix.
With that in mind, a team led by researchers at MIT’s Computer Science and Artificial Intelligence Lab (CSAIL) has developed a robotic system that can detect if an object is paper, metal or plastic.
The team’s “RoCycle” system includes a soft teflon hand that uses tactile sensors on its fingertips to detect an object’s size and stiffness. Compatible with any robotic arm, RoCycle was found to be 85% accurate at detecting materials when stationary, and 63% accurate on an actual simulated conveyor belt. Its most common error was identifying paper-covered metal tins as paper, which the team says would be improved by adding more sensors along the contact surface.
“Our robot’s sensorized skin provides haptic feedback that allows it to differentiate between a wide range of objects, from the rigid to the squishy,” said MIT professor Daniela Rus, senior author on a new paper about RoCycle that will be presented later this month at the IEEE International Conference on Soft Robotics in Seoul, South Korea. “Computer vision alone will not be able to solve the problem of giving machines human-like perception, so being able to use tactile input is of vital importance.”
A collaboration with Yale University, RoCycle directly demonstrates the limits of sight-based sorting: it can reliably distinguish between two identical-looking Starbucks cups made of paper and plastic that would give vision systems trouble.
Rus says that the project is part of her larger goal to reduce the back-end cost of recycling, in order to incentivize more cities and countries to create their own programs. Today recycling centers aren’t particularly automated: their main kinds of machinery include optical sorters that use different wavelength light to distinguish between plastics, magnetic sorters that separate out iron and steel products, and aluminum sorters that use eddy currents to remove non-magnetic metals.
This is a problem for one very big reason: just last month China raised its standards for the cleanliness of recycled goods it accepts from the US, meaning that most of our single-stream recycling is now sent to landfills.
“If a system like RoCycle could be deployed on a wide scale, we’d potentially be able to have the convenience of single-stream recycling with the lower contamination rates of multi-stream recycling,” said Ph.D. student Lillian Chin, lead author on the new paper.
It’s surprisingly hard to develop machines that can distinguish between paper, plastic and metal, which shows how impressive a feat it is for humans. When we pick up an object, we can immediately recognize many of its qualities even with our eyes closed, like whether it’s large and stiff or small and soft. By feeling the object and understanding how that relates to the softness of our own fingertips, we are able to learn how to handle a wide range of objects without dropping or breaking them.
This kind of intuition is tough to program into robots. Traditional rigid robot hands have to know an object’s exact location and size to be able to then calculate a precise motion path. Soft hands made of materials like rubber are much more flexible, but have a different problem: because they’re powered by fluidic forces, they have a balloon-like structure that can easily puncture quite easily.
How RoCycle works
Rus’ team used a motor-driven hand made of a relatively new material called “auxetics.” Most materials get narrower when pulled on, like a rubber band when you stretch it; auxetics, meanwhile, actually get wider. The MIT team took this concept and put a twist on it, quite literally: they created auxetics that, when cut, either twist to the left or right. Combining a “left-handed” and “right-handed” auxetic for each of the hand’s two large fingers makes them interlock and oppose each other’s rotation, enabling more dynamic movement. The team calls this “handed-shearing auxetics”, or HSA.
“In contrast to soft robots, whose fluid-driven approach requires air pumps and compressors, HSA combines twisting with extension, meaning that you’re able to use regular motors,” said Chin.
The team’s gripper first uses its “strain sensor” to estimate an object’s size, and then uses its two pressure sensors to measure the force needed to grasp an object. These metrics along with calibration data on the size and stiffnesses of objects of different material types is what gives the gripper a sense of what material the object is made of. Since the tactile sensors are also conductive, they can detect metal by how much it changes the electrical signal.
“In other words, we estimate the size and measure the pressure difference between the current closed hand and what a normal open hand should look like,” said Chin. “We use this pressure difference and size to classify the specific object based on information about different objects that we’ve already measured.”
RoCycle builds on a set of sensors that the team developed for a different paper to estimate an object’s size and stiffness. Those sensors could detect the radius of an object to within 30 percent accuracy, and tell the difference between “hard” and “soft” objects with 78 percent accuracy. The team’s hand is also almost completely puncture-resistant: it was able to be scraped by a sharp lid and punctured by a needle more than twenty times, with minimal structural damage.
As a next step, the researchers plan to build out the system so that it can combine tactile data with actual video data from a robot’s cameras. This would allow them to further improve its accuracy and potentially allow for even more nuanced differentiation between different kinds of materials.
Chin and Rus co-wrote the RoCycle paper alongside MIT postdoctoral associate Jeffrey Lipton, as well as PhD student Michelle Yuen and professor Rebecca Kramer-Bottiglio of Yale University.
This project was supported in part by Amazon, JD, the Toyota Research Institute and the National Science Foundation. The complete paper is available here.