New research will help cyber-physical systems understand human activities
ARC Lab students with the Baxter robot salad-making demonstration at the DARPA Forum in St. Louis.
UMIACS Associate Research Scientist Cornelia Fermüller is the principal investigator and Professor John Baras (ECE/ISR) and ISR-affiliated Professor Yiannis Aloimonos(CS/UMIACS) are the co-PIs on a three-year, $800K NSF Cyber-Physical Systems grant, “MONA LISA - Monitoring and Assisting with Actions.” The research will be conducted in the Autonomy Robotics Cognition (ARC) Laboratory.
Cyber-physical systems of the near future will collaborate with humans. Such cognitive systems will need to understand what the humans are doing, interpret human action in real-time and predict humans' immediate intention in complex, noisy and cluttered environments.
This research will develop a new three-layer architecture, motivated by biological perception and control, for cognitive cyber-physical systems that can understand complex human activities, focusing specifically on manipulation activities.
At the bottom layer are vision processes that detect, recognize and track humans, their body parts, objects, tools and object geometry. The middle layer contains symbolic models of human activity. It assembles through a grammatical description the recognized signal components of the previous layer into a representation of the ongoing activity. Cognitive control is at the top layer, deciding which parts of the scene will be processed next and which algorithms will be applied where. It modulates the vision processes by gathering additional knowledge when needed, and directs attention by controlling the active vision system to direct its sensors to specific places.
Thus, the bottom layer is perception, the middle layer is cognition, and the top layer is control. All layers have access to a knowledge base, built in offline processes, which contains semantics about actions.
The feasibility of the approach will be demonstrated through developing a smart manufacturing system, called MONA LISA, which assists humans in assembly tasks. This system will monitor humans as they perform assembly tasks, recognize assembly actions, and determine whether they are correct. It will communicate to the human possible errors and suggest ways to proceed.
The system will have advanced visual sensing and perception; action understanding grounded in robotics and human studies; semantic and procedural-like memory and reasoning, and a control module linking high-level reasoning and low-level perception for real time, reactive and proactive engagement with the human assembler.
The research will bring new tools and methodology to sensor networks and robotics and is applicable to a large variety of sectors and applications, including smart manufacturing. Being able to analyze human behavior using vision sensors will have an impact on many sectors, from healthcare and advanced driver assistance to human-robot collaboration.
ARC Lab holds inaugural open house
Alumnus Fumin Zhang promoted to full professor at Georgia Tech
Alumna Naomi Leonard wins Hendrik W. Bode Lecture Prize
Aloimonos gives keynote address at Bremen University Talks
Alum Xiaobo Tan named 'Foundation Professor' at Michigan State
Robots learn kitchen skills by watching YouTube videos
Martins, Gupta, Aloimonos speak at 'Fostering Excellence in Robotics'
Derek Paley leads 'Robots and Roaches' workshop
Sarah Bergbreiter engineers submillimeter-sized robotic systems
Thank you, Lockheed Martin!
September 16, 2015