Recent & Upcoming Talks

Yasemin Bekiroglu: Learning and multi-modal sensing for robotic grasping and manipulation

Robots are envisioned as capable machines who easily navigate and interact in a world built for humans. However, looking around us we see robots mainly confined to factories only performing repetitive tasks in environments built such as to circumvent their limitations. The central question of my research is how we can create robots that are capable of adapting such that they can co-inhabit our world. This means designing systems that are capable of functioning in unstructured environments that are continuously changing with unlimited combination of shapes, sizes, appearance, and positions of objects, able to understand, adapt and learn from humans, and importantly do so from small amounts of data. In specific my work focuses on grasping and manipulation, fundamental aspects to enable a robot to interact with humans in our environment, along with dexterity (e.g. to use objects/tools successfully) and high-level reasoning (e.g. to decide about which object/tool to use). Despite decades of research, robust autonomous grasping and manipulation approaching human skills remains an elusive goal. One main difficulty lies in dealing with the inevitable uncertainties in how a robot perceives the world. Our environment is dynamic, has a complex structure and sensory measurements are noisy and associated with a large degree of uncertainty which poses challenges to avoid failures. In my research I have developed methodologies that enable a robot to interact with natural objects and learn about object properties and relations between tasks and sensory streams. I have developed tools that allow a robot to use multiple streams of sensory data in a complementary fashion. In this talk I will specifically address how a robot can use vision and touch to address grasp related questions e.g. estimating unknown object properties such as shape, grasp stability estimation before manipulating objects that can be used to trigger plan corrections, and grasp adaptation/correction.

Yasemin Bekiroglu: Towards Robust and Goal-oriented Robotic Grasping and Manipulation

Robots are envisioned as capable machines who easily navigate and interact in a world built for humans. However, looking around us we see robots mainly confined to factories only performing repetitive tasks in environments built such as to circumvent their limitations. The central question of my research is how we can create robots that are capable of adapting such that they can co-inhabit our world. This means designing systems that are capable of functioning in unstructured environments that are continuously changing with unlimited combination of shapes, sizes, appearance, and positions of objects, able to understand, adapt and learn from humans, and importantly do so from small amounts of data. In specific my work focuses on grasping and manipulation, fundamental aspects to enable a robot to interact with humans in our environment, along with dexterity (e.g. to use objects/tools successfully) and high-level reasoning (e.g. to decide about which object/tool to use). Despite decades of research, robust autonomous grasping and manipulation remains an elusive goal. The difficulty lies in dealing with the inevitable uncertainties in how a robot perceives the world. Our environment is dynamic, has a complex structure and sensory measurements are noisy and associated with a large degree of uncertainty. In real-world settings, these issues can lead to grasp failures with serious consequences. In my research I have developed methodologies that enable a robot to interact with natural objects and learn about object properties and relations between tasks and sensory streams. I have developed tools that allow a robot to use multiple streams of sensory data in a complementary fashion. In this talk I will specifically address how a robot can use vision and touch to address grasp related questions e.g. estimating unknown object properties or grasp success before manipulating objects that can be used to trigger plan corrections.

Keshi He: Toward AI-powered Robots Served for IndustrialProductions and Healthcare

Recent advancement of AI technologies like deep learning or deep reinforcement learning greatly enhance robot’s capacities in many practical applications. In this talk, I will focus on how computer vision, deep learning and sensor fusion can be leveraged for better robotic perception and manipulation which enables the robots to better serve for industrial application and healthcare.I will start with our current work that exploits deep learning-based method to achieve the accurate estimation of object postures in the real scene. Bin picking robots have been explored for years.With the advancement in deep learning on point cloud, the object detection and pose estimation are intelligently achieved without explicit feature extraction. However, existing methods are only capable of estimating the bounding box but not the precise 3D position of an object. We propose anew learning-based object detection and pose estimation method for bin picking problem. We leverage PointNet and its following ideas for estimating 6 DoF parameters of objects. Our network estimates the translations and then rotations, and finally, we integrate the estimated postures by using mean-shift clustering. We trained our network using the data generated by physical simulations and stereo measurement system simulations. We demonstrate that our network can achieve the accurate estimation of object postures in the real scene without any transfer learning techniques. Furthermore, I will present our recent work on Robot-Enhanced therapy for children with autism spectrum disorder (ASD). Traditionally, diagnosis of ASD depend on doctor’s looking at a patient’s behavior and development, which is subjective and time consuming. We present a novel sensing-enhanced therapy system including three cameras and two Kinects to automatically, accurately and non-intrusively monitor ASD children in multiple scenarios. Involved technics are including but not limited to face expression, human motion analysis and gaze estimation.Furthermore, we propose a multimodal approach to detect ASD severity in participants undergoing treatment for ASD.Finally, I will present our earlier work on an ultrasound-based human-machine interface(HMI) for dexterous prosthetic control. Surface electromyography (EMG) is widely investigated in HMI by decoding movement intention to intuitively control intelligent devices, but could not offer satisfactory solutions for finger motion classification. We build a novel B mode ultrasound image-based HMI. Then, I present experiments on a finger motion task where this ultrasound-based HMI has a better average accuracy (95.88%) than the EMG-based HMI(90.14%) for finer motion recognition. Furthermore, I will discuss how to use machine learning and image process methods to improve the robustness of our proposed HMI in practice.

Rika Antonova: Adaptive Kernels and Priors for Data-efficient Bayesian Optimization

This talk will describe a framework for constructing simulation-based kernels for ultra data-efficient Bayesian optimization (BO). We consider challenging settings that allow only 10-20 hardware trials, with access to only mid/low-fidelity simulators. First, I will describe how we can construct an informed kernel by embedding the space of simulated trajectories into a lower-dimensional space of latent paths. Our sequential variational autoencoder handles large-scale learning from ample simulated data; its modular design ensures quick adaptation to close the sim-to-real gap. The approach does not require domain-specific knowledge. Hence, we are able to demonstrate on hardware that the same architecture works for different areas of robotics: locomotion and manipulation. For domains with severe sim-to-real mismatch, I will describe our variant of BO which ensures that discrepancies between simulation and reality do not hinder online adaption. Using task-oriented grasping as an example, I will demonstrate how this approach helps quick recovery in case of corrupted/degraded simulation. My longer-term research vision is to build priors from simulation without requiring a specific simulation scenario. So, I will conclude by providing the motivation for this direction, and will describe our initial work on ’timescale fidelity’ priors. Such priors could help transfer-aware models and simulators to automatically adjust their timestep/frequency or planning horizon.