Seminar
Imaging the World One Photon at a Time
Abstract: The heart of a camera and one of the pillars for computer vision is the digital photodetector, a device that forms images by collecting billions of photons traveling through the physical world and into the lens of a camera. While the photodetectors used by cellphones or professional DSLR cameras are designed to aggregate as [...]
Carnegie Mellon University
Lesson Learned from Two Decades of Robotics Development and Thoughts on Where We Go from Here
Abstract: In this talk, Herman Herman will offer various lessons learned from developing various robots for the last 2 decades at the National Robotics Engineering Center. He will also offer his perspective on the future of autonomous robots in various industries, including self-driving cars, material handling and consumer robotics. Bio: Dr. Herman Herman is the [...]
Factor Graphs for Robot Perception
Abstract: Factor graphs have become a popular tool for modeling robot perception problems. Not only can they model the bipartite relationship between sensor measurements and variables of interest for inference, but they have also been instrumental in devising novel inference algorithms that exploit the spatial and temporal structure inherent in these problems. I will overview [...]
Towards better methods of video generation
Abstract: Learning to generate future frames of a video sequence is a challenging research problem with great relevance to reinforcement learning, planning and robotics. Existing approaches either fail to capture the full distribution of outcomes, or yield blurry generations, or both. In this talk I will address two important aspects of video generations: (i) what [...]
Light-Sensitive Displays
Abstract: Nobel prize winner M. G. Lippmann described his dream of an ideal display as a “window into the world.” “While the current most perfect photographic print only shows one aspect of reality, reduced to a single image fixed in a plane, the direct view of reality offers, as we know, infinitely more variety.” Changing [...]
Acquiring and Transferring Generalizable Vision-based Robot Skills
Abstract: In recent years, there have been great advances in policy learning for goal-oriented agents. However, there are still major challenges brought by real-world constraints for teaching highly generalizable and versatile robot policies in a cost efficient and safe manner. In this talk, I will argue that instead of aiming to teach large motion repertoires [...]
Learning to localize and anonymize objects with indirect supervision
Abstract: Computer vision has made great strides for problems that can be learned with direct supervision, in which the goal can be precisely defined (e.g., drawing a box that tightly-fits an object). However, direct supervision is often not only costly, but also challenging to obtain when the goal is more ambiguous. In this talk, I [...]
What People See in a Robot: A New Look at Human-Like Appearance
Abstract: A long-standing question in HRI is what effects a robot’s human-like appearance has on various psychological responses. A substantial literature has demonstrated such effects on liking, trust, ascribed intelligence, and so on. Much of this work has relied on a construct of uni-dimensional low to high human-likeness. I introduce evidence for an alternative view according to which [...]
Safe Learning in Robotics
Abstract: A great deal of research in recent years has focused on robot learning. In many applications, guarantees that specifications are satisfied throughout the learning process are paramount. For the safety specification, we present a controller synthesis technique based on the computation of reachable sets, using optimal control and game theory. In the first part [...]
Bipolar Robotics – From the Arctic to the Antarctic with a stop for Fisheries in the middle latitudes.
Abstract: The Arctic, Antarctic and Greenland remain some of the least explored parts of the planet. This talk looks at efforts over the last decade to explore areas under-ice which have traditionally been difficult to access. The focus of the talk will be on the robots, the role of communications over low bandwidth acoustic links, [...]
Video Compression for Recognition & Video Recognition for Compression
Abstract: Training robust deep video representations has proven to be much more challenging than learning deep image representations. One reason is: videos are huge and highly redundant. The 'true' and interesting signal often drowns in too much irrelevant data. In the first part of the talk, I will show how to train a deep network [...]
Learning Robot Manipulation Skills through Experience and Generalization
Abstract: In the future, robots could be used to take care of the elderly, perform household chores, and assist in hazardous situations. However, such applications require robots to manipulate objects in unstructured and everyday environments. Hence, in order to perform a wide range of tasks, robots will need to learn manipulation skills that generalize between [...]
Signal to Symbol (via Skills)
Abstract: While recent years have seen dramatic progress in the development of affordable, general-purpose robot hardware, the capabilities of that hardware far exceed our ability to write software to adequately control. The key challenge here is one of abstraction: generally capable behavior requires high-level reasoning and planning, but perception and actuation must ultimately be performed [...]
Carnegie Mellon University
Autonomous drone cinematographer: Using artistic principles to create smooth, safe, occlusion-free trajectories for aerial filming
Abstract: Autonomous aerial cinematography has the potential to enable automatic capture of aesthetically pleasing videos without requiring human intervention, empowering individuals with the capability of high-end film studios. Current approaches either only handle off-line trajectory generation, or offer strategies that reason over short time horizons and simplistic representations for obstacles, which result in jerky movement and [...]
Visual SLAM with Semantic Scene understanding
Abstract: Simultaneous localization and mapping (SLAM) has been widely used in autonomous robots and virtual reality. It estimates the sensor motion and maps the environment at the same time. The classic sparse feature point map of visual SLAM is limited for many advanced tasks including robot navigation and interactions, which usually require a high-level understanding of [...]
Multi-Modal Geometric Learning for Grasping
Abstract: In this talk, we will describe methods to enable robots to grasp novel objects using multi-modal data and machine. The starting point is an architecture to enable robotic grasp planning via shape completion using a single occluded depth view of objects. Shape completion is accomplished through the use of a 3D CNN. The network [...]
Building a Force-Controlled Actuator (Company)
Abstract: In 2014, I was lucky enough to be one of 5 people to start HEBI Robotics, with the dream of eventually making the task of building custom robots as easy as building with Lego. A few years later we are now 10 people, and our first product, a series of modular force-controlled actuators, is [...]
Minimalist Visual Perception and Navigation for Consumer Drones
Abstract: Consumer drone developers often face the challenge of achieving safe autonomous navigation under very tight size, weight, power, and cost constraints. In this talk, I will present our recent results towards a minimalist, but complete perception and navigation solution utilizing only a low-cost monocular visual-inertial sensor suite. I will start with an introduction of [...]
Social Perception for Machines
Abstract: Despite decades of progress, machines remain intelligent tools rather than collaborative partners in individual human enterprise. A key reason is that machine perception of inter-personal communication is largely unsolved and a computationally accessible representation of such behavior remains elusive. In this talk, I will describe our research arc over the past decade at CMU [...]
Geometry Processing in The Wild
Abstract: Geometric data abounds, but our algorithms for geometry processing are failing. Whether from medical imagery, free-form architecture, self-driving cars, or 3D-printed parts, geometric data is often messy, riddled with "defects" that cause algorithms to crash or behave unpredictably. The traditional philosophy assumes geometry is given with 100% certainty and that algorithms can use whatever [...]
Self Driving Cars and AI: Transforming our cities and our lives
Abstract: Artificial intelligence and machine learning are critical to reaching full autonomy in self driving cars. I will present two autonomy systems along with the use of machine learning in each of them. I will summarize recent progress in commercializing these systems and make some observations about the potential impact of these systems in our [...]
Robotic Morphing Matter
Abstract: Morphing matter harnesses the programmability in material structures and compositions to achieve transformative behaviors and integrates sensing, actuation, and computation to create adaptive and responsive material systems. These material systems can be leveraged to design soft robots, self-assembling furniture, adaptive fabrics, and self-folding foods. In this talk, Lining presents the recent works in the [...]
The Mechanical Side of Artificial Intelligence
Abstract: Artificial Intelligence typically focuses on perception, learning, and control methods to enable autonomous robots to make and act on decisions in real environments. On the contrary, our research is focused on the design, mechanics, materials, and manufacturing of novel robot platforms that make the perception, control, or action easier or more robust for natural, unstructured, and [...]
Three surprises and a story of prison education
Abstract: I will talk about three results that surprised me. First, I will show that the free configuration space of an elastic wire is path-connected, a result that makes easy a manipulation planning problem that was thought to be hard. Second, I will show a linear relationship between stimulation parameters, skin impedance, and sensation intensity [...]
Carnegie Mellon University
Toward intuitive human controlled MAVs: motion primitives based teleoperation
Abstract: Humans excel at composing high-level plans that achieve a complex, multimodal objective; however, achieving proficiency in teleoperating multi-rotor aerial vehicles (MAVs) in unstructured environments with stability and safety requires significant skill and training. In this talk, we present human-in-the-loop control of a MAV via teleoperation using motion primitives that addresses these concerns. We show [...]
Robots Learning from Human Teachers
Abstract: In this talk I will cover some of the recent work out of the Socially Intelligent Machines Lab at UT Austin (http://sim.ece.utexas.edu/research.html). The vision of our research is to enable robots to function in dynamic human environments by allowing them to flexibly adapt their skill set via learning interactions with end-users. We explore the ways in which [...]
Carnegie Mellon University
Improving Multirotor Trajectory Tracking Performance using Learned Dynamics Models
Abstract: Multirotors and other aerial vehicles have recently seen a surge in popularity, partly due to a rise in industrial applications such as inspection, surveillance, exploration, package delivery, cinematography, and others. Crucial to multirotors' successes in these applications, and enabling their suitability for other applications, is the ability to accurately track trajectories at high speed [...]
Carnegie Mellon University
Automatic Real-time Anomaly Detection for Autonomous Aerial Vehicles
Abstract: The recent incidents with Boeing 737 Max 8 aircraft have raised concerns about the safety and reliability of autopilots and autonomous operations. There is a growing need for methods to monitor the status of aircraft and report any faults and anomalies to the human pilot or to the autopilot to deal with the emergency [...]
Creative Robots with Deep Reinforcement Learning
Recent advances in Deep Reinforcement Learning (DRL) algorithms provided us with the possibility of adding intelligence to robots. Recently, we have been applying a variety of DRL algorithms to the tasks that modern control theory may not be able to solve. We observed intriguing creativity from robots when they are constrained in reaching a certain [...]
Teruko Yata Memorial Lecture – Understanding Human Behavior for Robotic Assistance and Collaboration
Speaker: Henny Admoni, Assistant Professor, Robotics Institute Carnegie Mellon University Title: Understanding Human Behavior for Robotic Assistance and Collaboration . Human-robot collaboration has the potential to transform the way people work and live. Researchers are currently developing robots that assist people in public spaces, on the job, and in their homes. To be effective assistants, these robots [...]
Active Learning in Robot Motion Control
Abstract: Motion motivated by information needs can be found throughout natural systems, yet there is comparatively little work in robotics on analyzing and synthesizing motion for information. Instead, engineering analysis of robots and animal motion typically depends on defining objectives and rewards in terms of states and errors on states. This is how we formulate [...]
Event Cameras: Image Reconstruction, Convolutions and Color
Abstract: Event cameras are novel, bio-inspired visual sensors, whose pixels output asynchronous and independent timestamped spikes at local intensity changes, called ‘events’. Event cameras offer advantages over conventional frame-based cameras in terms of latency, high dynamic range (HDR) and temporal resolution. Event cameras do not output conventional image frames, thus, image reconstruction from events enables [...]
From Farm to Takeoff: Ground and Aerial Robots for Biological Systems Analysis
Abstract: Biological and agricultural environments are dynamic, unstructured, and uncertain, posing challenges for environmental data collection at the necessary spatial and temporal scales to enable meaningful systems analysis. Small unmanned systems, however, can overcome some of these challenges by enabling autonomous or human-assisted image-based and in situ environmental data collection. This talk will present a suite of [...]
Tracking Beyond Detection
Abstract: The majority of existing vision-based methods perform multi-object tracking in the image domain. Yet, in mobile robotics and autonomous driving scenarios, pixel-precise object localization and trajectory estimation in 3D space are of fundamental importance. Furthermore, the leading paradigms for vision-based multi-object tracking and trajectory prediction heavily rely on object detectors and effectively limit tracking [...]
Exploiting Deviations from Ideal Visual Recurrence
Abstract: Visual repetitions are abundant in our surrounding physical world: small image patches tend to reoccur within a natural image, and across different rescaled versions thereof. Similarly, semantic repetitions appear naturally inside an object class within image datasets, as a result of different views and scales of the same object. We studied deviations from these [...]
Attending to Pixels, Embedding Pixels, Predicting Pixels
Abstract: Nowadays splashy applications heavily depend on meticulously annotated datasets, data-driven and learning-based methods, among which pixel labeling plays an important role yet often lacks interpretability. In this talk, I will discuss how we deal with pixels with better interpretability. Firstly, I'll introduce the pixel embedding framework that allows for clustering pixels into discrete groups [...]
Automatically Supervised Learning: Two more steps on a long journey
Abstract: I will talk about two recent pieces of work that attempt to move towards learning with less reliance on labeled data. In the first, part, I will talk about how the surrogate task of predicting the motion of objects can induce complex representations in neural networks without any labeled data. In the second part of [...]
Geometric Deep Learning for Perceiving and Modeling Humans
Abstract: Perceiving and modeling shape and appearance of the human body from single images is a severely under-constrained problem that not only requires large volumes of data, but also prior knowledge. In this talk I will present recent solutions on how deep learning can leverage on geometric reasoning to address tasks like 3D estimation of [...]
Human-Level Learning of Driving Primitives through Bayesian Nonparametric Statistics
Abstract: Understanding and imitating human driver behavior has benefited for autonomous driving in terms of perception, control, and decision-making. However, the complexity of multi-vehicle interaction behavior is far messier than human beings can cope with because of the limited prior knowledge and capability of dealing with high-dimensional and large-scale sequential data. In this talk, I [...]
Formalizing Teamwork in Human-Robot Interaction
Abstract: Robots out in the world today work for people but not with people. Before robots can work closely with ordinary people as part of a human-robot team in a home or office setting, robots need the ability to acquire a new mix of functional and social skills. Working with people requires a shared understanding [...]
Knowledge Transfer Graph for Deep Collaborative Learning
Abstract: In this talk I will present our latest research about knowledge transfer graph for Deep Collaborative Learning (DCL), which is a method that incorporates Knowledge Distillation and Deep Mutual Learning. DCL is represented by a directional graph where each model is represented by a node, and the propagation of knowledge from the source node to the [...]
AI in Space – From Earth Orbit to Mars and Beyond!
Abstract: Artificial Intelligence is playing an increasing role in our everyday lives and the business marketplace. This trend extends to the space sector, where AI has already shown considerable success and has the potential to revolutionize almost every aspect of space exploration. We first highlight a number of success stories of the tremendous impact of [...]
Microsystems-inspired robotics
Abstract: The ability to manufacture micro-scale sensors and actuators has inspired the robotics community for over 30 years. There have been huge success stories; MEMS inertial sensors have enabled an entire market of low-cost, small UAVs. However, the promise of ant-scale robots has largely failed. Ants can move high speeds on surfaces from picnic tables [...]
Carnegie Mellon University
Self-Supervised Learning on Mobile Robots Using Acoustics, Vibration, and Visual Models to Build Rich Semantic Terrain Maps
Abstract: Humans and robots would benefit from having rich semantic maps of the terrain in which they operate. Mobile robots equipped with sensors and perception software could build such maps as they navigate through a new environment. This information could then be used by humans or robots for better localization and path planning, as well [...]
Robotic Grippers for Planetary Applications
Abstract: The previous generation of NASA missions to the outer solar system discovered salt water oceans on Europa and Enceladus, each with more liquid water than Earth – compelling targets to look for extraterrestrial life. Closer to home, JAXA and NASA have imaged sky-light entrances to lava tube caves on the Moon more than 100 [...]
Some New Designs of Convolutional and Recurrent Networks
Abstract: Convolutional networks (CNNs) and recurrent networks have driven the great engineering success of deep learning in recent years. However, as academics, we still wonder whether they are indeed the ultimate models of choice. Especially, CNNs seem unable to characterize predictive uncertainty, and they are highly dependent on small filters on small, rectangular neighborhoods. On [...]
Improving Multi-fingered Robot Manipulation by Unifying Learning and Planning
Abstract: Multi-fingered hands offer autonomous robots increased dexterity, versatility, and stability over simple two-fingered grippers. Naturally, this increased ability comes with increased complexity in planning and executing manipulation actions. As such, I propose combining model-based planning with learned components to improve over purely data-driven or purely-model based approaches to manipulation. This talk examines multi-fingered autonomous [...]
Language and Interaction in Minecraft
Abstract: I will discuss a research program aimed at building a Minecraft assistant, in order to facilitate the study of agents that can complete tasks specified by dialogue, and eventually, to learn from dialogue interactions. I will describe the tools and platform we have built allowing players to interact with the agents and to record those interactions, and [...]
Design, Modeling and Control of a Robot Bat: From Bio-inspiration to Engineering Solutions
Abstract: In this talk, I will describe our recent work building a biologically-inspired bat robot. Bats have a complex skeletal morphology, with both ball-and-socket and revolute joints that interconnect the bones and muscles to create a musculoskeletal system with over 40 degrees of freedom, some of which are passive. Replicating this biological system in a [...]
Attentive Human Action Recognition
Abstract: Enabling computers to recognize human actions in video has the potential to revolutionize many areas that benefit society such as clinical diagnosis, human-computer interaction, and social robotics. Human action recognition, however, is tremendously challenging for computers due to the subtlety of human actions and the complexity of video data. Critical to the success of [...]
Deep Learning for Robotics
Abstract: Programming robots remains notoriously difficult. Equipping robots with the ability to learn would by-pass the need for what otherwise often ends up being time-consuming task specific programming. This talk will describe recent progress in deep reinforcement learning (robots learning through their own trial and error), in apprenticeship learning (robots learning from observing people), and [...]
Temporal Modeling and Data Synthesis for Visual Understanding
Abstract: In this talk, I will present two recent pieces of work on leveraging temporal information and synthetic data to enhance video and image understanding. In the first part, I will introduce a progressive learning framework, Spatio-TEmporalProgressive (STEP), for action detection in videos. STEP is able to more effectively make use of longer temporal information, [...]
Multiple Drone Vision and Cinematography
Abstract: The aim of drone cinematography is to develop innovative intelligent single- and multiple-drone platforms for media production to cover outdoor events (e.g., sports) that are typically distributed over large expanses, ranging, for example, from a stadium to an entire city. The drone or drone team, to be managed by the production director and his/her [...]
Modeling, Design, and Analysis for Intelligent Vehicles: Intersection Management, Security-Aware Design, and Automotive Design Automation
Abstract: Advanced Driver Assistance Systems (ADAS), autonomous functions, and connected applications bring a revolution to automotive systems and software. In this talk, several research topics in the domain of automotive systems and software will be introduced: (1) graph-based modeling, scheduling, and verification for intersection management, (2) security-aware design and analysis considering timing, game theory, and [...]
VR facial animation via multiview image translation
Abstract: A key promise of Virtual Reality (VR) is the possibility of remote social interaction that is more immersive than any prior telecommunication media. However, existing social VR experiences are mediated by inauthentic digital representations of the user (i.e., stylized avatars). These stylized representations have limited the adoption of social VR applications in precisely those [...]
Neural Volumes: Learning Dynamic Renderable Volumes from Images
Abstract: Modeling and rendering of dynamic scenes is challenging, as natural scenes often contain complex phenomena such as thin structures, evolving topology, translucency, scattering, occlusion, and biological motion. Mesh-based reconstruction and tracking often fail in these cases, and other approaches (e.g., light field video) typically rely on constrained viewing conditions, which limit interactivity. We [...]
RI40 Seminar: From Direct-Drive to SuperLimb Bionics
In 1980-81 the first Direct-Drive robot was developed at the CMU Robotics Institute. After almost 40 years, Direct-Drive has a renewed interest in the leg robotics community. Robotic legs powered by direct-drive or low gear-reduction motors can better interact with the ground and absorb impacts. In this seminar I will talk about robot design in [...]
Tartan AUV: A Dive into Carnegie Mellon’s RoboSub Team
Abstract: Founded last year, Tartan AUV is Carnegie Mellon’s undergraduate underwater robotics team which competes annually in the RoboSub competition. RoboSub teams must design, build, and test autonomous underwater vehicles that compete each August to complete tasks related to underwater navigation, object detection and manipulation, and acoustic beacon localization. In this talk we will provide [...]
DNA and gammaPNA in programmable nanomaterials for sensing, robotics and manufacturing
Abstract: When programmable nanomaterials are used in conjunction with rapid microfabrication techniques like two photon polymerization, it becomes possible to rapidly prototype microstructures with nanoscale components. In this research presentation I introduce DNA nanotechnology using a commonly used simple nanotube motif, and I will illustrate how nucleic acid nanotubes can be used in sensing, robotics [...]
Towards Lightweight Real-time Hand Reconstruction in Challenging
Abstract: Humans naturally use their hands to interact and communicate with their surroundings. Reconstructing these complex and dexterous hand interactions enables sign-language recognition and translation, better assistive robots, and more immersive human-computer interaction (e.g. for AR and VR). To make hand reconstruction usable for the aforementioned applications and to a wide set of users, the [...]
Hybrid Methods for the Integration of Heterogeneous Multimodal Biomedical Data
Abstract: The prevalence of smartphones and wearable devices for health monitoring and widespread use of electronic health records have led to a surge in heterogeneous multimodal healthcare data, collected at an unprecedented scale. My research focuses on developing machine learning techniques that learn salient representations of multimodal, heterogeneous data for biomedical predictive models. The first [...]
The Robots are Coming – to your Farm! AKA: Autonomous and Intelligent Robots in Unstructured Field Environments
Abstract: What if a team of collaborative autonomous robots grew your food for you? In this talk, I will discuss some key advances in robotics, machine learning, and autonomy that will one day enable teams of small robots to grow food for you in your backyard in a fundamentally more sustainable way than modern mega-farms! [...]
Self-Driving Cars & AI: Transforming our Cities and our Lives
Abstract: Recent algorithmic and hardware improvements resulted in several success stories in the field of Artificial Intelligence (AI) which impact our daily lives. However, despite its ubiquity, AI is only just starting to make advances in what may arguably have the largest societal impact thus far, the nascent field of autonomous driving. At Uber ATG, [...]
Improving Robot and Deep Reinforcement Learning via Quality Diversity and Open-Ended Algorithms
Abstract: Quality Diversity (QD) algorithms are those that seek to produce a diverse set of high-performing solutions to problems. I will describe them and a number of their positive attributes. I will then summarize our Nature paper on how they, when combined with Bayesian Optimization, produce a learning algorithm that enables robots, after being damaged, to adapt in 1-2 minutes [...]
Go, fastMRI, and Minecraft: Exploring the limits of AI
Abstract: The application of AI across various domains demonstrates both the promise of existing techniques but also their limitations. In this talk, I explore three recent projects and how they shed light on the progress of AI and the challenges to come. These projects include ELF OpenGo a reimplementation of AlphaZero, fastMRI for reducing the time [...]
Towards Weakly-Supervised Visual Understanding
Abstract: Learning with weak and self-supervisions recently emerged as compelling tools towards leveraging vast amounts of unlabeled or partially-labeled data. In this talk, I will present some of the latest advances in weakly-supervised visual scene understanding from NVIDIA. Specifically, I will summarize and discuss some challenges and potential solutions in weakly-supervised learning, and introduce our [...]
Imaging without focusing: A computational approach to miniaturizing cameras
Abstract: Miniaturization of cameras is key to enabling new applications in areas such as connected devices, wearables, implantable medical devices, in vivo microscopy, and micro-robotics. Recently, lenses were identified as the main bottleneck in miniaturization of cameras. Standard smaller lens-system camera modules have a thickness of about 10 mm or higher, and reducing the size [...]
Towards photo-realistic face digitization from monocular videos
Abstract: Recent advances in face capture now enable digitizing high-quality 3D faces for the entertainment industry. Standardized digitization solutions, however, require tailor-made capture systems and extensive manual work, making them expensive and hard to deploy. With the advent of commodity sensors, new lightweight approaches that push the boundaries of human digitization have been introduced, slowly [...]
Toward telelocomotion: human sensorimotor control of contact-rich robot dynamics
Abstract: Human interaction with the physical world is increasingly mediated by automation -- planes assist pilots, cars assist drivers, and robots assist surgeons. Such semi-autonomous machines will eventually pervade our world, doing dull and dirty work, assisting the elderly and disabled, and responding to disasters. Recent results (e.g. from the DARPA Robotics Challenge) demonstrate that, [...]
Formal Synthesis for Robots
Abstract: In this talk I will describe how formal methods such as synthesis – automatically creating a system from a formal specification – can be leveraged to design robots, explain and provide guarantees for their behavior, and even identify skills they might be missing. I will discuss the benefits and challenges of synthesis techniques and [...]
Reconstructing 3D Human Avatars from Monocular Images
Abstract: Statistical 3D human body models have helped us to better understand human shape and motion and already enabled exciting new applications. However, if we want to learn detailed, personalized, and clothed models of human shape, motion, and dynamics, we require new approaches that learn from ubiquitous data such as plain RGB-images and video. I [...]
Extreme Motions in Biological and Engineered Systems
Abstract: Dr. Temel’s work mainly focuses on understanding the dynamics and energetics of extreme motions in small-scale natural and synthetic systems. Small-scale biological systems achieve extraordinary accelerations, speeds, and forces that can be repeated with minimal costs throughout the life of the organism. Zeynep uses analytical and computational models as well as physical prototypes to learn about these systems, test [...]
Reasoning about complex media from weak multi-modal supervision
Abstract: In a world of abundant information targeting multiple senses, and increasingly powerful media, we need new mechanisms to model content. Techniques for representing individual channels, such as visual data or textual data, have greatly improved, and some techniques exist to model the relationship between channels that are “mirror images” of each other and contain [...]
CANCELLED
Building Trust in Real World Applications of Vision Based Machine Learning
Abstract: In all machine learning problems, there is an explicit trade off between cost and benefit. In real world vision problems, this optimization becomes increasingly difficult since those trade offs directly impact technology and product development as well as business strategy. For any successful business case, it is critical that the cost/benefit trade offs in [...]
Knowledge Infused Deep Learning
Abstract: This talk is motivated by the following thesis: Background knowledge is key to intelligent decision making. While deep learning methods have made significant strides over the last few years, they often lack the context in which they operate. Knowledge Graphs (and more generally multi-relational graphs) provide a flexible framework to capture and represent knowledge [...]
Yes, That’s a Robot in Your Grocery Store. Now what?
Abstract: Retail stores are becoming ground zero for indoor robotics. Fleet of different robots have to coexist with each others and humans every day, navigating safely, coordinating missions, and interacting appropriately with people, all at large scale. For us roboticists, stores are giant labs where we're learning what doesn't work and iterating. If we get [...]
Learning to Reconstruct 3D Humans
Abstract: Recent advances in 2D perception have led to very successful systems, able to estimate the 2D pose of humans with impressive robustness. However, our interactions with the world are fundamentally 3D, so to be able to understand, explain and predict these interactions, it is crucial to reconstruct people in 3D. In this talk, I [...]
CANCELLED
Abstract: Before learning robots can be deployed in the real world, it is critical that probabilistic guarantees can be made about the safety and performance of such systems. In recent years, safe reinforcement learning algorithms have enjoyed success in application areas with high-quality models and plentiful data, but robotics remains a challenging domain for scaling [...]
Deep Learning for Understanding Dynamic Visual Data
Abstract: Perceiving dynamic environments from visual inputs allows autonomous agents to understand and interact with the world and is a core topic in Artificial Intelligence. The success of deep learning motivates us to apply deep learning techniques to the perception of dynamic visual data. However, how to design and apply deep neural networks to effectively [...]
Optimizing for coordination with people
https://youtu.be/AQ-w5o2oGI8 Abstract: From autonomous cars to quadrotors to mobile manipulators, robots need to co-exist and even collaborate with humans. In this talk, we will explore how our formalism for decision making needs to change to account for this interaction, and dig our heels into the subtleties of modeling human behavior -- sometimes strategic, often irrational, [...]
Analyzing Grasp Contact via Thermal Imaging
Abstract: Grasping and manipulating objects is an important human skill. Because contact between hand and object is fundamental to grasping, measuring it can lead to important insights. However, observing contact through external sensors is challenging because of occlusion and the complexity of the human hand. I will discuss the use of thermal cameras to capture [...]
Fast Foveation for LIDARs, Projectors and Cameras
Abstract: Most cameras today capture images without considering scene content. In contrast, animal eyes have fast mechanical movements that control how the scene is imaged in detail by the fovea, where visual acuity is highest. This concentrates computational (i.e. neuronal) resources in places where they are most needed. The prevalence of foveation, and the wide [...]
Learning to See Through Occlusions and Obstructions
Virtual VASC: https://cmu.zoom.us/j/249106600 Abstract: Photography allows us to capture and share memorable moments of our lives. However, 2D images appear flat due to the lack of depth perception and may suffer from poor imaging conditions such as taking photos through reflecting or occluding elements. In this talk, I will present our recent efforts to [...]
Detectron2 in Object Detection Research
Virtual VASC: https://cmu.zoom.us/j/249106600 Abstract: Detectron2 is Facebook's library for object detection and segmentation. It has been used widely in FAIR's research and Facebook's products. This talk will introduce detectron2 with a focus on its use in object detection research, including the lessons we learned from building it, as well as the new research enabled [...]
Fairness in visual recognition
Virtual VASC Seminar: https://cmu.zoom.us/j/249106600 Abstract: Computer vision models trained on unparalleled amounts of data hold promise for making impartial, well-informed decisions in a variety of applications. However, more and more historical societal biases are making their way into these seemingly innocuous systems. Visual recognition models have exhibited bias by inappropriately correlating age, gender, sexual [...]
Bio-inspired depth sensing using computational optics
Virtual Seminar: https://cmu.zoom.us/j/249106600 Abstract: Jumping spiders rely on accurate depth perception for predation and navigation. They accomplish depth perception, despite their tiny brains, by using specialized optics. Each principal eye includes a multitiered retina that simultaneously receives multiple images with different amounts of defocus, and distance is decoded from these images with seemingly little [...]
Task-specific Vision DNN Models and Their Relation for Explaining Different Areas of the Visual Cortex
Virtual VASC Seminar: https://cmu.zoom.us/j/249106600 Abstract: Deep Neural Networks (DNNs) are state-of-the-art models for many vision tasks. We propose an approach to assess the relationship between visual tasks and their task-specific models. Our method uses Representation Similarity Analysis (RSA), which is commonly used to find a correlation between neuronal responses from brain data and models. [...]
End-to-end Generative 3D Human Shape and Pose Models and Active Human Sensing
Virtual VASC Seminar: https://cmu.zoom.us/j/249106600 Title: End-to-end Generative 3D Human Shape and Pose Models and Active Human Sensing Abstract: I will review some of our recent work in 3d human modeling, synthesis, and active vision. I will present our new, end-to-end trainable nonlinear statistical 3d human shape and pose models of different resolutions (GHUM and GHUMLite) as [...]
Telling Left from Right: Learning Spatial Correspondence Between Sight and Sound
Virtual VASC Seminar: https://cmu.zoom.us/j/92741882813?pwd=R1R0eGRaeXFHTEF2VWNwY2VIZmU5Zz09 Abstract: Self-supervised audio-visual learning aims to capture useful representations of video by leveraging correspondences between visual and audio inputs. Existing approaches have focused primarily on matching semantic information between the sensory streams. In my talk, I’ll describe a novel self-supervised task to leverage an orthogonal principle: matching spatial information in the [...]
The Topology of Learning
Zoom Virtual Meeting: https://cmu.zoom.us/j/92178295543?pwd=L2dwZU5SbDY5NzZZNzZ4ZmFUclRqQT09 Abstract: Deep Neural Networks (DNNs) have revolutionized computer vision. We now have DNNs that achieve top results in many computer vision problems, including object recognition, facial expression analysis, and semantic segmentation, to name but a few. Unfortunately, the rise in performance has come with a cost. DNNs have become so [...]
Implicit Neural Scene Representations
Virtual Zoom Seminar: https://cmu.zoom.us/j/92178295543?pwd=L2dwZU5SbDY5NzZZNzZ4ZmFUclRqQT09 Abstract How we represent signals has major implications for the algorithms we build to analyze them. Today, most signals are represented discretely: Images as grids of pixels, shapes as point clouds, audio as grids of amplitudes, etc. If images weren't pixel grids - would we be using convolutional neural networks [...]
Computational Imaging: Beyond the Limits Imposed by Lenses
Virtual VASC Seminar: https://cmu.zoom.us/j/92587238250?pwd=S0paYUVBUXozQkFTclMwRUg0MzBNZz09 Abstract: The lens has long been a central element of cameras, since its early use in the mid-nineteenth century by Niepce, Talbot, and Daguerre. The role of the lens, from the Daguerrotype to modern digital cameras, is to refract light to achieve a one-to-one mapping between a point in the scene and a point on the sensor. This effect enables the sensor to compute a particular two-dimensional (2D) [...]
Beyond ROS: Using a Data Connectivity Framework to build and run Autonomous Systems
Virtual FRC Seminar: Seminar recording: https://cmu.zoom.us/rec/share/x84qF7_q8TlIcpHoyG_DRa58O6i8aaa8hCAW_fEPxEkBGjBVPyzW_lK0YW30RfJ3?startTime=1598551489000 Passcode: qu6)ePH9 Abstract: Next-generation robotics will need more than the current ROS code in order to comply with the interoperability, security and scalability requirements for commercial deployments. This session will provide a technical overview of ROS, ROS2 and the Data Distribution Service™ (DDS) protocol for data connectivity in safety-critical cyber-physical [...]
Learning 3D Reconstruction in Function Space
Virtual VASC Seminar: https://cmu.zoom.us/j/96635002737?pwd=RkxGVlJaUTlhcDdGeVBPcnpTS015dz09 Abstract: In this talk, I will show several recent results of my group on learning neural implicit 3D representations, departing from the traditional paradigm of representing 3D shapes explicitly using voxels, point clouds or meshes. Implicit representations have a small memory footprint and allow for modeling arbitrary 3D toplogies at [...]
Scaling Probabilistically Safe Learning to Robotics
Abstract: Before learning robots can be deployed in the real world, it is critical that probabilistic guarantees can be made about the safety and performance of such systems. In recent years, safe reinforcement learning algorithms have enjoyed success in application areas with high-quality models and plentiful data, but robotics remains a challenging domain for [...]
Compositional Representations for Visual Recognition
Virtual VASC - https://cmu.zoom.us/j/99437689110?pwd=cWxuQkIwWlFFZEk0QkVDUVFiN0lTdz09 Abstract: Compositionality is the ability for a model to recognize a concept based on its parts or constituents. This ability is essential to use language effectively as there exists a very large combination of plausible objects, attributes, and actions in the world. We posit that visual recognition models should be [...]
From kinematic to energetic design and control of wearable robots for agile human locomotion
Abstract: Even with the help of modern prosthetic and orthotic (P&O) devices, lower-limb amputees and stroke survivors often struggle to walk in the home and community. Emerging powered P&O devices could actively assist patients to enable greater mobility, but these devices are currently designed to produce a small set of pre-defined motions. Finite state machines [...]
Making 3D Predictions with 2D Supervision
Abstract: Building computer vision systems that understand 3D shape are important for applications including autonomous vehicles, graphics, and VR / AR. If we assume 3D shape supervision, we can now build systems that do a reasonable job at predicting 3D shapes from images. However, 3D supervision is difficult to obtain at scale; therefore we should [...]
The World’s Tiniest Space Program
Abstract: The aerospace industry has experienced a dramatic shift over the last decade: Flying a spacecraft has gone from something only national governments and large defense contractors could afford to something a small startup can accomplish on a shoestring budget. A virtuous cycle has developed where lower costs have led to more launches and the [...]
Perceiving 3D Human-Object Spatial Arrangements from a Single Image In-the-wild
Abstract: We live in a 3D world that is dynamic—it is full of life, with inhabitants like people and animals who interact with their environment through moving their bodies. Capturing this complex world in 3D from images has a huge potential for many applications such as compelling mixed reality applications that can interact with people [...]
A future with affordable Self-driving vehicles
(Video to appear once approved) Abstract: We are on the verge of a new era in which robotics and artificial intelligence will play an important role in our daily lives. Self-driving vehicles have the potential to redefine transportation as we understand it today. Our roads will become safer and less congested, while parking spots will be repurposed as leisure [...]
Detection of Photo Manipulation with Media Forensics
Abstract: Rapid progress in machine learning, computer vision and graphics leads to successive democratization of media manipulation capabilities. While convincing photo and video manipulation used to require substantial time and skill, modern editors bring (semi-) automated tools that can be used by everyone. Some of the most recent examples include manipulation of human faces, e.g., [...]
Robotics and Biosystems
Abstract: Research at the Center for Robotics and Biosystems at Northwestern University encompasses bio-inspiration, neuromechanics, human-machine systems, and swarm robotics, among other topics. In this talk I will give an overview of some of our recent work on in-hand manipulation, robot locomotion on yielding ground, and human-robot systems. Biography: Kevin Lynch received the B.S.E. degree [...]
Advancing the State of the Art of Computer Vision for Billions of Users
Abstract: At Google, advancing the state of the art of computer vision is very impactful as there are billions of users of Google products, many of which require high-quality, artifact-free images. I will share what we learned from successfully launching core computer vision techniques for various Google products, including PhotoScan (Photos), seamless Google Street View [...]
Learning-based 6D Object Pose Estimation in Real-world Conditions
Abstract: Estimating the 6D pose, i.e., 3D rotation and 3D translation, of objects relative to the camera from a single input image has attracted great interest in the computer vision community. Recent works typically address this task by training a deep network to predict the 6D pose given an image as input. While effective on [...]
SubT Fall Update Webinar Led by CMU’s Robotics Institute faculty members Sebastian Scherer and Matt Travers, as well as OSU’s Geoff Hollinger
We invite you to meet members of the award-winning Team Explorer, the CMU DARPA Subterranean Challenge team, and learn more about this groundbreaking competition. Some of the world's top universities have entered the DARPA Subterranean Challenge, developing technologies to map, navigate, and search underground environments. Led by CMU's Robotics Institute faculty members Sebastian Scherer and Matt [...]
Deep Learning: (still) Not Robust
Abstract: One of the key limitations of deep learning is its inability to generalize to new domains. This talk studies recent attempts at increasing neural network robustness to both natural and adversarial distribution shifts. Robustness to adversarial examples, inputs crafted specifically to fool machine learning models, are arguably the most difficult type of domain shift. [...]
Drones in Public: distancing and communication with all users
Abstract: This talk will focus on the role of human-robot interaction with drones in public spaces and be focused on two individual research areas: proximal interactions in shared spaces and improved communication with both end-users and bystanders. Prior work on human-interaction with aerial robots has focused on communication from the users or about the intended direction [...]
End-to-End ‘One Networks’: Learning Regularizers for Least Squares via Deep Neural Networks
Abstract: Linear Restoration Problems (or Linear Inverse Problems) involve reconstructing images or videos from noisy measurement vectors. Notable examples include denoising, inpainting, super-resolution, compressive sensing, deblurring and frame prediction. Often, multiple such tasks should be solved simultaneously, e.g., through Regularized Least Squares, where each individual problem is underdetermined (overcomplete) with infinitely many solutions from which [...]
Data Scalability for Robot Learning
Abstract: Recent progress in robot learning has demonstrated how robots can acquire complex manipulation skills from perceptual inputs through trial and error, particularly with the use of deep neural networks. Despite these successes, the generalization and versatility of robots across environment conditions, tasks, and objects remains a major challenge. And, unfortunately, our existing algorithms and [...]
Carnegie Mellon University
Learning to Generalize beyond Training
Abstract: Generalization, i.e., the ability to adapt to novel scenarios, is the hallmark of human intelligence. While we have systems that excel at cleaning floors, playing complex games, and occasionally beating humans, they are incredibly specific in that they only perform the tasks they are trained for and are miserable at generalization. One of the [...]
Detecting Image Synthesis — Shallow and Deep
Abstract: The proliferation of synthetic media are subject to malicious usages such as disinformation campaigns, posing potential threats to media integrity and democracy. A way to combat this is developing forensics algorithms to identify manipulated media. In the beginning of the talk, I will discuss how one can train a model to detect photos manipulated [...]
Deep Learning to Distinguish Recalled but Benign Mammography Images in Breast Cancer Screening
Abstract: Breast cancer screening using the standard mammography exam currently exhibits a high false recall rate (11.6% for women in the U.S.). Only a low proportion (0.5%) of women who were recalled for additional workup were actually found to have breast cancer. As a result of the unnecessary stress and follow-up work from these false [...]
The Plenoptic Camera
Abstract: Imagine a futuristic version of Google Street View that could dial up any possible place in the world, at any possible time. Effectively, such a service would be a recording of the plenoptic function—the hypothetical function described by Adelson and Bergen that captures all light rays passing through space at all times. While the plenoptic function [...]
Photorealistic Reconstruction of Landmarks and People using Implicit Scene Representation
Abstract: Reconstructing scenes to synthesize novel views is a long standing problem in Computer Vision and Graphics. Recently, implicit scene representations have shown novel view synthesis results of unprecedented quality, like the ones of Neural Radiance Fields (NeRF), which use the weights of a multi-layer perceptron to model the volumetric density and color of a [...]