layout | title | subtitle | description | show_sidebar | hide_footer | permalink | hero_image | image | published |
---|---|---|---|---|---|---|---|---|---|
page |
Tartan SLAM Series Fall Edition |
Fall 2021 interactive series of talks, tutorials, and learning on SLAM |
Fall 2021 interactive series of talks, tutorials, and learning on SLAM |
false |
false |
/tartanslamseries2/ |
/img/slam_series/tartanSLAMbannerFall.png |
/img/slam_series/thumbAllFallv8.png |
true |
The goal of this series is to expand the understanding of those both new and experienced with SLAM. Sessions will include research talks, as well as introductions to various themes of SLAM and thought provoking open-ended discussions. The lineup of events aim to foster fun, provocative discussions on robotics.
Join our mailing list for Zoom links, updates and reminders about the presentations. Our Tartan SLAM Series Discord server, aims to foster an inclusive learning community for SLAM. If you are an expert or newcomer, we invite you to join the discord server to build the community.
Join our Mailing List Subscribe to our YouTube Channel Join our Discord
Each talk format is 40 minutes of presentation and 20 minutes of open-ended discussion.
Check out our previous edition of the Tartan SLAM Series here.
Presenter | Session Title | Date/Time | YouTube Link | |
---|---|---|---|---|
Professor in Interactive Computing Georgia Tech |
Factor Graphs for Perception and Action |
30 Sept 2021 1:30 PM EST |
|
|
Outline:
|
||||
Professor of Electrical Engineering University of Chile |
Unifying the SLAM Back and Front Ends with Random Finite Sets |
7 Oct 2021 12:30 PM EST |
|
|
Professor of Informatics and Mathematics Technical University of Munich |
Deep and Direct Visual SLAM |
14 Oct 2021 12:00 PM EST |
|
|
Assistant Professor, Naval Architecture and Marine Engineering University of Michigan |
Deep Learning for Marine Robot Perception |
18 Oct 2021 12:00 PM EST |
|
|
Professor of Robotics and Perception University of Zurich |
SLAM: from Frames to Events |
21 Oct 2021 12:00 PM EST |
|
|
Associate Professor, Institute for Aerospace Studies University of Toronto |
Unlocking Dynamic Cameras for Visual Navigation Expand Contents |
25 Oct 2021 12:00 PM EST |
|
|
Abstract: Gimbal-stabilized dynamic cameras provide many advantages in robotic applications governed by highly dynamic motion profiles and uneven feature distributions, due to their ability to provide smooth image capture independent of robot motion. In order to integrate information received from gimballed cameras, an accurate time-varying extrinsic calibration between the dynamic camera and other sensors, such as static cameras and IMUs, needs to be determined. In this talk, I will first present our work on the extrinsic calibration for dynamic and static camera clusters. I will then talk about our recent efforts to perform the calibration between a dynamic camera and an IMU, online and in flight while presenting results in simulation and real hardware data. |
||||
Professor University of Bonn |
Dynamic Environments in Least Squares SLAM |
28 Oct 2021 12:00 PM EST |
|
|
Agricultural Engineer U.S. Department of Agriculture |
Reconstructing small things in large spaces, and other reconstruction stories. Expand Contents |
1 Nov 2021 12:00 PM EST |
|
|
Bio: Amy Tabb holds degrees from Sweet Briar College (B.A. Math/Computer Science and Music), Duke University (M.A. Musicology), and Purdue University (M.S. and Ph.D. Electrical and Computer Engineering) and is a Research Agricultural Engineer at a US Department of Agriculture, Agricultural Research Service laboratory in Kearneysville, West Virginia. There, she has been engaged in creating systems for automation in the tree fruit industry. Her research interests are within the fields of computer vision and robotics, in particular robust three-dimensional reconstruction and perception in outdoor conditions. Abstract: Three-dimensional reconstruction is an intermediate step needed for many applications in agriculture, including automation and phenotyping of plants and fruits. The nature of agricultural objects is that they have low texture compared to objects in typical computer vision datasets and consequently many classical approaches do not work well for camera pose localization and reconstruction. In this talk, I will discuss some work I have done on reconstructing a range of objects, from leafless trees with a robot-camera system to individual fruits with a tabletop system. Throughout, I will discuss failures and motivations for choosing one approach over another, as well as why working on three-dimensional reconstruction in agriculture has led to work on calibration systems. |
||||
Assistant Professor ETH Zurich |
SLAM and beyond: advancing vision-based Robotic Perception |
4 Nov 2021 12:00 PM EST |
|
|
Professor, Institute for Aerospace Studies University of Toronto |
Where Can Machine Learning Help Robotic State Estimation? Expand Contents |
11 Nov 2021 12:30 PM EST |
|
|
Abstract: Classic state estimation tools (e.g., determining position/velocity of a robot from noisy sensor data) have been in use since the 1960s, perhaps the most famous technique being the Kalman filter. For difficult-to-model nonlinear systems with rich sensing (e.g., almost any real-world robot), clever adaptations are needed to the classic tools. In this talk, I will first briefly summarize a few ideas that have become standard practice in our group over the last several years: continuous-time trajectory estimation (and its connection to sparse Gaussian process regression) as well as estimation on matrix Lie groups (to handle rotations cleanly). I will also discuss two new frameworks we have been pursuing lately: exactly sparse Gaussian variational inference (ESGVI) and Koopman state estimation (KoopSE). ESGVI seeks to minimize the Kullback-Leibler divergence between a Gaussian state estimate and the full Bayesian posterior; however, the framework also easily allows for parameter learning through Expectation Maximization and we’ve used this to learn simple parameters such as constant system matrices and covariances, but also to model rich sensors using Deep Neural Networks and learn the weights from data. KoopSE takes a different approach by lifting a nonlinear system into a high-dimensional Reproducing Kernel Hilbert Space where we can treat it as linear and apply classic estimation tools; it also allows for the system to be learned from training data quite efficiently. For all these ideas, I will give simple intuitive explanations of the mathematics and show some examples of things working in practice. |
||||
University of Washington |
Learning to Navigate in the Real World without a Metric Map |
18 Nov 2021 12:30 PM EST |
|
|
Senior Research Scientist CSIRO Data61 |
CSIRO's Wildcat SLAM: A Robust SLAM System for Robot Teams in the Wild |
22 Nov 2021 5:00 PM EST |
|
|
Principal Investigator and Research Technologist NASA JPL |
Resilient Robotic Autonomy Under Uncertainty |
29 Nov 2021 2:00 PM EST |
|
|
Senior Researcher Czech Institute of Informatics, Robotics and Cybernetics |
Long-Term Visual Localization |
2 Dec 2021 12:00 PM EST |
|
|
Professor for Computer Science University of Freiburg |
Probabilistic and Deep Learning Techniques for Robot Navigation and Automated Driving |
9 Dec 2021 12:00 PM EST |
|