Publications

  

Combining 3D Shape, Color, and Motion for Robust Anytime Tracking
David Held, Jesse Levinson, Sebastian Thrun, Silvio Savarese.
Robotics: Science and Systems (RSS), 2014.
Improving on our ICRA 2013 paper, this new approach enables real-time probabilistic object tracking. Now, computational time is allocated dynamically according to the shape of the track’s posterior distribution. The algorithm is “anytime”, allowing speed or accuracy to be optimized based on the needs of the application. We currently use this method to track all dynamic obstacles seen by our autonomous vehicle, in real-time, with significantly improved accuracy compared to our previous Kalman-filter based approach.

pdf (rss)Project Page - Supplementary material, C++ code, poster, presentationbib

Group Induction
Alex Teichman and Sebastian Thrun.
Proc. of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2013.
Tracking-based semi-supervised learning, as originally presented at RSS2011, was an offline algorithm. This is fine in some contexts, but ideally a user could provide new hand-labeled training examples online, as the system runs, without retraining from scratch. Qualitatively, this would mean the ability to point out - from the back seat of your autonomous car - a few examples of, say, an elliptical bike or sk8poler, and the algorithm would start learning to recognize them on the fly without you having to do anything else. Group induction is a mathematical framework for this kind of learning.

pdfbib

Automatic Online Calibration of Cameras and Lasers
Jesse Levinson and Sebastian Thrun.
Robotics: Science and Systems (RSS), 2013.
Extending previous work on offline 6-DOF calibration of 3D laser sensors to 2D cameras, this paper presents two new real-time techniques that enable camera-laser calibration online, automatically, and in arbitrary environments. The first is a probabilistic monitoring algorithm that can detect a sudden mis-calibration in a fraction of a second. The second is a continuous calibration optimizer that adjusts transform offsets in real time, tracking gradual sensor drift as it occurs. Together, these techniques allow significantly greater flexibility and adaptability of robots in unknown and potentially harsh environments.

pdfbib

Precision Tracking with Sparse 3D and Dense Color 2D Data
David Held, Jesse Levinson, and Sebastian Thrun.
International Conference on Robotics and Automation (ICRA), 2013.
Precision tracking is important for predicting the behavior of other cars in autonomous driving. We present a novel method to combine sparse laser data with a high-resolution camera image to achieve accurate velocity estimates of moving objects. We present a color-augmented, pre-filtered grid search algorithm to align the points from a tracked object, thereby obtaining much more precise estimates of the tracked vehicle’s velocity than were possible with previous methods.

pdfbib

Automatic Calibration of Cameras and Lasers in Arbitrary Scenes
Jesse Levinson and Sebastian Thrun.
International Symposium on Experimental Robotics (ISER), 2012.
This paper presents a new algorithm for automatically calibrating cameras to multi-beam lasers on a mobile robot given a series of frames from both sensors. Our method does not require the use of a known calibration target, nor does it require any hand labeling of correspondences. Even without these requirements, by leveraging unsupervised data, it still outperforms previous state-of-the-art calibration techniques by a significant margin.

bib

Online, semi-supervised learning for long-term interaction with object recognition systems
Alex Teichman and Sebastian Thrun.
Invited talk at RSS Workshop on Long-term Operation of Autonomous Robotic Systems in Changing Environments, 2012.
Tracking-based semi-supervised learning, as originally presented at RSS2011, was an offline algorithm. This is fine in some contexts, but ideally a user could provide new hand-labeled training examples online, as the system runs, without retraining from scratch. Qualitatively, this would mean the ability to point out - from the back seat of your autonomous car - a few examples of, say, an elliptical bike or sk8poler, and tracking-based semi-supervised learning would start learning to recognize them on the fly without you having to do anything else. This talk discusses some preliminary work in this direction.

presentation

Tracking-based semi-supervised learning
Alex Teichman and Sebastian Thrun.
International Journal of Robotics Research (IJRR), 2012.
Extended journal version of previous work with the same title. More experiments, more intuition as to how the method works.

pdf (sage)bib

A Probabilistic Framework for Object Detection in Images using Context and Scale
David Held, Jesse Levinson, and Sebastian Thrun.
International Conference on Robotics and Automation (ICRA), 2012.
Detecting cars in real-world images is an important task for autonomous driving, yet it remains unsolved. The system described in this paper takes advantage of context and scale to build a monocular single-frame image-based car detector that significantly outperforms previous state-of-the-art methods. By using a calibrated camera and localization on a road map, we are able to obtain context and scale information from a single image without the use of a 3D laser.

pdfbib

Practical object recognition in autonomous driving and beyond
Alex Teichman and Sebastian Thrun.
IEEE Workshop on Advanced Robotics and its Social Impacts (ARSO), 2011.
This paper gives an overview of the recent object recognition research in our lab and what is needed to make it a fully functional, high accuracy object recognition system that is applicable beyond perception for autonomous driving.

pdfbib

Tracking-based semi-supervised learning
Alex Teichman and Sebastian Thrun.
Robotics: Science and Systems (RSS), 2011.
Building on previous work, we introduce a simple semi-supervised learning method that uses tracking information to find new, useful training examples automatically. This method achieves nearly the same accuracy as before, but with about two orders of magnitude less human labeling effort.

pdfbibprojectRSS proceedings

Towards 3D object recognition via classification of arbitrary object tracks
Alex Teichman, Jesse Levinson, and Sebastian Thrun.
International Conference on Robotics and Automation (ICRA), 2011.
Breaking down the object recognition problem into segmentation, tracking, and track classification components, we show an accurate and real-time method of classifying tracked objects as car, pedestrian, bicyclist, or 'other'.

pdfbibdataset

Towards fully autonomous driving: systems and algorithms
Jesse Levinson, Jake Askeland, Jan Becker, Jennifer Dolson, David Held, Soeren Kammel, J. Zico Kolter, Dirk Langer, Oliver Pink, Vaughan Pratt, Michael Sokolsky, Ganymed Stanek, David Stavens, Alex Teichman, Moritz Werling, and Sebastian Thrun.
Intelligent Vehicles Symposium, 2011.
This paper is a broad summary of recent work on Junior, Stanford's autonomous vehicle. Topics covered include object recognition, sensor calibration, planning, control, etc.

pdfbib

Traffic Light Mapping, Localization, and State Detection for Autonomous Vehicles
Jesse Levinson, Jake Askeland, Jennifer Dolson, and Sebastian Thrun.
International Conference on Robotics and Automation (ICRA), 2011.
We present a passive camera-based pipeline for traffic light state detection using imperfect vehicle localization and assuming prior knowledge of traffic light location. To achieve robust real-time detections in a variety of lighting conditions, we combine several probabilistic stages that explicitly account for the corresponding sources of sensor and data uncertainty.

pdfbib

Automatic laser calibration, mapping, and localization for autonomous vehicles
Jesse Levinson.
Thesis (Ph.D.), Stanford University, 2011.
This dissertation presents several related algorithms that enable important capabilities for self-driving vehicles. These include offline mapping and online map-based localization techniques using GPS, IMU, and lasers, online localization without a prerecorded map as used in the DARPA Urban Challenge, intrinsic and extrinsic calibration algorithms for multi-beam lasers, and realtime detection of traffic lights.

pdfStanford Librarybib

Unsupervised Calibration for Multi-beam Lasers
Jesse Levinson and Sebastian Thrun.
International Symposium on Experimental Robotics (ISER), 2010.
This paper introduces an unsupervised solution for solving the intrinsic and extrinsic calibration properties of a multi-beam laser on a mobile robot in arbitrary, unknown environments. By defining and optimizing an objective function that rewards 3D consistency between points seem by different beams, we are able to calibrate internal angles, range offsets, and remittance response curves for each beam in addition to the 6-DOF position of the laser relative to the vehicle's inertial frame.

pdfbib

Robust Vehicle Localization in Urban Environments Using Probabilistic Maps
Jesse Levinson and Sebastian Thrun.
International Conference on Robotics and Automation (ICRA), 2010.
We extend previous work on localization using GPS, IMU, and LIDAR data by modeling the environment as a probabilistic grid in which every cell is represented as its own gaussian distribution over remittance values. This approach offers higher precision, the ability to learn and improve maps over time, and increased robustness to environment changes and dynamic obstacles.

pdfbib

Exponential family sparse coding with application to self-taught learning
Honglak Lee, Rajat Raina, Alex Teichman, and Andrew Y. Ng.
International Joint Conference on Artificial Intelligence (IJCAI), 2009.

pdfbib

Map-Based Precision Vehicle Localization in Urban Environments
Jesse Levinson and Sebastian Thrun.
Robotics: Science and Systems (RSS), 2007.
GPS-based inertial guidance systems do not provide sufficient accuracy for many urban navigation applications, including autonomous navigation. We propose a technique for high-accuracy localization of moving vehicles that utilizes maps of urban environments. Our approach integrates GPS, IMU, wheel odometry, and LIDAR data to generate high-resolution environment maps. We use offline GraphSLAM techniques to align intersections and regions of self-overlap, and a particle filter to localize the vehicle relative to these maps in real time.

pdfbib

Contents © 2014 Stanford Autonomous Driving Team - Powered by Nikola