SlideShare a Scribd company logo
1 of 45
Download to read offline
Driving Behaviors for ADAS
and Autonomous Driving XIV
Yu Huang
Yu.huang07@gmail.com
Sunnyvale, California
Outline
• TNT: Target-driveN Trajectory Prediction (CVPR’20)
• Driving Through Ghosts: Behavioral Cloning with False Positives (8’29)
• LiRaNet: E2E Trajectory Prediction using Spatio-Temporal Radar Fusion (10.15)
• SimAug: Learning Robust Representations from Simulation for Trajectory
Prediction (ECCV’20)
• Learning Lane Graph Representations for Motion Forecasting (ECCV’20)
• Implicit Latent Variable Model for Scene-Consistent Motion Forecasting (ECCV’20)
• Perceive, Predict, and Plan: Safe Motion Planning Through Interpretable Semantic
Representations (ECCV’20)
TNT: Target-driveN Trajectory Prediction
• This key insight is that for prediction within a moderate time horizon, the
future modes can be effectively captured by a set of target states.
• This leads to target-driven trajectory prediction (TNT) framework.
• TNT has three stages which are trained end-to-end.
• It first predicts an agent’s potential target states T steps into the future, by encoding
its interactions with the environment and the other agents.
• TNT then generates trajectory state sequences conditioned on targets.
• A final stage estimates trajectory likelihoods and a final compact set of trajectory
predictions is selected.
• This is in contrast to previous work which models agent intents as latent
variables, and relies on test-time sampling to generate diverse trajectories.
• Benchmark TNT on trajectory prediction of vehicles and pedestrians,
outperform state-of-the-art on Argoverse Forecasting, INTERACTION,
Stanford Drone and an in-house Pedestrian-at-Intersection dataset.
TNT: Target-driveN Trajectory Prediction
Illustration of the TNT framework when applied to the vehicle future trajectory prediction task. TNT
consists of three stages: (a) target prediction which proposes a set of plausible targets (stars)
among all candidates (diamonds). (b) target-conditioned motion estimation which estimates a
trajectory (distribution) towards each selected target, (c) scoring and selection which ranks
trajectory hypotheses and selects a final set of trajectory predictions with likelihood scores.
TNT: Target-driveN Trajectory Prediction
TNT model overview. Scene context is first encoded as the model’s inputs. Then follows the core
three stages of TNT: (a) target prediction which proposes an initial set of M targets; (b) target-
conditioned motion estimation which estimates a trajectory for each target; (c) scoring and selection
which ranks trajectory hypotheses and outputs a final set of K predicted trajectories.
TNT: Target-driveN Trajectory Prediction
TNT supports flexible choices of targets. Vehicle target candidate points
are sampled from the lane centerlines. Pedestrian target candidate
points are sampled from a virtual grid centered on the pedestrian.
TNT: Target-driveN Trajectory Prediction
Driving Through Ghosts: Behavioral Cloning with
False Positives
• In the context of behavioral cloning, perceptual errors at training time can
lead to learning difficulties or wrong policies, as expert demonstrations
might be inconsistent with the perceived world state.
• In this work, propose a behavioral cloning approach that can safely
leverage imperfect perception without being conservative.
• The core is a representation of perceptual uncertainty for learning to plan.
• It propose a new probabilistic birds-eye-view semantic grid to encode the
noisy output of object perception systems.
• Then leverage expert demonstrations to learn an imitative driving policy
using this probabilistic representation.
• Using the CARLA simulator, it can safely overcome critical false positives
that would otherwise lead to catastrophic failures or conservative behavior.
Driving Through Ghosts: Behavioral Cloning with
False Positives
It is a probabilistic birds-eye-view semantic representation, Soft BEV, for imitation learning
under perceptual uncertainty. It enables learning safer policies that can ignore false positives.
Driving Through Ghosts: Behavioral Cloning with
False Positives
• The observations o in a birds-eye-view grid, i.e., an NxMxD-dim. tensor where each dimension k
represents a category of estimated state (e.g., an object or feature type) together with the
respective estimated confidences.
• Each slice is a matrix of NxM, where each element corresponds to the presence of an estimated
object or feature of type k at that location, weighted by its estimated confidence.
• The resulting input representation is referred as the Soft BEV.
• It models a driving agent via a deep convolutional policy network with input of Soft BEV.
• The CNN outputs way-points along the future trajectory, used by a PID controller to compute the
control signals for the steering and throttle of the vehicle.
• It consists of a ResNet-18 base network acting as an encoder, followed by three deconvolutional
layers which also have as an input the current speed signal.
• For each of the potential high-level commands (“go left”, “go right”, “go straight”, “follow the
road”), the network predicts multiple output heat-maps which are then converted into way-points
by spatial soft-argmax layers.
• Based on the high-level command, the respective head is used to predict the way-points.
Driving Through Ghosts: Behavioral Cloning with
False Positives
Experimental setup: The CARLA simulator provides ground truth features. Perception
noise is applied to the dynamic features, which are then fused into an uncertainty-
scaled birds-eye view representation, the Soft BEV. Together with high-level
commands and speed information it is fed to a CNN that predicts way-points.
Driving Through Ghosts: Behavioral Cloning with
False Positives
LiRaNet: E2E Trajectory Prediction using Spatio-
Temporal Radar Fusion
• LiRaNet, a end-to-end trajectory prediction method which utilizes radar sensor
information along with widely used lidar and high definition (HD) maps.
• Automotive radar provides rich, complementary information, allowing for longer
range vehicle detection as well as instantaneous radial velocity measurements.
• However, there are factors that make the fusion of lidar and radar information
challenging, such as the relatively low angular resolution of radar measurements,
their sparsity and the lack of exact time synchronization with lidar.
• To overcome these challenges, propose an efficient spatio-temporal radar feature
extraction scheme which achieves state-of-the-art performance on multiple
large-scale datasets.
LiRaNet: E2E Trajectory Prediction using Spatio-
Temporal Radar Fusion
An example scene from X17k in bird’s eye view where lidar points (light blue) and radar point velocities (orange)
are visualized with labels (white) for current, past and future frames. Vehicle A is a turning bus that has multiple
radar points across frames. By effectively combining them over space and time a full 2D velocity and turning
rate can be recovered. Vehicle B shows the high positional noise that inherently comes with radar. Vehicle C
shows a case with sparse lidar points where implicitly associating them across time can be challenging.
However, radar points present around C can add context for the model to detect and predict the trajectory.
LiRaNet: E2E Trajectory Prediction using Spatio-
Temporal Radar Fusion
LiRaNet overview: The radar feature extraction network (A) extracts spatio-temporal features from raw radar points in
2 steps: (1) for each frame we create a graph between the BEV grid cells and radar points to learn spatial features of
each cell using a non-rigid convolution, (2) these spatial features are further fused temporally by stacking across
channel dimension and using an MLP to get a radar feature volume. This feature volume is then fused with feature
volumes from other sensors and fed to a joint perception-prediction network (B) which produces detections and their
future trajectories. An example prediction for a scene from X17k can be seen in (C).
LiRaNet: E2E Trajectory Prediction using Spatio-
Temporal Radar Fusion
• Input domain consists of radar points and output domain consists of BEV cells.
• For each cell j, calculate the features hm
j for the sweep m as
• where Am
j is the set of associated radar points, xm
i is the 2D coordinates of the associated
radar point, xm
j is the 2D coordinate of the BEV cell’s center, ⊕ denotes the
concatenation operation, fm
i is the feature vector for the radar point and gm() is an multi-
layer perceptron (MLP) with learnable weights shared across all the cells.
• Calculate Am
j, using nearest neighbor algorithm with a distance threshold.
• By using a threshold larger than the size of a cell, this method compensates for positional
errors in radar.
• For each cell j, calculate the final spatio-temporal feature vector (hj ) by concatenating
the per sweep features hj,m and using an MLP to combine them.
LiRaNet: E2E Trajectory Prediction using Spatio-
Temporal Radar Fusion
RadarisfusedwithLiDARandMaptogenerate
multi-sensorfeaturevolumeinBEVwhichisused
fordetectionandtrajectoryprediction.Each
learnableoperatorisdenotedas(op,kernel,
stride,numberoffeatures).Convdenotes
convolutionandResdenotesaresidualblock.
LiRaNet: E2E Trajectory Prediction using Spatio-
Temporal Radar Fusion
SimAug: Learning Robust Representations from
Simulation for Trajectory Prediction
• This paper studies the problem of predicting future trajectories of people in
unseen cameras of novel scenarios and views.
• It approaches this problem through the real-data-free setting in which the model
is trained only on 3D simulation data and applied out-of-the-box to a wide variety
of real cameras.
• It proposes to learn robust representation through augmenting the simulation
training data such that the representation can better generalize to unseen real-
world test data.
• The key idea is to mix the feature of the hardest camera view with the adversarial
feature of the original view.
• It is referred as SimAug, achieving results on three real-world benchmarks using
zero real training data, and state-of-the-art performance in the Stanford Drone
and the VIRAT/ActEV dataset when using in-domain training data.
• Code and models are released at https://next.cs.cmu.edu/simaug.
SimAug: Learning Robust Representations from
Simulation for Trajectory Prediction
SimAug that is trained on simulation and tested on real unseen videos. Each
training trajectory is represented by multi-view segmentation features extracted
from the simulator. SimAug mixes the feature of the hardest camera view with
the adversarial feature of the original view.
SimAug: Learning Robust Representations from
Simulation for Trajectory Prediction
• Each time given a camera view, use it as an anchor to search for the “hardest" view that is most
inconsistent with what the model has learned.
• It uses the classification loss as the criteria and compute:
• For the original view, generate an adversarial trajectory by the targeted-FGSM attack
• The attack tries to make the model predict the future locations in the selected “hardest" camera
view rather than the original view.
• In essence, the resulting adversarial feature is “warped" to the “hardest” camera view by a small
perturbation.
• By defending against such adversarial trajectory, the model learns representations that are robust
against variances in camera views.
SimAug: Learning Robust Representations from
Simulation for Trajectory Prediction
• It mix up the trajectory locations of the selected view and the adversarial
trajectory locations by a convex combination function over their features and
one-hot location labels.
• where [yh+1, …, yT ] = Lh+1:T are the ground-truth locations of the original view.
• The one-hot (·) function projects the location in xy coordinates into an one-hot
embedding over the predefined grid used in our backbone trajectory prediction
model.
SimAug: Learning Robust Representations from
Simulation for Trajectory Prediction
• Backbone network is the Multiverse model (CVPR’20).
• The training algorithm’s pseudo-code is:
SimAug: Learning Robust Representations from
Simulation for Trajectory Prediction
SimAug: Learning Robust Representations from
Simulation for Trajectory Prediction
Learning Lane Graph Representations for Motion
Forecasting
• A motion forecasting model that exploits a novel structured map representation
as well as actor-map interactions.
• Instead of encoding vectorized maps as raster images, construct a lane graph
from raw map data to explicitly preserve the map structure.
• To capture the complex topology and long range dependencies of the lane graph,
propose LaneGCN which extends graph convolutions with multiple adjacency
matrices and along-lane dilation.
• To capture the complex interactions between actors and maps, exploit a fusion
network consisting of four types of interactions, actor-to-lane, lane-to-lane, lane-
to-actor and actor-to-actor.
• Powered by LaneGCN and actor-map interactions, the model is able to predict
accurate and realistic multi-modal trajectories.
• This approach significantly outperforms the state-of-the-art on the large scale
Argoverse motion forecasting benchmark.
Learning Lane Graph Representations for Motion
Forecasting
It constructs a lane graph from raw map data and use LaneGCN to extract map features. In parallel,
ActorNet extracts actor features from observed past trajectories. Then it uses FusionNet to model the
Interactions between actors themselves and the map, and predict the future trajectories.
Learning Lane Graph Representations for Motion
Forecasting
The model is composed of four modules. (1) ActorNet receives the past actor trajectories as input, and uses 1D convolution to
extract actor node features. (2) MapNet constructs a lane graph from HD maps, and uses a LaneGCN to exact lane node
features. (3) FusionNet is a stack of 4 interaction blocks. The actor to lane block fuses real-time traffic information from actor
nodes to lane nodes. The lane to lane block propagates information over the lane graph and updates lane features. The lane to
actor block fuses updated map information from lane nodes to actor nodes. The actor to actor block performs interactions among
actors. It uses another LaneGCN for the lane to lane block, and spatial attention layers for the other blocks. (4) The prediction
header uses after-fusion actor features to produce multi-modal trajectories.
Learning Lane Graph Representations for Motion
Forecasting
LaneGCN is a stack of 4 multi-scale LaneConv residual
blocks, each of which consists of a LaneConv (1, 2, 4, 8,
16, 32) and a linear layer with a residual connection. All
layers have 128 feature channels.
Left: The lane centerline of interest, its predecessor,
successor, left and right neighbor are denoted with red,
orange, blue, purple, and green lines, respectively.
Each centerline is given as a sequence of BEV points
(hollow circles). Right: Derived lane graph with an
example lane node. The lane node of interest, its
predecessor, successor, left and right neighbor are
denoted with red, orange, blue, purple and green
circles respectively.
Learning Lane Graph Representations for Motion
Forecasting
LaneConv Operator:
parameterize the node feature as follows,
LaneConv operator as
Ai and Wi are the adjacency and the weight matrices
corresponding to the i-th connection type respectively.
MLP indicates a multi-layer perceptron and the
two subscripts refer to shape and location,
respectively.
k-dilation LaneConv operator as
Ak
pre is the k-th matrix power of Apre.
In regular grid graphs, a dilated convolution operator can effectively
capture the long range dependency by enlarging the receptive field.
Learning Lane Graph Representations for Motion
Forecasting
LaneGCN: LaneConv(k1, .., kC) this multi- scale layer.
• In the model, use spatial attention and LaneGCN to capture a complete set of actor-map interactions.
FusionNet:
• Build a stack of four fusion modules to capture all information between actors and lane nodes, i.e.,
actors to lanes (A2L), lanes to lanes (L2L), lanes to actors (L2A) and actors to actors (A2A).
• L2L module by LaneGCN, but other three modules by:
Prediction Header: The header has two branches, a regression branch to predict the trajectory of each
mode and a classification branch to predict the confidence score of each mode.
• regression branch
• classification branch
Learning Lane Graph Representations for Motion
Forecasting
Implicit Latent Variable Model for Scene-
Consistent Motion Forecasting
• In this paper, aim to learn scene-consistent motion forecasts of complex
urban trac directly from sensor data.
• In particular, propose to characterize the joint distribution over future
trajectories via an implicit latent variable model.
• It models the scene as an interaction graph and employ powerful graph
neural networks to learn a distributed latent representation of the scene.
• Coupled with a deterministic decoder, obtain trajectory samples that are
consistent across trac participants, achieving state-of-the-art results in
motion forecasting and interaction understanding.
• This motion forecasts result in safer and more comfortable motion
planning.
Implicit Latent Variable Model for Scene-
Consistent Motion Forecasting
Graphical models of trajectory distribution. Dashed arrows/circles denote that only some approaches within
the group use those components. Double circle in (c) denotes that it is a deterministic mapping of its inputs.
Actor Feature Extraction. Given LiDAR and maps, the backbone CNN detects the actors in the scene, and
individual feature vectors per actor are extracted via RRoI Align, followed by a CNN with spatial pooling.
Implicit Latent Variable Model for Scene-
Consistent Motion Forecasting
Implicit Latent Variable Model encodes the scene into a latent space, from which it can efficiently sample
multiple future realizations in parallel, each with socially consistent trajectories.
Implicit Latent Variable Model for Scene-
Consistent Motion Forecasting
Implicit Latent Variable Model for Scene-
Consistent Motion Forecasting
Perceive, Predict, and Plan: Safe Motion Planning
Through Interpretable Semantic Representations
• In this paper propose an end-to-end learnable network that performs joint
perception, prediction and motion planning for self-driving vehicles and
produces interpretable intermediate representations.
• Unlike existing neural motion planners, this motion planning costs are
consistent with perception and prediction estimates.
• This is achieved by a novel differentiable semantic occupancy representation
that is explicitly used as cost by the motion planning process.
• This network is learned end-to-end from human demonstrations.
• The experiments in a large-scale manual-driving dataset and closed-loop
simulation show that the proposed model significantly outperforms state-of-
the-art planners in imitating the human behaviors while producing much
safer trajectories.
Perceive, Predict, and Plan: Safe Motion Planning
Through Interpretable Semantic Representations
The overview of end-to-end learnable autonomy system that takes raw sensor data, an
HD map and a high level route as input and produces safe maneuvers for the self-
driving vehicle via our novel semantic interpretable intermediate representations.
Perceive, Predict, and Plan: Safe Motion Planning
Through Interpretable Semantic Representations
Semantic classes in occupancy forecasting. Colors match between drawing and hierarchy. Shadowed
area corresponds to the SDV route. Black vehicle, pedestrian and bike icons represent the agents'
true current location.
Perceive, Predict, and Plan: Safe Motion Planning
Through Interpretable Semantic Representations
Inference diagram of the perception and recurrent occupancy forecasting
model. || symbolizes concatenation along the feature dimension, ⊕ element-
wise sum and ∆ bilinear interpolation used to downscale the occupancy.
Perceive, Predict, and Plan: Safe Motion Planning
Through Interpretable Semantic Representations
Examples of the motion planner cost functions: (a) collision, (b) driving-path,
(c) lane boundary, (d) traffic light, (e) comfort, (f) route, (g) progress.
Cost related comfort, traffic rules and progress in the route:
Safe cost:
Perceive, Predict, and Plan: Safe Motion Planning
Through Interpretable Semantic Representations
Learn the model parameters by exploiting these two loss functions:
Semantic Occupancy Loss:
Planning Loss:
Perceive, Predict, and Plan: Safe Motion Planning
Through Interpretable Semantic Representations
Driving behaviors for adas and autonomous driving xiv

More Related Content

What's hot

Deep Learning’s Application in Radar Signal Data II
Deep Learning’s Application in Radar Signal Data IIDeep Learning’s Application in Radar Signal Data II
Deep Learning’s Application in Radar Signal Data IIYu Huang
 
Driving behaviors for adas and autonomous driving XII
Driving behaviors for adas and autonomous driving XIIDriving behaviors for adas and autonomous driving XII
Driving behaviors for adas and autonomous driving XIIYu Huang
 
3-d interpretation from single 2-d image for autonomous driving II
3-d interpretation from single 2-d image for autonomous driving II3-d interpretation from single 2-d image for autonomous driving II
3-d interpretation from single 2-d image for autonomous driving IIYu Huang
 
Multi sensor calibration by deep learning
Multi sensor calibration by deep learningMulti sensor calibration by deep learning
Multi sensor calibration by deep learningYu Huang
 
Depth Fusion from RGB and Depth Sensors IV
Depth Fusion from RGB and Depth Sensors  IVDepth Fusion from RGB and Depth Sensors  IV
Depth Fusion from RGB and Depth Sensors IVYu Huang
 
Pedestrian behavior/intention modeling for autonomous driving V
Pedestrian behavior/intention modeling for autonomous driving VPedestrian behavior/intention modeling for autonomous driving V
Pedestrian behavior/intention modeling for autonomous driving VYu Huang
 
3-d interpretation from single 2-d image IV
3-d interpretation from single 2-d image IV3-d interpretation from single 2-d image IV
3-d interpretation from single 2-d image IVYu Huang
 
Prediction and planning for self driving at waymo
Prediction and planning for self driving at waymoPrediction and planning for self driving at waymo
Prediction and planning for self driving at waymoYu Huang
 
Simulation for autonomous driving at uber atg
Simulation for autonomous driving at uber atgSimulation for autonomous driving at uber atg
Simulation for autonomous driving at uber atgYu Huang
 
Fisheye-Omnidirectional View in Autonomous Driving III
Fisheye-Omnidirectional View in Autonomous Driving IIIFisheye-Omnidirectional View in Autonomous Driving III
Fisheye-Omnidirectional View in Autonomous Driving IIIYu Huang
 
Pedestrian behavior/intention modeling for autonomous driving IV
Pedestrian behavior/intention modeling for autonomous driving IVPedestrian behavior/intention modeling for autonomous driving IV
Pedestrian behavior/intention modeling for autonomous driving IVYu Huang
 
BEV Semantic Segmentation
BEV Semantic SegmentationBEV Semantic Segmentation
BEV Semantic SegmentationYu Huang
 
Driving behaviors for adas and autonomous driving XIII
Driving behaviors for adas and autonomous driving XIIIDriving behaviors for adas and autonomous driving XIII
Driving behaviors for adas and autonomous driving XIIIYu Huang
 
Driving Behavior for ADAS and Autonomous Driving VII
Driving Behavior for ADAS and Autonomous Driving VIIDriving Behavior for ADAS and Autonomous Driving VII
Driving Behavior for ADAS and Autonomous Driving VIIYu Huang
 
Camera-based road Lane detection by deep learning III
Camera-based road Lane detection by deep learning IIICamera-based road Lane detection by deep learning III
Camera-based road Lane detection by deep learning IIIYu Huang
 
camera-based Lane detection by deep learning
camera-based Lane detection by deep learningcamera-based Lane detection by deep learning
camera-based Lane detection by deep learningYu Huang
 
Pedestrian behavior/intention modeling for autonomous driving III
Pedestrian behavior/intention modeling for autonomous driving IIIPedestrian behavior/intention modeling for autonomous driving III
Pedestrian behavior/intention modeling for autonomous driving IIIYu Huang
 
Depth Fusion from RGB and Depth Sensors III
Depth Fusion from RGB and Depth Sensors  IIIDepth Fusion from RGB and Depth Sensors  III
Depth Fusion from RGB and Depth Sensors IIIYu Huang
 
Driving Behavior for ADAS and Autonomous Driving III
Driving Behavior for ADAS and Autonomous Driving IIIDriving Behavior for ADAS and Autonomous Driving III
Driving Behavior for ADAS and Autonomous Driving IIIYu Huang
 
Jointly mapping, localization, perception, prediction and planning
Jointly mapping, localization, perception, prediction and planningJointly mapping, localization, perception, prediction and planning
Jointly mapping, localization, perception, prediction and planningYu Huang
 

What's hot (20)

Deep Learning’s Application in Radar Signal Data II
Deep Learning’s Application in Radar Signal Data IIDeep Learning’s Application in Radar Signal Data II
Deep Learning’s Application in Radar Signal Data II
 
Driving behaviors for adas and autonomous driving XII
Driving behaviors for adas and autonomous driving XIIDriving behaviors for adas and autonomous driving XII
Driving behaviors for adas and autonomous driving XII
 
3-d interpretation from single 2-d image for autonomous driving II
3-d interpretation from single 2-d image for autonomous driving II3-d interpretation from single 2-d image for autonomous driving II
3-d interpretation from single 2-d image for autonomous driving II
 
Multi sensor calibration by deep learning
Multi sensor calibration by deep learningMulti sensor calibration by deep learning
Multi sensor calibration by deep learning
 
Depth Fusion from RGB and Depth Sensors IV
Depth Fusion from RGB and Depth Sensors  IVDepth Fusion from RGB and Depth Sensors  IV
Depth Fusion from RGB and Depth Sensors IV
 
Pedestrian behavior/intention modeling for autonomous driving V
Pedestrian behavior/intention modeling for autonomous driving VPedestrian behavior/intention modeling for autonomous driving V
Pedestrian behavior/intention modeling for autonomous driving V
 
3-d interpretation from single 2-d image IV
3-d interpretation from single 2-d image IV3-d interpretation from single 2-d image IV
3-d interpretation from single 2-d image IV
 
Prediction and planning for self driving at waymo
Prediction and planning for self driving at waymoPrediction and planning for self driving at waymo
Prediction and planning for self driving at waymo
 
Simulation for autonomous driving at uber atg
Simulation for autonomous driving at uber atgSimulation for autonomous driving at uber atg
Simulation for autonomous driving at uber atg
 
Fisheye-Omnidirectional View in Autonomous Driving III
Fisheye-Omnidirectional View in Autonomous Driving IIIFisheye-Omnidirectional View in Autonomous Driving III
Fisheye-Omnidirectional View in Autonomous Driving III
 
Pedestrian behavior/intention modeling for autonomous driving IV
Pedestrian behavior/intention modeling for autonomous driving IVPedestrian behavior/intention modeling for autonomous driving IV
Pedestrian behavior/intention modeling for autonomous driving IV
 
BEV Semantic Segmentation
BEV Semantic SegmentationBEV Semantic Segmentation
BEV Semantic Segmentation
 
Driving behaviors for adas and autonomous driving XIII
Driving behaviors for adas and autonomous driving XIIIDriving behaviors for adas and autonomous driving XIII
Driving behaviors for adas and autonomous driving XIII
 
Driving Behavior for ADAS and Autonomous Driving VII
Driving Behavior for ADAS and Autonomous Driving VIIDriving Behavior for ADAS and Autonomous Driving VII
Driving Behavior for ADAS and Autonomous Driving VII
 
Camera-based road Lane detection by deep learning III
Camera-based road Lane detection by deep learning IIICamera-based road Lane detection by deep learning III
Camera-based road Lane detection by deep learning III
 
camera-based Lane detection by deep learning
camera-based Lane detection by deep learningcamera-based Lane detection by deep learning
camera-based Lane detection by deep learning
 
Pedestrian behavior/intention modeling for autonomous driving III
Pedestrian behavior/intention modeling for autonomous driving IIIPedestrian behavior/intention modeling for autonomous driving III
Pedestrian behavior/intention modeling for autonomous driving III
 
Depth Fusion from RGB and Depth Sensors III
Depth Fusion from RGB and Depth Sensors  IIIDepth Fusion from RGB and Depth Sensors  III
Depth Fusion from RGB and Depth Sensors III
 
Driving Behavior for ADAS and Autonomous Driving III
Driving Behavior for ADAS and Autonomous Driving IIIDriving Behavior for ADAS and Autonomous Driving III
Driving Behavior for ADAS and Autonomous Driving III
 
Jointly mapping, localization, perception, prediction and planning
Jointly mapping, localization, perception, prediction and planningJointly mapping, localization, perception, prediction and planning
Jointly mapping, localization, perception, prediction and planning
 

Similar to Driving behaviors for adas and autonomous driving xiv

Path Planning And Navigation
Path Planning And NavigationPath Planning And Navigation
Path Planning And Navigationguest90654fd
 
Path Planning And Navigation
Path Planning And NavigationPath Planning And Navigation
Path Planning And Navigationguest90654fd
 
Hybrid autonomousnavigation p_limaye-et-al_3pgabstract
Hybrid autonomousnavigation p_limaye-et-al_3pgabstractHybrid autonomousnavigation p_limaye-et-al_3pgabstract
Hybrid autonomousnavigation p_limaye-et-al_3pgabstractPushkar Limaye
 
EE660_Report_YaxinLiu_8448347171
EE660_Report_YaxinLiu_8448347171EE660_Report_YaxinLiu_8448347171
EE660_Report_YaxinLiu_8448347171Yaxin Liu
 
Swarm.Robotics Research Report IEEE
Swarm.Robotics Research Report IEEESwarm.Robotics Research Report IEEE
Swarm.Robotics Research Report IEEEAsad Masood
 
A ROS IMPLEMENTATION OF THE MONO-SLAM ALGORITHM
A ROS IMPLEMENTATION OF THE MONO-SLAM ALGORITHMA ROS IMPLEMENTATION OF THE MONO-SLAM ALGORITHM
A ROS IMPLEMENTATION OF THE MONO-SLAM ALGORITHMcsandit
 
Udacity-Didi Challenge Finalists
Udacity-Didi Challenge FinalistsUdacity-Didi Challenge Finalists
Udacity-Didi Challenge FinalistsDavid Silver
 
Visual odometry & slam utilizing indoor structured environments
Visual odometry & slam utilizing indoor structured environmentsVisual odometry & slam utilizing indoor structured environments
Visual odometry & slam utilizing indoor structured environmentsNAVER Engineering
 
Report bep thomas_blanken
Report bep thomas_blankenReport bep thomas_blanken
Report bep thomas_blankenxepost
 
The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)theijes
 
LiDAR-based Autonomous Driving III (by Deep Learning)
LiDAR-based Autonomous Driving III (by Deep Learning)LiDAR-based Autonomous Driving III (by Deep Learning)
LiDAR-based Autonomous Driving III (by Deep Learning)Yu Huang
 
Automatic Target Recognition Using Recurrent Neural Networks
Automatic Target Recognition Using Recurrent Neural NetworksAutomatic Target Recognition Using Recurrent Neural Networks
Automatic Target Recognition Using Recurrent Neural NetworksAbhishek Jain
 
fusion of Camera and lidar for autonomous driving II
fusion of Camera and lidar for autonomous driving IIfusion of Camera and lidar for autonomous driving II
fusion of Camera and lidar for autonomous driving IIYu Huang
 
IOT-WSN: SURVEY ON POSITIONING TECHNIQUES
IOT-WSN: SURVEY ON POSITIONING TECHNIQUESIOT-WSN: SURVEY ON POSITIONING TECHNIQUES
IOT-WSN: SURVEY ON POSITIONING TECHNIQUESijassn
 
fusion of Camera and lidar for autonomous driving I
fusion of Camera and lidar for autonomous driving Ifusion of Camera and lidar for autonomous driving I
fusion of Camera and lidar for autonomous driving IYu Huang
 

Similar to Driving behaviors for adas and autonomous driving xiv (20)

DICTA 2017 poster
DICTA 2017 posterDICTA 2017 poster
DICTA 2017 poster
 
Path Planning And Navigation
Path Planning And NavigationPath Planning And Navigation
Path Planning And Navigation
 
Path Planning And Navigation
Path Planning And NavigationPath Planning And Navigation
Path Planning And Navigation
 
Hybrid autonomousnavigation p_limaye-et-al_3pgabstract
Hybrid autonomousnavigation p_limaye-et-al_3pgabstractHybrid autonomousnavigation p_limaye-et-al_3pgabstract
Hybrid autonomousnavigation p_limaye-et-al_3pgabstract
 
EE660_Report_YaxinLiu_8448347171
EE660_Report_YaxinLiu_8448347171EE660_Report_YaxinLiu_8448347171
EE660_Report_YaxinLiu_8448347171
 
Swarm.Robotics Research Report IEEE
Swarm.Robotics Research Report IEEESwarm.Robotics Research Report IEEE
Swarm.Robotics Research Report IEEE
 
Robotics Localization
Robotics LocalizationRobotics Localization
Robotics Localization
 
A ROS IMPLEMENTATION OF THE MONO-SLAM ALGORITHM
A ROS IMPLEMENTATION OF THE MONO-SLAM ALGORITHMA ROS IMPLEMENTATION OF THE MONO-SLAM ALGORITHM
A ROS IMPLEMENTATION OF THE MONO-SLAM ALGORITHM
 
Udacity-Didi Challenge Finalists
Udacity-Didi Challenge FinalistsUdacity-Didi Challenge Finalists
Udacity-Didi Challenge Finalists
 
All projects
All projectsAll projects
All projects
 
Visual odometry & slam utilizing indoor structured environments
Visual odometry & slam utilizing indoor structured environmentsVisual odometry & slam utilizing indoor structured environments
Visual odometry & slam utilizing indoor structured environments
 
Report bep thomas_blanken
Report bep thomas_blankenReport bep thomas_blanken
Report bep thomas_blanken
 
The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)
 
LiDAR-based Autonomous Driving III (by Deep Learning)
LiDAR-based Autonomous Driving III (by Deep Learning)LiDAR-based Autonomous Driving III (by Deep Learning)
LiDAR-based Autonomous Driving III (by Deep Learning)
 
Automatic Target Recognition Using Recurrent Neural Networks
Automatic Target Recognition Using Recurrent Neural NetworksAutomatic Target Recognition Using Recurrent Neural Networks
Automatic Target Recognition Using Recurrent Neural Networks
 
fusion of Camera and lidar for autonomous driving II
fusion of Camera and lidar for autonomous driving IIfusion of Camera and lidar for autonomous driving II
fusion of Camera and lidar for autonomous driving II
 
IOT-WSN: SURVEY ON POSITIONING TECHNIQUES
IOT-WSN: SURVEY ON POSITIONING TECHNIQUESIOT-WSN: SURVEY ON POSITIONING TECHNIQUES
IOT-WSN: SURVEY ON POSITIONING TECHNIQUES
 
C1802022430
C1802022430C1802022430
C1802022430
 
fusion of Camera and lidar for autonomous driving I
fusion of Camera and lidar for autonomous driving Ifusion of Camera and lidar for autonomous driving I
fusion of Camera and lidar for autonomous driving I
 
Where Next
Where NextWhere Next
Where Next
 

More from Yu Huang

Application of Foundation Model for Autonomous Driving
Application of Foundation Model for Autonomous DrivingApplication of Foundation Model for Autonomous Driving
Application of Foundation Model for Autonomous DrivingYu Huang
 
The New Perception Framework in Autonomous Driving: An Introduction of BEV N...
The New Perception Framework  in Autonomous Driving: An Introduction of BEV N...The New Perception Framework  in Autonomous Driving: An Introduction of BEV N...
The New Perception Framework in Autonomous Driving: An Introduction of BEV N...Yu Huang
 
Data Closed Loop in Simulation Test of Autonomous Driving
Data Closed Loop in Simulation Test of Autonomous DrivingData Closed Loop in Simulation Test of Autonomous Driving
Data Closed Loop in Simulation Test of Autonomous DrivingYu Huang
 
Techniques and Challenges in Autonomous Driving
Techniques and Challenges in Autonomous DrivingTechniques and Challenges in Autonomous Driving
Techniques and Challenges in Autonomous DrivingYu Huang
 
BEV Joint Detection and Segmentation
BEV Joint Detection and SegmentationBEV Joint Detection and Segmentation
BEV Joint Detection and SegmentationYu Huang
 
BEV Object Detection and Prediction
BEV Object Detection and PredictionBEV Object Detection and Prediction
BEV Object Detection and PredictionYu Huang
 
Fisheye based Perception for Autonomous Driving VI
Fisheye based Perception for Autonomous Driving VIFisheye based Perception for Autonomous Driving VI
Fisheye based Perception for Autonomous Driving VIYu Huang
 
Fisheye/Omnidirectional View in Autonomous Driving V
Fisheye/Omnidirectional View in Autonomous Driving VFisheye/Omnidirectional View in Autonomous Driving V
Fisheye/Omnidirectional View in Autonomous Driving VYu Huang
 
Fisheye/Omnidirectional View in Autonomous Driving IV
Fisheye/Omnidirectional View in Autonomous Driving IVFisheye/Omnidirectional View in Autonomous Driving IV
Fisheye/Omnidirectional View in Autonomous Driving IVYu Huang
 
Prediction,Planninng & Control at Baidu
Prediction,Planninng & Control at BaiduPrediction,Planninng & Control at Baidu
Prediction,Planninng & Control at BaiduYu Huang
 
Cruise AI under the Hood
Cruise AI under the HoodCruise AI under the Hood
Cruise AI under the HoodYu Huang
 
LiDAR in the Adverse Weather: Dust, Snow, Rain and Fog (2)
LiDAR in the Adverse Weather: Dust, Snow, Rain and Fog (2)LiDAR in the Adverse Weather: Dust, Snow, Rain and Fog (2)
LiDAR in the Adverse Weather: Dust, Snow, Rain and Fog (2)Yu Huang
 
Scenario-Based Development & Testing for Autonomous Driving
Scenario-Based Development & Testing for Autonomous DrivingScenario-Based Development & Testing for Autonomous Driving
Scenario-Based Development & Testing for Autonomous DrivingYu Huang
 
How to Build a Data Closed-loop Platform for Autonomous Driving?
How to Build a Data Closed-loop Platform for Autonomous Driving?How to Build a Data Closed-loop Platform for Autonomous Driving?
How to Build a Data Closed-loop Platform for Autonomous Driving?Yu Huang
 
Annotation tools for ADAS & Autonomous Driving
Annotation tools for ADAS & Autonomous DrivingAnnotation tools for ADAS & Autonomous Driving
Annotation tools for ADAS & Autonomous DrivingYu Huang
 
Data pipeline and data lake for autonomous driving
Data pipeline and data lake for autonomous drivingData pipeline and data lake for autonomous driving
Data pipeline and data lake for autonomous drivingYu Huang
 
Open Source codes of trajectory prediction & behavior planning
Open Source codes of trajectory prediction & behavior planningOpen Source codes of trajectory prediction & behavior planning
Open Source codes of trajectory prediction & behavior planningYu Huang
 
Lidar in the adverse weather: dust, fog, snow and rain
Lidar in the adverse weather: dust, fog, snow and rainLidar in the adverse weather: dust, fog, snow and rain
Lidar in the adverse weather: dust, fog, snow and rainYu Huang
 
Autonomous Driving of L3/L4 Commercial trucks
Autonomous Driving of L3/L4 Commercial trucksAutonomous Driving of L3/L4 Commercial trucks
Autonomous Driving of L3/L4 Commercial trucksYu Huang
 
3-d interpretation from single 2-d image V
3-d interpretation from single 2-d image V3-d interpretation from single 2-d image V
3-d interpretation from single 2-d image VYu Huang
 

More from Yu Huang (20)

Application of Foundation Model for Autonomous Driving
Application of Foundation Model for Autonomous DrivingApplication of Foundation Model for Autonomous Driving
Application of Foundation Model for Autonomous Driving
 
The New Perception Framework in Autonomous Driving: An Introduction of BEV N...
The New Perception Framework  in Autonomous Driving: An Introduction of BEV N...The New Perception Framework  in Autonomous Driving: An Introduction of BEV N...
The New Perception Framework in Autonomous Driving: An Introduction of BEV N...
 
Data Closed Loop in Simulation Test of Autonomous Driving
Data Closed Loop in Simulation Test of Autonomous DrivingData Closed Loop in Simulation Test of Autonomous Driving
Data Closed Loop in Simulation Test of Autonomous Driving
 
Techniques and Challenges in Autonomous Driving
Techniques and Challenges in Autonomous DrivingTechniques and Challenges in Autonomous Driving
Techniques and Challenges in Autonomous Driving
 
BEV Joint Detection and Segmentation
BEV Joint Detection and SegmentationBEV Joint Detection and Segmentation
BEV Joint Detection and Segmentation
 
BEV Object Detection and Prediction
BEV Object Detection and PredictionBEV Object Detection and Prediction
BEV Object Detection and Prediction
 
Fisheye based Perception for Autonomous Driving VI
Fisheye based Perception for Autonomous Driving VIFisheye based Perception for Autonomous Driving VI
Fisheye based Perception for Autonomous Driving VI
 
Fisheye/Omnidirectional View in Autonomous Driving V
Fisheye/Omnidirectional View in Autonomous Driving VFisheye/Omnidirectional View in Autonomous Driving V
Fisheye/Omnidirectional View in Autonomous Driving V
 
Fisheye/Omnidirectional View in Autonomous Driving IV
Fisheye/Omnidirectional View in Autonomous Driving IVFisheye/Omnidirectional View in Autonomous Driving IV
Fisheye/Omnidirectional View in Autonomous Driving IV
 
Prediction,Planninng & Control at Baidu
Prediction,Planninng & Control at BaiduPrediction,Planninng & Control at Baidu
Prediction,Planninng & Control at Baidu
 
Cruise AI under the Hood
Cruise AI under the HoodCruise AI under the Hood
Cruise AI under the Hood
 
LiDAR in the Adverse Weather: Dust, Snow, Rain and Fog (2)
LiDAR in the Adverse Weather: Dust, Snow, Rain and Fog (2)LiDAR in the Adverse Weather: Dust, Snow, Rain and Fog (2)
LiDAR in the Adverse Weather: Dust, Snow, Rain and Fog (2)
 
Scenario-Based Development & Testing for Autonomous Driving
Scenario-Based Development & Testing for Autonomous DrivingScenario-Based Development & Testing for Autonomous Driving
Scenario-Based Development & Testing for Autonomous Driving
 
How to Build a Data Closed-loop Platform for Autonomous Driving?
How to Build a Data Closed-loop Platform for Autonomous Driving?How to Build a Data Closed-loop Platform for Autonomous Driving?
How to Build a Data Closed-loop Platform for Autonomous Driving?
 
Annotation tools for ADAS & Autonomous Driving
Annotation tools for ADAS & Autonomous DrivingAnnotation tools for ADAS & Autonomous Driving
Annotation tools for ADAS & Autonomous Driving
 
Data pipeline and data lake for autonomous driving
Data pipeline and data lake for autonomous drivingData pipeline and data lake for autonomous driving
Data pipeline and data lake for autonomous driving
 
Open Source codes of trajectory prediction & behavior planning
Open Source codes of trajectory prediction & behavior planningOpen Source codes of trajectory prediction & behavior planning
Open Source codes of trajectory prediction & behavior planning
 
Lidar in the adverse weather: dust, fog, snow and rain
Lidar in the adverse weather: dust, fog, snow and rainLidar in the adverse weather: dust, fog, snow and rain
Lidar in the adverse weather: dust, fog, snow and rain
 
Autonomous Driving of L3/L4 Commercial trucks
Autonomous Driving of L3/L4 Commercial trucksAutonomous Driving of L3/L4 Commercial trucks
Autonomous Driving of L3/L4 Commercial trucks
 
3-d interpretation from single 2-d image V
3-d interpretation from single 2-d image V3-d interpretation from single 2-d image V
3-d interpretation from single 2-d image V
 

Recently uploaded

Processing & Properties of Floor and Wall Tiles.pptx
Processing & Properties of Floor and Wall Tiles.pptxProcessing & Properties of Floor and Wall Tiles.pptx
Processing & Properties of Floor and Wall Tiles.pptxpranjaldaimarysona
 
IVE Industry Focused Event - Defence Sector 2024
IVE Industry Focused Event - Defence Sector 2024IVE Industry Focused Event - Defence Sector 2024
IVE Industry Focused Event - Defence Sector 2024Mark Billinghurst
 
Call Girls in Nagpur Suman Call 7001035870 Meet With Nagpur Escorts
Call Girls in Nagpur Suman Call 7001035870 Meet With Nagpur EscortsCall Girls in Nagpur Suman Call 7001035870 Meet With Nagpur Escorts
Call Girls in Nagpur Suman Call 7001035870 Meet With Nagpur EscortsCall Girls in Nagpur High Profile
 
MANUFACTURING PROCESS-II UNIT-2 LATHE MACHINE
MANUFACTURING PROCESS-II UNIT-2 LATHE MACHINEMANUFACTURING PROCESS-II UNIT-2 LATHE MACHINE
MANUFACTURING PROCESS-II UNIT-2 LATHE MACHINESIVASHANKAR N
 
College Call Girls Nashik Nehal 7001305949 Independent Escort Service Nashik
College Call Girls Nashik Nehal 7001305949 Independent Escort Service NashikCollege Call Girls Nashik Nehal 7001305949 Independent Escort Service Nashik
College Call Girls Nashik Nehal 7001305949 Independent Escort Service NashikCall Girls in Nagpur High Profile
 
VIP Call Girls Service Hitech City Hyderabad Call +91-8250192130
VIP Call Girls Service Hitech City Hyderabad Call +91-8250192130VIP Call Girls Service Hitech City Hyderabad Call +91-8250192130
VIP Call Girls Service Hitech City Hyderabad Call +91-8250192130Suhani Kapoor
 
Sheet Pile Wall Design and Construction: A Practical Guide for Civil Engineer...
Sheet Pile Wall Design and Construction: A Practical Guide for Civil Engineer...Sheet Pile Wall Design and Construction: A Practical Guide for Civil Engineer...
Sheet Pile Wall Design and Construction: A Practical Guide for Civil Engineer...Dr.Costas Sachpazis
 
HARMONY IN THE NATURE AND EXISTENCE - Unit-IV
HARMONY IN THE NATURE AND EXISTENCE - Unit-IVHARMONY IN THE NATURE AND EXISTENCE - Unit-IV
HARMONY IN THE NATURE AND EXISTENCE - Unit-IVRajaP95
 
Porous Ceramics seminar and technical writing
Porous Ceramics seminar and technical writingPorous Ceramics seminar and technical writing
Porous Ceramics seminar and technical writingrakeshbaidya232001
 
Architect Hassan Khalil Portfolio for 2024
Architect Hassan Khalil Portfolio for 2024Architect Hassan Khalil Portfolio for 2024
Architect Hassan Khalil Portfolio for 2024hassan khalil
 
(ANJALI) Dange Chowk Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
(ANJALI) Dange Chowk Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...(ANJALI) Dange Chowk Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
(ANJALI) Dange Chowk Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...ranjana rawat
 
Introduction and different types of Ethernet.pptx
Introduction and different types of Ethernet.pptxIntroduction and different types of Ethernet.pptx
Introduction and different types of Ethernet.pptxupamatechverse
 
VIP Call Girls Service Kondapur Hyderabad Call +91-8250192130
VIP Call Girls Service Kondapur Hyderabad Call +91-8250192130VIP Call Girls Service Kondapur Hyderabad Call +91-8250192130
VIP Call Girls Service Kondapur Hyderabad Call +91-8250192130Suhani Kapoor
 
GDSC ASEB Gen AI study jams presentation
GDSC ASEB Gen AI study jams presentationGDSC ASEB Gen AI study jams presentation
GDSC ASEB Gen AI study jams presentationGDSCAESB
 
MANUFACTURING PROCESS-II UNIT-5 NC MACHINE TOOLS
MANUFACTURING PROCESS-II UNIT-5 NC MACHINE TOOLSMANUFACTURING PROCESS-II UNIT-5 NC MACHINE TOOLS
MANUFACTURING PROCESS-II UNIT-5 NC MACHINE TOOLSSIVASHANKAR N
 
Software Development Life Cycle By Team Orange (Dept. of Pharmacy)
Software Development Life Cycle By  Team Orange (Dept. of Pharmacy)Software Development Life Cycle By  Team Orange (Dept. of Pharmacy)
Software Development Life Cycle By Team Orange (Dept. of Pharmacy)Suman Mia
 
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptx
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptxDecoding Kotlin - Your guide to solving the mysterious in Kotlin.pptx
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptxJoão Esperancinha
 
Call Girls Delhi {Jodhpur} 9711199012 high profile service
Call Girls Delhi {Jodhpur} 9711199012 high profile serviceCall Girls Delhi {Jodhpur} 9711199012 high profile service
Call Girls Delhi {Jodhpur} 9711199012 high profile servicerehmti665
 
Biology for Computer Engineers Course Handout.pptx
Biology for Computer Engineers Course Handout.pptxBiology for Computer Engineers Course Handout.pptx
Biology for Computer Engineers Course Handout.pptxDeepakSakkari2
 

Recently uploaded (20)

Processing & Properties of Floor and Wall Tiles.pptx
Processing & Properties of Floor and Wall Tiles.pptxProcessing & Properties of Floor and Wall Tiles.pptx
Processing & Properties of Floor and Wall Tiles.pptx
 
IVE Industry Focused Event - Defence Sector 2024
IVE Industry Focused Event - Defence Sector 2024IVE Industry Focused Event - Defence Sector 2024
IVE Industry Focused Event - Defence Sector 2024
 
Call Girls in Nagpur Suman Call 7001035870 Meet With Nagpur Escorts
Call Girls in Nagpur Suman Call 7001035870 Meet With Nagpur EscortsCall Girls in Nagpur Suman Call 7001035870 Meet With Nagpur Escorts
Call Girls in Nagpur Suman Call 7001035870 Meet With Nagpur Escorts
 
MANUFACTURING PROCESS-II UNIT-2 LATHE MACHINE
MANUFACTURING PROCESS-II UNIT-2 LATHE MACHINEMANUFACTURING PROCESS-II UNIT-2 LATHE MACHINE
MANUFACTURING PROCESS-II UNIT-2 LATHE MACHINE
 
College Call Girls Nashik Nehal 7001305949 Independent Escort Service Nashik
College Call Girls Nashik Nehal 7001305949 Independent Escort Service NashikCollege Call Girls Nashik Nehal 7001305949 Independent Escort Service Nashik
College Call Girls Nashik Nehal 7001305949 Independent Escort Service Nashik
 
VIP Call Girls Service Hitech City Hyderabad Call +91-8250192130
VIP Call Girls Service Hitech City Hyderabad Call +91-8250192130VIP Call Girls Service Hitech City Hyderabad Call +91-8250192130
VIP Call Girls Service Hitech City Hyderabad Call +91-8250192130
 
Sheet Pile Wall Design and Construction: A Practical Guide for Civil Engineer...
Sheet Pile Wall Design and Construction: A Practical Guide for Civil Engineer...Sheet Pile Wall Design and Construction: A Practical Guide for Civil Engineer...
Sheet Pile Wall Design and Construction: A Practical Guide for Civil Engineer...
 
HARMONY IN THE NATURE AND EXISTENCE - Unit-IV
HARMONY IN THE NATURE AND EXISTENCE - Unit-IVHARMONY IN THE NATURE AND EXISTENCE - Unit-IV
HARMONY IN THE NATURE AND EXISTENCE - Unit-IV
 
Porous Ceramics seminar and technical writing
Porous Ceramics seminar and technical writingPorous Ceramics seminar and technical writing
Porous Ceramics seminar and technical writing
 
★ CALL US 9953330565 ( HOT Young Call Girls In Badarpur delhi NCR
★ CALL US 9953330565 ( HOT Young Call Girls In Badarpur delhi NCR★ CALL US 9953330565 ( HOT Young Call Girls In Badarpur delhi NCR
★ CALL US 9953330565 ( HOT Young Call Girls In Badarpur delhi NCR
 
Architect Hassan Khalil Portfolio for 2024
Architect Hassan Khalil Portfolio for 2024Architect Hassan Khalil Portfolio for 2024
Architect Hassan Khalil Portfolio for 2024
 
(ANJALI) Dange Chowk Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
(ANJALI) Dange Chowk Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...(ANJALI) Dange Chowk Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
(ANJALI) Dange Chowk Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
 
Introduction and different types of Ethernet.pptx
Introduction and different types of Ethernet.pptxIntroduction and different types of Ethernet.pptx
Introduction and different types of Ethernet.pptx
 
VIP Call Girls Service Kondapur Hyderabad Call +91-8250192130
VIP Call Girls Service Kondapur Hyderabad Call +91-8250192130VIP Call Girls Service Kondapur Hyderabad Call +91-8250192130
VIP Call Girls Service Kondapur Hyderabad Call +91-8250192130
 
GDSC ASEB Gen AI study jams presentation
GDSC ASEB Gen AI study jams presentationGDSC ASEB Gen AI study jams presentation
GDSC ASEB Gen AI study jams presentation
 
MANUFACTURING PROCESS-II UNIT-5 NC MACHINE TOOLS
MANUFACTURING PROCESS-II UNIT-5 NC MACHINE TOOLSMANUFACTURING PROCESS-II UNIT-5 NC MACHINE TOOLS
MANUFACTURING PROCESS-II UNIT-5 NC MACHINE TOOLS
 
Software Development Life Cycle By Team Orange (Dept. of Pharmacy)
Software Development Life Cycle By  Team Orange (Dept. of Pharmacy)Software Development Life Cycle By  Team Orange (Dept. of Pharmacy)
Software Development Life Cycle By Team Orange (Dept. of Pharmacy)
 
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptx
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptxDecoding Kotlin - Your guide to solving the mysterious in Kotlin.pptx
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptx
 
Call Girls Delhi {Jodhpur} 9711199012 high profile service
Call Girls Delhi {Jodhpur} 9711199012 high profile serviceCall Girls Delhi {Jodhpur} 9711199012 high profile service
Call Girls Delhi {Jodhpur} 9711199012 high profile service
 
Biology for Computer Engineers Course Handout.pptx
Biology for Computer Engineers Course Handout.pptxBiology for Computer Engineers Course Handout.pptx
Biology for Computer Engineers Course Handout.pptx
 

Driving behaviors for adas and autonomous driving xiv

  • 1. Driving Behaviors for ADAS and Autonomous Driving XIV Yu Huang Yu.huang07@gmail.com Sunnyvale, California
  • 2. Outline • TNT: Target-driveN Trajectory Prediction (CVPR’20) • Driving Through Ghosts: Behavioral Cloning with False Positives (8’29) • LiRaNet: E2E Trajectory Prediction using Spatio-Temporal Radar Fusion (10.15) • SimAug: Learning Robust Representations from Simulation for Trajectory Prediction (ECCV’20) • Learning Lane Graph Representations for Motion Forecasting (ECCV’20) • Implicit Latent Variable Model for Scene-Consistent Motion Forecasting (ECCV’20) • Perceive, Predict, and Plan: Safe Motion Planning Through Interpretable Semantic Representations (ECCV’20)
  • 3. TNT: Target-driveN Trajectory Prediction • This key insight is that for prediction within a moderate time horizon, the future modes can be effectively captured by a set of target states. • This leads to target-driven trajectory prediction (TNT) framework. • TNT has three stages which are trained end-to-end. • It first predicts an agent’s potential target states T steps into the future, by encoding its interactions with the environment and the other agents. • TNT then generates trajectory state sequences conditioned on targets. • A final stage estimates trajectory likelihoods and a final compact set of trajectory predictions is selected. • This is in contrast to previous work which models agent intents as latent variables, and relies on test-time sampling to generate diverse trajectories. • Benchmark TNT on trajectory prediction of vehicles and pedestrians, outperform state-of-the-art on Argoverse Forecasting, INTERACTION, Stanford Drone and an in-house Pedestrian-at-Intersection dataset.
  • 4. TNT: Target-driveN Trajectory Prediction Illustration of the TNT framework when applied to the vehicle future trajectory prediction task. TNT consists of three stages: (a) target prediction which proposes a set of plausible targets (stars) among all candidates (diamonds). (b) target-conditioned motion estimation which estimates a trajectory (distribution) towards each selected target, (c) scoring and selection which ranks trajectory hypotheses and selects a final set of trajectory predictions with likelihood scores.
  • 5. TNT: Target-driveN Trajectory Prediction TNT model overview. Scene context is first encoded as the model’s inputs. Then follows the core three stages of TNT: (a) target prediction which proposes an initial set of M targets; (b) target- conditioned motion estimation which estimates a trajectory for each target; (c) scoring and selection which ranks trajectory hypotheses and outputs a final set of K predicted trajectories.
  • 6. TNT: Target-driveN Trajectory Prediction TNT supports flexible choices of targets. Vehicle target candidate points are sampled from the lane centerlines. Pedestrian target candidate points are sampled from a virtual grid centered on the pedestrian.
  • 8. Driving Through Ghosts: Behavioral Cloning with False Positives • In the context of behavioral cloning, perceptual errors at training time can lead to learning difficulties or wrong policies, as expert demonstrations might be inconsistent with the perceived world state. • In this work, propose a behavioral cloning approach that can safely leverage imperfect perception without being conservative. • The core is a representation of perceptual uncertainty for learning to plan. • It propose a new probabilistic birds-eye-view semantic grid to encode the noisy output of object perception systems. • Then leverage expert demonstrations to learn an imitative driving policy using this probabilistic representation. • Using the CARLA simulator, it can safely overcome critical false positives that would otherwise lead to catastrophic failures or conservative behavior.
  • 9. Driving Through Ghosts: Behavioral Cloning with False Positives It is a probabilistic birds-eye-view semantic representation, Soft BEV, for imitation learning under perceptual uncertainty. It enables learning safer policies that can ignore false positives.
  • 10. Driving Through Ghosts: Behavioral Cloning with False Positives • The observations o in a birds-eye-view grid, i.e., an NxMxD-dim. tensor where each dimension k represents a category of estimated state (e.g., an object or feature type) together with the respective estimated confidences. • Each slice is a matrix of NxM, where each element corresponds to the presence of an estimated object or feature of type k at that location, weighted by its estimated confidence. • The resulting input representation is referred as the Soft BEV. • It models a driving agent via a deep convolutional policy network with input of Soft BEV. • The CNN outputs way-points along the future trajectory, used by a PID controller to compute the control signals for the steering and throttle of the vehicle. • It consists of a ResNet-18 base network acting as an encoder, followed by three deconvolutional layers which also have as an input the current speed signal. • For each of the potential high-level commands (“go left”, “go right”, “go straight”, “follow the road”), the network predicts multiple output heat-maps which are then converted into way-points by spatial soft-argmax layers. • Based on the high-level command, the respective head is used to predict the way-points.
  • 11. Driving Through Ghosts: Behavioral Cloning with False Positives Experimental setup: The CARLA simulator provides ground truth features. Perception noise is applied to the dynamic features, which are then fused into an uncertainty- scaled birds-eye view representation, the Soft BEV. Together with high-level commands and speed information it is fed to a CNN that predicts way-points.
  • 12. Driving Through Ghosts: Behavioral Cloning with False Positives
  • 13. LiRaNet: E2E Trajectory Prediction using Spatio- Temporal Radar Fusion • LiRaNet, a end-to-end trajectory prediction method which utilizes radar sensor information along with widely used lidar and high definition (HD) maps. • Automotive radar provides rich, complementary information, allowing for longer range vehicle detection as well as instantaneous radial velocity measurements. • However, there are factors that make the fusion of lidar and radar information challenging, such as the relatively low angular resolution of radar measurements, their sparsity and the lack of exact time synchronization with lidar. • To overcome these challenges, propose an efficient spatio-temporal radar feature extraction scheme which achieves state-of-the-art performance on multiple large-scale datasets.
  • 14. LiRaNet: E2E Trajectory Prediction using Spatio- Temporal Radar Fusion An example scene from X17k in bird’s eye view where lidar points (light blue) and radar point velocities (orange) are visualized with labels (white) for current, past and future frames. Vehicle A is a turning bus that has multiple radar points across frames. By effectively combining them over space and time a full 2D velocity and turning rate can be recovered. Vehicle B shows the high positional noise that inherently comes with radar. Vehicle C shows a case with sparse lidar points where implicitly associating them across time can be challenging. However, radar points present around C can add context for the model to detect and predict the trajectory.
  • 15. LiRaNet: E2E Trajectory Prediction using Spatio- Temporal Radar Fusion LiRaNet overview: The radar feature extraction network (A) extracts spatio-temporal features from raw radar points in 2 steps: (1) for each frame we create a graph between the BEV grid cells and radar points to learn spatial features of each cell using a non-rigid convolution, (2) these spatial features are further fused temporally by stacking across channel dimension and using an MLP to get a radar feature volume. This feature volume is then fused with feature volumes from other sensors and fed to a joint perception-prediction network (B) which produces detections and their future trajectories. An example prediction for a scene from X17k can be seen in (C).
  • 16. LiRaNet: E2E Trajectory Prediction using Spatio- Temporal Radar Fusion • Input domain consists of radar points and output domain consists of BEV cells. • For each cell j, calculate the features hm j for the sweep m as • where Am j is the set of associated radar points, xm i is the 2D coordinates of the associated radar point, xm j is the 2D coordinate of the BEV cell’s center, ⊕ denotes the concatenation operation, fm i is the feature vector for the radar point and gm() is an multi- layer perceptron (MLP) with learnable weights shared across all the cells. • Calculate Am j, using nearest neighbor algorithm with a distance threshold. • By using a threshold larger than the size of a cell, this method compensates for positional errors in radar. • For each cell j, calculate the final spatio-temporal feature vector (hj ) by concatenating the per sweep features hj,m and using an MLP to combine them.
  • 17. LiRaNet: E2E Trajectory Prediction using Spatio- Temporal Radar Fusion RadarisfusedwithLiDARandMaptogenerate multi-sensorfeaturevolumeinBEVwhichisused fordetectionandtrajectoryprediction.Each learnableoperatorisdenotedas(op,kernel, stride,numberoffeatures).Convdenotes convolutionandResdenotesaresidualblock.
  • 18. LiRaNet: E2E Trajectory Prediction using Spatio- Temporal Radar Fusion
  • 19. SimAug: Learning Robust Representations from Simulation for Trajectory Prediction • This paper studies the problem of predicting future trajectories of people in unseen cameras of novel scenarios and views. • It approaches this problem through the real-data-free setting in which the model is trained only on 3D simulation data and applied out-of-the-box to a wide variety of real cameras. • It proposes to learn robust representation through augmenting the simulation training data such that the representation can better generalize to unseen real- world test data. • The key idea is to mix the feature of the hardest camera view with the adversarial feature of the original view. • It is referred as SimAug, achieving results on three real-world benchmarks using zero real training data, and state-of-the-art performance in the Stanford Drone and the VIRAT/ActEV dataset when using in-domain training data. • Code and models are released at https://next.cs.cmu.edu/simaug.
  • 20. SimAug: Learning Robust Representations from Simulation for Trajectory Prediction SimAug that is trained on simulation and tested on real unseen videos. Each training trajectory is represented by multi-view segmentation features extracted from the simulator. SimAug mixes the feature of the hardest camera view with the adversarial feature of the original view.
  • 21. SimAug: Learning Robust Representations from Simulation for Trajectory Prediction • Each time given a camera view, use it as an anchor to search for the “hardest" view that is most inconsistent with what the model has learned. • It uses the classification loss as the criteria and compute: • For the original view, generate an adversarial trajectory by the targeted-FGSM attack • The attack tries to make the model predict the future locations in the selected “hardest" camera view rather than the original view. • In essence, the resulting adversarial feature is “warped" to the “hardest” camera view by a small perturbation. • By defending against such adversarial trajectory, the model learns representations that are robust against variances in camera views.
  • 22. SimAug: Learning Robust Representations from Simulation for Trajectory Prediction • It mix up the trajectory locations of the selected view and the adversarial trajectory locations by a convex combination function over their features and one-hot location labels. • where [yh+1, …, yT ] = Lh+1:T are the ground-truth locations of the original view. • The one-hot (·) function projects the location in xy coordinates into an one-hot embedding over the predefined grid used in our backbone trajectory prediction model.
  • 23. SimAug: Learning Robust Representations from Simulation for Trajectory Prediction • Backbone network is the Multiverse model (CVPR’20). • The training algorithm’s pseudo-code is:
  • 24. SimAug: Learning Robust Representations from Simulation for Trajectory Prediction
  • 25. SimAug: Learning Robust Representations from Simulation for Trajectory Prediction
  • 26. Learning Lane Graph Representations for Motion Forecasting • A motion forecasting model that exploits a novel structured map representation as well as actor-map interactions. • Instead of encoding vectorized maps as raster images, construct a lane graph from raw map data to explicitly preserve the map structure. • To capture the complex topology and long range dependencies of the lane graph, propose LaneGCN which extends graph convolutions with multiple adjacency matrices and along-lane dilation. • To capture the complex interactions between actors and maps, exploit a fusion network consisting of four types of interactions, actor-to-lane, lane-to-lane, lane- to-actor and actor-to-actor. • Powered by LaneGCN and actor-map interactions, the model is able to predict accurate and realistic multi-modal trajectories. • This approach significantly outperforms the state-of-the-art on the large scale Argoverse motion forecasting benchmark.
  • 27. Learning Lane Graph Representations for Motion Forecasting It constructs a lane graph from raw map data and use LaneGCN to extract map features. In parallel, ActorNet extracts actor features from observed past trajectories. Then it uses FusionNet to model the Interactions between actors themselves and the map, and predict the future trajectories.
  • 28. Learning Lane Graph Representations for Motion Forecasting The model is composed of four modules. (1) ActorNet receives the past actor trajectories as input, and uses 1D convolution to extract actor node features. (2) MapNet constructs a lane graph from HD maps, and uses a LaneGCN to exact lane node features. (3) FusionNet is a stack of 4 interaction blocks. The actor to lane block fuses real-time traffic information from actor nodes to lane nodes. The lane to lane block propagates information over the lane graph and updates lane features. The lane to actor block fuses updated map information from lane nodes to actor nodes. The actor to actor block performs interactions among actors. It uses another LaneGCN for the lane to lane block, and spatial attention layers for the other blocks. (4) The prediction header uses after-fusion actor features to produce multi-modal trajectories.
  • 29. Learning Lane Graph Representations for Motion Forecasting LaneGCN is a stack of 4 multi-scale LaneConv residual blocks, each of which consists of a LaneConv (1, 2, 4, 8, 16, 32) and a linear layer with a residual connection. All layers have 128 feature channels. Left: The lane centerline of interest, its predecessor, successor, left and right neighbor are denoted with red, orange, blue, purple, and green lines, respectively. Each centerline is given as a sequence of BEV points (hollow circles). Right: Derived lane graph with an example lane node. The lane node of interest, its predecessor, successor, left and right neighbor are denoted with red, orange, blue, purple and green circles respectively.
  • 30. Learning Lane Graph Representations for Motion Forecasting LaneConv Operator: parameterize the node feature as follows, LaneConv operator as Ai and Wi are the adjacency and the weight matrices corresponding to the i-th connection type respectively. MLP indicates a multi-layer perceptron and the two subscripts refer to shape and location, respectively. k-dilation LaneConv operator as Ak pre is the k-th matrix power of Apre. In regular grid graphs, a dilated convolution operator can effectively capture the long range dependency by enlarging the receptive field.
  • 31. Learning Lane Graph Representations for Motion Forecasting LaneGCN: LaneConv(k1, .., kC) this multi- scale layer. • In the model, use spatial attention and LaneGCN to capture a complete set of actor-map interactions. FusionNet: • Build a stack of four fusion modules to capture all information between actors and lane nodes, i.e., actors to lanes (A2L), lanes to lanes (L2L), lanes to actors (L2A) and actors to actors (A2A). • L2L module by LaneGCN, but other three modules by: Prediction Header: The header has two branches, a regression branch to predict the trajectory of each mode and a classification branch to predict the confidence score of each mode. • regression branch • classification branch
  • 32. Learning Lane Graph Representations for Motion Forecasting
  • 33. Implicit Latent Variable Model for Scene- Consistent Motion Forecasting • In this paper, aim to learn scene-consistent motion forecasts of complex urban trac directly from sensor data. • In particular, propose to characterize the joint distribution over future trajectories via an implicit latent variable model. • It models the scene as an interaction graph and employ powerful graph neural networks to learn a distributed latent representation of the scene. • Coupled with a deterministic decoder, obtain trajectory samples that are consistent across trac participants, achieving state-of-the-art results in motion forecasting and interaction understanding. • This motion forecasts result in safer and more comfortable motion planning.
  • 34. Implicit Latent Variable Model for Scene- Consistent Motion Forecasting Graphical models of trajectory distribution. Dashed arrows/circles denote that only some approaches within the group use those components. Double circle in (c) denotes that it is a deterministic mapping of its inputs. Actor Feature Extraction. Given LiDAR and maps, the backbone CNN detects the actors in the scene, and individual feature vectors per actor are extracted via RRoI Align, followed by a CNN with spatial pooling.
  • 35. Implicit Latent Variable Model for Scene- Consistent Motion Forecasting Implicit Latent Variable Model encodes the scene into a latent space, from which it can efficiently sample multiple future realizations in parallel, each with socially consistent trajectories.
  • 36. Implicit Latent Variable Model for Scene- Consistent Motion Forecasting
  • 37. Implicit Latent Variable Model for Scene- Consistent Motion Forecasting
  • 38. Perceive, Predict, and Plan: Safe Motion Planning Through Interpretable Semantic Representations • In this paper propose an end-to-end learnable network that performs joint perception, prediction and motion planning for self-driving vehicles and produces interpretable intermediate representations. • Unlike existing neural motion planners, this motion planning costs are consistent with perception and prediction estimates. • This is achieved by a novel differentiable semantic occupancy representation that is explicitly used as cost by the motion planning process. • This network is learned end-to-end from human demonstrations. • The experiments in a large-scale manual-driving dataset and closed-loop simulation show that the proposed model significantly outperforms state-of- the-art planners in imitating the human behaviors while producing much safer trajectories.
  • 39. Perceive, Predict, and Plan: Safe Motion Planning Through Interpretable Semantic Representations The overview of end-to-end learnable autonomy system that takes raw sensor data, an HD map and a high level route as input and produces safe maneuvers for the self- driving vehicle via our novel semantic interpretable intermediate representations.
  • 40. Perceive, Predict, and Plan: Safe Motion Planning Through Interpretable Semantic Representations Semantic classes in occupancy forecasting. Colors match between drawing and hierarchy. Shadowed area corresponds to the SDV route. Black vehicle, pedestrian and bike icons represent the agents' true current location.
  • 41. Perceive, Predict, and Plan: Safe Motion Planning Through Interpretable Semantic Representations Inference diagram of the perception and recurrent occupancy forecasting model. || symbolizes concatenation along the feature dimension, ⊕ element- wise sum and ∆ bilinear interpolation used to downscale the occupancy.
  • 42. Perceive, Predict, and Plan: Safe Motion Planning Through Interpretable Semantic Representations Examples of the motion planner cost functions: (a) collision, (b) driving-path, (c) lane boundary, (d) traffic light, (e) comfort, (f) route, (g) progress. Cost related comfort, traffic rules and progress in the route: Safe cost:
  • 43. Perceive, Predict, and Plan: Safe Motion Planning Through Interpretable Semantic Representations Learn the model parameters by exploiting these two loss functions: Semantic Occupancy Loss: Planning Loss:
  • 44. Perceive, Predict, and Plan: Safe Motion Planning Through Interpretable Semantic Representations