Professional Documents
Culture Documents
Sensor Set Design Patterns For Autonomous Vehicles - Open Autonomous Driving PDF
Sensor Set Design Patterns For Autonomous Vehicles - Open Autonomous Driving PDF
Why does every self-driving car company use different sensors and different mounting
positions? Although there is still a large variety in sensor sets, there are some recurring
patterns to be found, like high resolution lidars mounted on the roof center of autonomous
vehicles. This article aims to provide an overview of these patterns and to explain the
reasoning behind the different approaches.
🛈 Disclaimer / scope
This article focusses on sensor positioning and the impact of the field of view (FOV) of
sensors on the scenarios autonomous cars can manage. Details about sensors and
sensor technology are not a major subject of this article.
This article focusses a more on camera and lidar than on radar. Other sensors like
ultrasonic or thermal cameras are highly interesting for autonomous driving as well but
are not a major subject of this article.
For the sake of simplicity, sensor coverage is visualized as continuous volumes.
The FOV of the model sensors used for the considerations below reflects “standard
sensors” used in autonomous driving, see Lidar comparison, Radar comparison.
Abbreviations
FOV: Field Of View
ADAS: Advanced Driver Assistance System
AV: Autonomous Vehicle
https://autonomous-driving.org/2019/01/25/positioning-sensors-for-autonomous-vehicles/ 1/18
6/12/2020 Sensor Set Design Patterns for Autonomous Vehicles – open autonomous driving
General Considerations
Given their individual features, the sensors of an autonomous car determine its operational
design domain (ODD) – the environments in which the car can be deployed safely. Identifying
the sensors required to “solve” an ODD is complex since there is a plethora of common cases
and edge cases to consider, while all sensors have individual traits and limitations – each
sensor is chosen as a trade-off between accuracy, range, sampling rate, FOV, cost and overall
system complexity.
Scenarios
A common way to identify the sensor set needed to solve an ODD is to capture a variety of
scenarios, that are fed into simulation to check under which circumstances a sensor set would
miss objects in the environment. The performance of different sensor sets is compared, and
promising configurations are improved iteratively. This data-driven approach allows to use
Machine Learning in order to identify the optimal configuration. Scenarios can be collected
automatically from cars in the field or can be designed manually. Below is a selection of
challenging scenarios and edge cases:
If a lidar sensor with small vertical FOV is mounted too low, none of the lidar beams it emits
might hit a protruding obstacle.
https://autonomous-driving.org/2019/01/25/positioning-sensors-for-autonomous-vehicles/ 2/18
6/12/2020 Sensor Set Design Patterns for Autonomous Vehicles – open autonomous driving
Protruding obstacle are not detected if the vertical resolution of the lidar sensor is too low, and
no beam hits the obstacle.
Cyclist passing by
When taking a turn to the right, special caution must be paid for traffic participants that could
be cut off by the turn. While big vehicles are comparably easy to detect (via Radar, e.g. for
lane change assist), light traffic participants such as cyclists can be significantly harder to spot,
sensors mounted on the roof center.
A cyclist riding next to the vehicle is not in the FOV of lidar sensor mounted on the roof center.
https://autonomous-driving.org/2019/01/25/positioning-sensors-for-autonomous-vehicles/ 3/18
6/12/2020 Sensor Set Design Patterns for Autonomous Vehicles – open autonomous driving
A centrally mounted sensor would not be able to perceive traffic coming from an occluded
crossing lane.
A sensor facing sideways close to the vehicle front can peek into the crossing lane and detect
the approaching cross-traffic.
https://autonomous-driving.org/2019/01/25/positioning-sensors-for-autonomous-vehicles/ 4/18
6/12/2020 Sensor Set Design Patterns for Autonomous Vehicles – open autonomous driving
Slope
Slope has a major impact on the requirements towards the vertical FOV of sensors. When
approaching ramps and underpasses, the coverage on other traffic participants is reduced. In
addition, the geometry of the road surface must be compensated for object detection;
otherwise ramps will be classified as obstacles, and actual obstacles out of the FOV will be
missed. Particular challenging are intersections with ascending or descending crossing lanes.
https://autonomous-driving.org/2019/01/25/positioning-sensors-for-autonomous-vehicles/ 5/18
6/12/2020 Sensor Set Design Patterns for Autonomous Vehicles – open autonomous driving
A ramp in front of the vehicle can be erroneously interpreted as an obstacle, while the second
vehicle is not covered by the sensor at all.
Sensor Colocation
When sensors are mounted apart from each other, they’ll have a different perspective on the
vehicle’s environment (parallax) which may lead to a situation in which one sensor perceives
an object, while it is occluded for the other sensor. This is a challenge for sensor data fusion:
for high level fusion (fusion of objects), there’ll be contradicting detections from the sensors.
For low level fusion (fusion of raw sensor data), the parallax will result in areas that can’t be
matched between the sensors and consequently ‘holes’ in the fused sensor data. These
effects are particularly strong for the close environment.
To reduce these issues to a minimum, different sensor modalities, particularly camera and
lidar, are often colocated: the sensors are mounted as close as possible to each other.
https://autonomous-driving.org/2019/01/25/positioning-sensors-for-autonomous-vehicles/ 6/18
6/12/2020 Sensor Set Design Patterns for Autonomous Vehicles – open autonomous driving
For a highway lane keeping feature with the driver supervising, a front facing radar and camera
can be sufficient.
For autonomous driving in urban environments, 360° horizontal coverage across all three
major sensor modalities (camera, lidar and radar) is required in order to reliably detect traffic
participants in the relevant scenarios while leveraging the different modalities to compensate
their respective weaknesses. In fact, 360° are necessary, but won’t suffice; a 360° spinning
lidar on top of the vehicle is insufficient to detect obstacles in many cases (see lidar section
below) and needs to be complemented by further lidars.
Commercial considerations
Having solved an ODD enables offering a product. Whether there’ll be a positive business
case for this product or not will depend on many further factors, such as the nature of the
service provided by the AV, region where the service is being provided etc.
This article won’t explore commercial considerations, but it is vital for sensor set design to
acknowledge that, apart from sensors’ contribution to solving the ODD, cost, reliability and
production scalability are decisive factors for the deployment of large fleets.
Sensor Modalities
Lidar
Lidars are a comparably new technology in the automotive industry; they used to be more
common in robotics. With the advance of ADAS and autonomous driving, this is about to
change. While first series cars use low resolution Lidars already, high resolution Lidars
allowing detection at range and even classification are currently reserved for self-driving cars
with generous sensor budgets.
There are various different lidar sensors on the market, different in wavelength, resolution, and
technology – to get familiar with the technology behind the sensors, read more here.
https://autonomous-driving.org/2019/01/25/positioning-sensors-for-autonomous-vehicles/ 7/18
6/12/2020 Sensor Set Design Patterns for Autonomous Vehicles – open autonomous driving
Variations
Pros
Cons
Blind spots for low objects in all directions, particularly to the rear.
Sensor must be elevated from roof to leverage full vertical FOV, the high position causes
mechanical challenges in case of abrupt deceleration and makes it impossible to enter
areas with low ceiling.
https://autonomous-driving.org/2019/01/25/positioning-sensors-for-autonomous-vehicles/ 8/18
6/12/2020 Sensor Set Design Patterns for Autonomous Vehicles – open autonomous driving
Variations
Pros
Cons
Higher complexity in terms of integration and point cloud fusion compared to single
sensor
Tilting of the sensors is limited by their vertical FOV. Too few tilting results in blind spots
near the vehicle while too much tilting limits coverage on elevated objects.
Front lidar
https://autonomous-driving.org/2019/01/25/positioning-sensors-for-autonomous-vehicles/ 9/18
6/12/2020 Sensor Set Design Patterns for Autonomous Vehicles – open autonomous driving
Variations
Pros
Cons
On slopes; sensor will return ground plane as an obstacle for ascending roads and no
returns for descending roads,
Insufficient overall coverage for urban AD if not combined with other sensors.
https://autonomous-driving.org/2019/01/25/positioning-sensors-for-autonomous-vehicles/ 10/18
6/12/2020 Sensor Set Design Patterns for Autonomous Vehicles – open autonomous driving
Variations:
Higher horizontal FOV and vertical FOV lidars enable full coverage of the vehicle’s sides.
Different positions: position behind front wheel case may simplify integration, position
below side mirror allows to look over low obstacles.
Pros:
Enables Crossing with occlusion scenario: The vehicle can detect cross-traffic at
intersections without entering and potentially blocking the traffic.
Cons:
Sensor with low vertical FOV will have issues with slopes
Camera
Cameras (as opposed to Lidars, radars and ultrasonic sensors) are passive sensors, that
collect light reflected from the environment. While cameras offer a comparably high resolution,
their data is subject to a variety of environment influences (darkness, rain, fog, low sun, spray,
snow, dust, …). Camera data needs to be interpreted by computationally expensive computer
vision algorithms or Deep Learning in order to provide useful information for ADAS and
autonomous driving, while the data from distance measuring sensors such as Lidar can used
(almost) as-is for collision avoidance. On the positive side camera data (once interpreted)
provides the semantically richest representation of the vehicle’s environment.
Wide angle (“Fisheye”) cameras have a large FOV, but their images are subject to lens
distortion that needs to be corrected before further processing. Lens distortion correction
removes parts of the image and consequently reduces the FOV. In addition, the areas close to
the edges of the image may be blurred and less reliable for image recognition tasks.
Many series cars are equipped with a set of wide-angle cameras to provide a bird’s-eye view
of the close environment that makes parking easier. Usually, 360° horizontal coverage with
cameras is reached by mounting several 180° cameras on the vehicle. For ADAS, cameras
are mounted around the body of the vehicle (front bumper, mirrors, trunk lid) while for self-
driving they are mounted on the vehicle’s roof, e.g. at the roof edge centers.
Variations:
Cameras can by tilted to better cover the close environment around the vehicle.
Cameras can be mounted on the corners of the vehicle instead of the roof edge centers,
or as a central “camera tower”, see below.
Pros
360° coverage.
https://autonomous-driving.org/2019/01/25/positioning-sensors-for-autonomous-vehicles/ 12/18
6/12/2020 Sensor Set Design Patterns for Autonomous Vehicles – open autonomous driving
Cons
There are several common ways of mounting cameras apart from the “cameras at the roof
edge centers” approach described above:
Cameras at the roof corners show larger areas of low sensor reliability compared to cameras
on the roof edge centers (highly depending on vehicle dimensions). Additionally, the highly
important view to the front is covered by the outer parts of the camera’s field of view; for
minimum distortion, the view to the front would be on the optical axis of a camera.
Instead of 180° cameras, 120° cameras that are less affected by lens distortion can be
mounted as a “camera tower” on the roof center. This can be combined with a central 360°
lidar to form on compact sensor module. As described for lidars above, the sensors mounted in
this configuration can’t observe objects occluded by the roof.
https://autonomous-driving.org/2019/01/25/positioning-sensors-for-autonomous-vehicles/ 13/18
6/12/2020 Sensor Set Design Patterns for Autonomous Vehicles – open autonomous driving
Front camera
Front cameras are the primary (and often only) sensor for environment recognition in current
series cars. They are comparably cheap but enable a variety of ADAS features – although they
are limited by their sensibility towards environment conditions, like all cameras.
Variations
Front cameras are typically mounted between the rear mirror and the windshield. The
windshield protects the camera and is cleaned by the wipers.
Alternatively, front cameras can be installed inside the vehicle between the dashboard
and the windshield, or outside of the vehicle on the bumper or at the front roof edge
center.
Stereo cameras can provide distance estimation.
Pros
Cons
Vertical FOV is limited by vehicle’s hood hence small objects in front of the vehicle are
occluded (unless camera is mounted on front bumper).
https://autonomous-driving.org/2019/01/25/positioning-sensors-for-autonomous-vehicles/ 14/18
6/12/2020 Sensor Set Design Patterns for Autonomous Vehicles – open autonomous driving
Insufficient overall coverage for urban AD if not combined with other sensors.
Front camera and surround view camera systems are usually unable to detect objects the are
close and elevated, such as traffic lights, because of their limited vertical FOV. This can be
solved by adding a wide-angle camera mounted at the front roof edge center. Depending on
the positioning of traffic lights in the target environment, it may be necessary to tilt these
cameras upwards.
https://autonomous-driving.org/2019/01/25/positioning-sensors-for-autonomous-vehicles/ 15/18
6/12/2020 Sensor Set Design Patterns for Autonomous Vehicles – open autonomous driving
Radar
Radars are already established in the automotive industry, they have been employed in series
cars since many years to enable ADAS features such as Adaptive Cruise Control (ACC) and
Autonomous Emergency Breaking (AEB).
Radars accurately measure distance and radial velocity. They are particularly good at
detecting metallic objects but are also able to detect non-metallic objects such as pedestrians
with reduced range.
Pros
A set of 4-6 short-range radars (SRR, ~30m range) and mid-range-radars (MRR, ~100m
range) suffices to reach 360° horizontal coverage and solve many typical scenarios in
urban ODDs.
Provides redundancy for lidar and camera while being more robust against challenging
environment conditions.
Cons
The narrow vertical FOV of conventional automotive radars poses severe challenges
when facing non-flat terrain.
https://autonomous-driving.org/2019/01/25/positioning-sensors-for-autonomous-vehicles/ 16/18
6/12/2020 Sensor Set Design Patterns for Autonomous Vehicles – open autonomous driving
Front radar
Pros
Mid-range radars (MRR, ~100m range) and long-range radars (LRR, ~250m range)
enable ADAS features like Adaptive Cruise Control (ACC) and Autonomous Emergency
Breaking (AEB).
More robust against challenging environment conditions like rain, fog and snow than lidar
and camera.
Cons:
Insufficient overall coverage for urban AD if not combined with other sensors.
Conclusion
Sensing for automated cars is like sensing for living beings – there is no one perfect sensor
configuration that prevails in all environments and fits all budgets. The range of available
sensor products is increasing fast, and there are multiple configurations that do the job, with
their respective pros and cons.
Providing background on sensor set design and exploring frequent patterns for the different
sensor modalities in this article did not aim for providing an absolute answer; instead it was
https://autonomous-driving.org/2019/01/25/positioning-sensors-for-autonomous-vehicles/ 17/18
6/12/2020 Sensor Set Design Patterns for Autonomous Vehicles – open autonomous driving
supposed to convey the thought process of sensor set design, enabling an understanding of
the rationale behind sensor choices.
If the concepts and patterns above intrigue you, a next step for getting more familiar with
sensors for autonomous driving is to take a deeper look into sensor technology and the sensor
products that are currently on the market. The next section lists helpful materials for this
purpose.
Additional resources
Public sensor configurations
Some companies working on self-driving cars have already released their sensor
configurations:
Sensor technology
Three Sensor Types Drive Autonomous Vehicles
Article by sensorsmag.com
Tesla & Google Disagree About LIDAR — Which Is Right?
Great article that compares camera, lidar, radar and ultrasonic sensors.
Three sensor types drive autonomous vehicles
Details on automotive cameras, lidars and radars as an extrapolation of current series
sensors, with a reasonable explanation of technology and physical principles.
Sensor specifications
Spec comparison sheets by autonomoustuff.com: Radar, Lidar
Product overview by autonomoustuff.com
incl. Velodyne HDL-64 lidar (spec sheet), Delphi ESR 2.5 radar (spec sheet)
https://autonomous-driving.org/2019/01/25/positioning-sensors-for-autonomous-vehicles/ 18/18