Professional Documents
Culture Documents
Driving School Syestem
Driving School Syestem
net/publication/224242658
The Driving School System: Learning Basic Driving Skills From a Teacher in a
Real Car
CITATIONS READS
8 16,741
11 authors, including:
Some of the authors of this publication are also working on these related projects:
1993 - 1995 : Face Processing with Elastic Bunch Graph Matching View project
All content following this page was uploaded by Aušra Vidugirienė on 12 May 2014.
Abstract—To offer increased security and comfort, advanced situation, they use control laws, from which they derive out-
driver-assistance systems (ADASs) should consider individual put signals to aid the driver [1]–[5]. However, interindividual
driving styles. Here, we present a system that learns a human’s differences in driving can be large [6], [7], and it is difficult
basic driving behavior and demonstrate its use as ADAS by issuing
alerts when detecting inconsistent driving behavior. In contrast for conventional control systems to accommodate these differ-
to much other work in this area, which is based on or obtained ences, leading to suboptimal driver support. This can finally
from simulation, our system is implemented as a multithreaded decrease the driver’s safety [8], [9].
parallel central processing unit (CPU)/graphics processing unit Such observations were our motivation to investigate and
(GPU) architecture in a real car and trained with real driving data build a system that automatically adapts to the driving style of
to generate steering and acceleration control for road following.
It also implements a method for detecting independently moving its users. In addition, current R&D efforts of the car industry
objects (IMOs) for spotting obstacles. Both learning and IMO focus on systems that explicitly take the driver and its behavior
detection algorithms are data driven and thus improve above the into account [10], [11]. In addition to the safety aspect, such
limitations of model-based approaches. The system’s ability to systems will also be more easily accepted by users because
imitate the teacher’s behavior is analyzed on known and unknown they will provide more comfort, which is an important sales
streets, and results suggest its use for steering assistance but limit
the use of the acceleration signal to curve negotiation. We propose argument.
that this ability to adapt to the driver can lead to better acceptance In the current study, we will describe a system based on
of ADAS, which is an important sales argument. imitation learning, i.e., a system that learns to interpret basic
Index Terms—Advanced individualized driver-assistance aspects of the road (lanes) in the same way as its driver,
system, driving, imitation learning, independently moving object reproducing the driver’s actions. In addition, we demonstrate
(IMO), real-time system. its use as a basic driver assistance system by issuing warning
signals if the driver deviates from his/her predicted default
behavior. The so-called DRIVSCO1 system is realized by a
A DVANCED driver-assistance systems (ADASs) that adapt
to the individual driver have high potential in the car
industry since they can reduce the risk of accidents while
multithreaded parallel central processing unit (CPU)/graphics
processing unit (GPU) architecture. It is vision based, operates
in real time on real roads, and also includes a method for data-
providing a high degree of comfort. Conventional systems are
driven detection of independently moving objects (IMOs). The
based on a general moment-to-moment assessment of road
latter is not the focus of this paper and is therefore only briefly
and driving parameters. To arrive at a judgment of the current
described. The system has been designed for use on motorways
and country roads.
Manuscript received March 12, 2010; revised November 5, 2010 and Before describing the details of our system and comparing
March 15, 2011; accepted April 6, 2011. Date of publication June 20, 2011; date it to the literature (see State of the Art), we shortly explain its
of current version December 5, 2011. This work was supported in part by the
European Commission under Project FP6-IST-FET (DRIVSCO) and in part by structure as a guideline for the reader (see Fig. 1). Thus, this
the Bernstein Focus Neurotechnology (BFNT) Göttingen. The Associate Editor paper is organized as follows: in Section I, the overall structure
for this paper was S. Tang. of the system is presented. It is compared with the state of the
I. Markelić and F. Wörgötter are with Georg-August-University Göttingen,
37077 Göttingen, Germany (e-mail: irene@physik3.gwdg.de; worgott@ art in Section II, with its realization explained in Section III and
physik3.gwdg.de). results presented in Section IV. In Section V, we conclude and
A. Kjær-Nielsen, L. Baunegaard With Jensen, and N. Krüger are
with Maersk McKinney Moller Institute, University of Southern Denmark,
discuss the presented work.
5230 Odense, Denmark (e-mail: akn@mmmi.sdu.dk; lbwj@mmmi.sdu.dk;
norbert@mmmi.sdu.dk).
K. Pauwels, N. Chumerin, and M. Van Hulle are with Katholieke Universiteit I. S YSTEM OVERVIEW
Leuven, 3000 Leuven, Belgium (e-mail: Karl.Pauwels@med.kuleuven.be;
Nikolay.Chumerin@med.kuleuven.be; Marc.VanHulle@med.kuleuven.be). The overall structure of the DRIVSCO system is shown
A. Rotter is with Hella KGaA Hueck & Co, 59552 Lippstadt, Germany in Fig. 1. The yellow box (“human senseact”) symbolizes
(e-mail: Alexander.Rotter@hella.com).
A. Vidugiriene and M. Tamosiunaite are with Vytautas Magnus the human driver who senses the environment, which is de-
University, 44248 Kaunas, Lithuania (e-mail: m.tamosiunaite@if.vdu.lt; noted by the “world” box, and responds to it with adequate
a.vidugiriene@if.vdu.lt). driving actions (“act”). At the same time, the system senses the
Color versions of one or more of the figures in this paper are available online
at http://ieeexplore.ieee.org.
Digital Object Identifier 10.1109/TITS.2011.2157690 1 This is short for Driving School.
Fig. 2. System integration in the car. (a) Mounting for the stereo camera
system. (b) Fixation of the computer in the trunk. (c) Installation of the monitor
in the car.
C. Dense Vision
The visual cues used for the IMO detection are computed
during the “dense vision” stage (see Fig. 3). Dense vision
algorithms process the visual signal in its entirety, as opposed
to sparse vision algorithms that only operate on interest points
such as edges or corners. The cues used here are dense disparity
(the horizontal difference in image location for corresponding
pixels in the rectified left and right image) and optical flow (the
2-D image motion of each pixel). The algorithms used rely on
the phase of quadrature pair Gabor filter responses [34] [see
Fig. 5(b)], which were extracted at multiple orientations and
scales, to establish correspondences. The GPU implementation
[35] of a phase-based algorithm [36] is used to compute optical
flow [see Fig. 5(c)]. This algorithm integrates the temporal
phase gradient across orientation and gradually refines its esti-
Fig. 5. Real-time IMO detection. Multiorientation multiscale Gabor filter
mates by traversing a Gabor pyramid from coarser to finer lev- responses (B) are extracted from the stereo image pair (A) and used to
els. The optical flow is computed for the left video stream only. compute dense optical flow (C) and stereo disparity (D). The horizontal and
The dense disparity algorithm [see Fig. 5(d)] is very similar vertical optical flow is color coded from −15 (dark red) to +15 pixels (dark
blue), and the stereo disparity is color coded from −50 (dark red) to +20
to the optical flow algorithm but operates on phase differences (dark blue) pixels. Combined with egomotion (E, which is extracted from the
between the left and right image, as opposed to temporal phase optical flow, not shown), these cues allow the extraction of independent motion
gradients. These aspects are described in more detail in [37]. (F, likelihood increases from blue to red). This independent motion signal is
gathered in a fixed region of the image (G), and when it exceeds a threshold,
a warning is issued.
D. IMO Detection and Warning
deviations are assigned to independent motion if they comply
Recent (offline) approaches for the detection of IMOs have with a 3-D translation model in a local region surrounding the
achieved very good results using model-based techniques [38], pixel [see Fig. 5(f)]. To detect a moving vehicle in front, the
[39]. However, limitations include the difficult detection of (pixelwise) independent motion signal is accumulated inside a
distant IMOs since they occupy only small patches in the image fixed region in the image [see the blue rectangle in Fig. 5(g)].
and are thus difficult to match and, further, objects for which no A warning is issued when more than 30% of the pixels within
model is provided cannot be detected. Here, we use a model- this box are considered independently moving.
free mechanism that is more general in the sense that it will
respond to any sufficiently large (11 × 11 pixels) moving
E. Lane Detection and Filtering
object. The IMO detection component combines dense vision
cues (optical flow and stereo; see Section III-C) in real time to A large number of lane detection algorithms have been
compute egomotion (the rotation and translation of the camera) reported, which can roughly be divided into feature- and model-
and independent motion (the parts of the image that move with based methods (see, e.g., [2]). The former detect street lanes
respect to the static environment) to detect an IMO in front. bottom-up, i.e., based on certain low-level features such as in-
The whole process is complex and cannot be described in tensity gradients, edges, and color, e.g., [44], whereas the latter
detail in this paper. See [40] for further information. Here, aim at identifying image parts corresponding to a predefined
we will give only a short overview summarized in Fig. 5. lane or street model, e.g., [45] and [46]. Both approaches have
A nonlinear instantaneoustime model [41] is used to extract known advantages and disadvantages: feature-based methods
egomotion from the optical flow. To obtain optimal estimates, can detect arbitrary shapes but might add erroneously detected
an iterative minimization procedure that relies on M-estimation image parts into the returned street detection. The more re-
is used [42] for outlier compensation. A total of 32 different stricted model-based methods are more robust to disturbances
initializations are explored to deal with local minima. The such as occlusions, noise, and shadows. However, they are
data-intensive parts of the algorithm entirely run on the GPU. restricted to predefined lane shapes, making them less flexible.
Independent motion is detected by evaluating the depth/flow To detect lanes in incoming image frames, we employ a simple
constraint [43] at each pixel. Deviations from this constraint and fast (real time, i.e., 20 Hz) feature-based algorithm, which
point to inconsistencies between the optical flow, disparity, and works similar to contour tracers used in computer vision: First,
egomotion and result from noise or independent motion. The edges and pixel orientations are computed by using standard
MARKELI Ć et al.: DRIVING SCHOOL SYSTEM: LEARNING BASIC DRIVING SKILL FROM TEACHER IN REAL CAR 1139
Fig. 7. (a) Extracted street boundaries described by polylines and the corner e = {s, sfut , afut } PAR entry. (2)
points. The vectors of the corner points (c0l/r , c1l/r , . . .) for the left and right
lanes constitute the visual state description. (b) Steering and (c) acceleration The action sequences spast/fut and afut
signals from 15 runs from the same track and driver. The mean is plotted in
black. “SNR” refers to signal-to-noise ratio. (d) The sequence closest to the
mean is shown in red. Black signals are considered to be sufficiently similar. spast = [st1 , st2 , . . . , stm ] (3)
be sufficiently similar, and the others (differently colored plots) sfut = [st , st+1 , . . . , st+n ] (4)
are those that were sorted out. The sequence closest to the mean
is plotted in red. We use these data for training the PAR, as afut = [at , at+1 , . . . , at+n ] (5)
explained in Section III-F2.
Driving data assigned to similar situations are averaged dur- with s and a denoting single steering and acceleration signal
ing training. Similarity is defined by the resemblance between values (actions).
the left and right street lanes of the observed image, and one The descriptions of the left and right street lanes (vleft/right )
already contained in the PAR, as well as a short sequence are linear approximations (polylines) of the extracted lanes
of previous steering data. A formal definition is given in in image coordinates, which were obtained by applying the
Section III-F3. Hence, for tracks that are being driven multiple Douglas–Peucker method [53]. We store this as vectors contain-
times, the system defines the default human driving behavior ing the corner points of the polylines, as visualized in Fig. 7(a).
over the mean of the action sequences obtained from driving on Thus
the same track, as shown by the black plot in Fig. 7(b). Note,
however, that learning starts with the first entry in the PAR, vright = [c0r , c1r , . . . , clr ] (6)
and additional information just improves the performance of
vleft = [c0l , c1l , . . . , cll ] (7)
the perception–action mapping. Single-example performance
is reliable for steering but remains restricted for acceleration
with ll and lr denoting the lengths of vleft and vright .
control (see Section V). Note that, as the curve shapes are fairly
During the training phase, a new entry e [see (2)] is added
similar, after some learning, the system is able to generalize into
to the PAR if it represents new knowledge, i.e., if no entries
unseen curves. This will be shown later (see Fig. 11).
are already available containing similar information. An entry
Since the signals predicted by the PAR should correspond
is added if 1) there is no other entry already in the PAR, with
to the default behavior of the human, i.e., the mean action
vleft and vright having the same lengths as those contained in
signal, we evaluate the performance of the PAR by comparing
the probed entry (in other words, if a completely new street
its output against the human action sequence closest to the mean
trajectory is observed), or 2) if there are such entries but none
signal, i.e., the red plot in Fig. 7(d) (which we do not include in
of them has a state vector similar to that probed for adding.
the training data).
Two state vectors are similar if the differences between their
PAR Training: The information stored as experience in the
components are each below a fixed threshold, that is, if the
PAR is a state vector s containing a description of extracted
following are fulfilled:
left and right street lanes (vleft , vright ) from a current image
frame It . Here, t denotes the time of the image observation.3
_v ≤ thresh_v (8)
Because we found that only a description of the street ahead
is not sufficient to distinguish different driving situations from _spast ≤ thresh_spast (9)
each other, e.g., a straight street can lead to acceleration or no
where _spast is the summed squared difference between the
3 Time is measured in discrete time steps according to the camera’s image
entries of spast of two state vectors, and _v = _vl + _vr and
capturing frequency. _vl/r are the sums of the normalized, weighted, and summed
MARKELI Ć et al.: DRIVING SCHOOL SYSTEM: LEARNING BASIC DRIVING SKILL FROM TEACHER IN REAL CAR 1141
TABLE I
C OMPARISON B ETWEEN THE P RESENCE OF AN ACTUAL L ANE
M ARKER IN THE S CENE AND THE O UTPUT OF THE D EVELOPED
L ANE -D ETECTION A LGORITHM
IV. R ESULTS
The performance of the IMO detection is assessed in two
ways, which together demonstrate the range of situations that
the algorithm can handle. On one hand, simulated scenes are
used with precisely known parameters to allow a rigid quantita-
tive evaluation. On the other hand, several real-world scenes are
processed to assess the qualitative performance. We evaluate
the detection of two moving objects and the magnitude of their
speed in a demanding simulated scene and find that both objects
are detected with a density of more than 50%, compared with
all pixels that belong to the objects. The magnitude of the speed
Fig. 10. Results for (left column) steering and (right column) acceleration
is calculated with an accuracy of better than 98%. The qual- prediction for known tracks. “cc” denotes the correlation coefficient value
itative performance was evaluated by applying the algorithm between the human-generated signal and the synthesized signal.
to a diverse set of real-world driving situations involving both
country roads and complex urban environments. A variety of
moving objects (cars, bikes, and pedestrians) were successfully
detected, whereas the car itself underwent a wide range of
egomotion speeds and steering actions. Further information can
be found in [40].
The developed lane detection algorithm was evaluated on a
randomly chosen sample of each tour of the available data set,
which comprised a variety of country roads with both clear and
Fig. 11. Results for steering (left) and acceleration (right) prediction on an
damaged lane markers (a total of 11.400 frames with humanly unknown track.
detectable lane markers present in 11.055 frames). In 98.9%
of the 11.055 frames, a valid lane description was extracted. steering signal is better approximated than acceleration, which
In 5.6% of these cases, only the left marker and, in 8.5%, is as expected from the computed SNR in Section III-F1.
only the right marker were detected. Both markers were found For testing the performance on an unknown track, we train
in 84.8% of these cases. This information is summarized in the PAR with all available data, except that from the track we
Table I. Examples of detected lanes are shown in Figs. 4, 7(a), test for. The result is shown in Fig. 11. We chose s05, which
and 12. contains a very sharp turn (see Fig. 12), leading to two typical
To evaluate how well the action predictions match the human phenomena, which are discussed here. The resulting steer and
default behavior, we use the provided data set. acceleration predictions in comparison to the human signal are
After filtering it, as explained in Section III-F1, approxi- shown in Fig. 11. The steering prediction is very close to the
mately 80 min of training data were obtained, resulting in a human signal, but the acceleration is less reliable. In particular,
PAR with 90 470 entries, which is adequate for evaluating the it can be seen from the plotted acceleration prediction in Fig. 11
system performance. First, we test the performance on a known that the sharp curve [which is entered around time step 200,
track, i.e., one that the system had seen before. For that, we train c.f., Fig. 12(a)] is accompanied by a deceleration in the human
the PAR with all runs but that closest to the mean, which we signal (between time steps 220 and 380) and that this is nicely
consider to resemble the human default behavior, as explained reflected by the algorithm. However, before and after the decel-
in Section III-F1, and we use for testing. The smoothed steering eration part, the predicted signal considerably differs from the
and acceleration signals from the algorithm are plotted against human data.
the signal generated by the human for s03, s05, and s06 in This phenomenon results from the fact that street parts with a
Fig. 10. As an additional measure of similarity, we compute lower curvature allow a great variance in the driver’s choice of
the correlation coefficient of the two signals (human and pre- speed, depending on hard-to-access variables including “mood”
diction). For steering and acceleration prediction, we obtain and “intention,” where curves, particularly sharp curves, sig-
0.99 and 0.81 for s03, 0.93 and 0.73 for s05, and 0.97 and nificantly restrict the driver’s behavior and thus make it more
0.67 for s06. Thus, all predictions are very good. However, the predictable. We therefore conclude that acceleration prediction
MARKELI Ć et al.: DRIVING SCHOOL SYSTEM: LEARNING BASIC DRIVING SKILL FROM TEACHER IN REAL CAR 1143
bad illumination conditions, the human’s sensing process is ob- [17] D. A. Pomerleau, “Alvinn: An autonomous land vehicle in a neural
viously limited; however, by using infrared light, the system’s network,” in Advances in Neural Information Processing Systems 1.
San Mateo, CA: Morgan Kaufmann, 1989.
sensing is less affected, given that the lane detection process is [18] D. A. Pomerleau, “Neural network based autonomous navigation,” in
adequately adapted to the new circumstances. Thus, the system Proc. NAVLAB, 1990, pp. 558–614.
can use its knowledge about driving acquired during the day [19] D. A. Pomerleau, “Efficient training of artificial neural networks for
autonomous navigation,” Neural Comput., vol. 3, no. 1, pp. 88–97, 1991.
to support the human in the more difficult situation of night [20] M. Arbib, “Neural network vision for robot driving,” in The Handbook
driving. of Brain Theory and Neural Networks. Cambridge, MA: MIT Press,
1999.
[21] M. Pasquier and R. J. Oentaryo, “Learning to drive the human way: A
step towards intelligent vehicle,” Int. J. Veh. Auton. Syst., vol. 6, no. 1/2,
R EFERENCES pp. 24–47, Dec. 2007. [Online]. Available: http://dx.doi.org/10.1504/
IJVAS.2008.016477
[1] S. Mammar, S. Glaser, and M. Netto, “Time to line crossing for [22] C. Sammut, S. Hurst, D. Kedzier, and D. Michie, “Learning to fly,” in
lane departure avoidance: A theoretical study and an experimental set- Proc. Mach. Learn., 1992, pp. 385–393.
ting,” IEEE Trans. Intell. Transp. Syst., vol. 7, no. 2, pp. 226–241, [23] A. Y. Ng and S. Russel, “Algorithms for inverse reinforcement learning,”
Jun. 2006. in Proc. ICML, 2000, pp. 663–670.
[2] J. McCall and M. Trivedi, “Video-based lane estimation and tracking for [24] P. Abbeel and A. Y. Ng, “Apprenticeship learning via inverse reinforce-
driver assistance: Survey, system, and evaluation,” IEEE Trans. Intell. ment learning,” in Proc. 21st Int. Conf. Mach. Learn., 2004, p. 1.
Transp. Syst., vol. 7, no. 1, pp. 20–37, Mar. 2006. [25] Official Diplecs Website, 2010. [Online]. Available: http://www.
[3] A. Amditis, M. Bimpas, G. Thomaidis, M. Tsogas, M. Netto, S. Mammar, diplecs.eu/
A. Beutner, N. Möhler, T. Wirthgen, S. Zipser, A. Etemad, M. Da Lio, and [26] C. H. Hwang, N. Massey, B. W. Miller, and K. Torkkola, “Hybrid intelli-
R. Cicilloni, “A situation-adaptive lane-keeping support system: Overview gence for driver assistance,” in Proc. FLAIRS, 2003, pp. 281–285.
of the safelane approach,” IEEE Trans. Intell. Transp. Syst., vol. 11, no. 3, [27] R. Freymann, “Driver assistance technology to enhance traffic safety,” in
pp. 617–629, Sep. 2010. Motion and Vibration Control. Berlin, Germany: Springer-Verlag, 2008,
[4] L. Li, F. Y. Wang, and Q. Zhou, “Integrated longitudinal and lateral pp. 71–81.
tire/road friction modeling and monitoring for vehicle motion control,” [28] K. Brookhuis and D. de Waard, “Limiting speed, towards an intelligent
IEEE Trans. Intell. Transp. Syst., vol. 7, no. 1, pp. 1–19, Mar. 2006. speed adapter (ISA),” Transp. Res. F, Traffic Psychol. Behav., vol. 2, no. 2,
[5] E. Bertolazzi, F. Biral, M. D. Lio, A. Saroldi, and F. Tango, “Supporting pp. 81–90, Jun. 1999.
drivers in keeping safe speed and safe distance: The saspence subpro- [29] A. Tahirovic, S. Konjicija, Z. Avdagic, G. Meier, and C. Wurmthaler,
ject within the European framework programme 6 integrating project “Longitudinal vehicle guidance using neural networks,” in Proc. CIRA,
prevent,” IEEE Trans. Intell. Transp. Syst., vol. 11, no. 3, pp. 525–538, 2005, pp. 685–688.
Sep. 2010. [30] D. Partouche, M. Pasquier, and A. Spalanzani, “Intelligent speed adapta-
[6] P. Ulleberg and T. Rundmo, “Personality, attitudes and risk perception as tion using a self-organizing neuro-fuzzy controller,” in Proc. IEEE Intell.
predictors of risky driving behaviour among young drivers,” Safety Sci., Veh. Symp., 2007, pp. 846–851.
vol. 41, no. 5, pp. 427–443, Jun. 2003. [31] H. Kwasnicka and M. Dudala, “Neuro-fuzzy driver learning from
[7] F. I. Kandil, A. Rotter, and M. Lappe, “Driving is smoother and more real driving observations,” in Proc. Artif. Intell. Control Manag., 2002,
stable when using the tangent point,” J. Vis., vol. 9, no. 1, pp. 1–11, pp. 81–89.
Jan. 2009. [32] I. Markelic, T. Kulvicius, M. Tamosiunaite, and F. Wörgötter, “Antici-
[8] K. A. Brookhuis, D. de Waard, and W. H. Janssen, “Behavioural im- patory driving for a robot-car based on supervised learning,” in Proc.
pacts of advanced driver assistance systems an overview,” Eur. J. Transp. ABiALS, 2008, pp. 267–282.
Infrastructure Res., vol. 1, no. 3, pp. 245–253, 2001. [33] D. W. Aha, Ed., “Editorial,” in Lazy Learning. Norwell, MA: Kluwer,
[9] A. Lindgren and F. Chen, “State of the art analysis: An overview of 1997, ser. Artificial Intelligence Review, pp. 7–10.
advanced driver assistance systems (ADAS) and possible human factors [34] J. Daugman, “Uncertainty relation for resolution in space, spatial fre-
issues,” in Proc. Human Factors Eco. Aspects Saf. Swedish Netw. Human quency, and orientation optimized by two-dimensional visual cortical
Factors Conf., 2007, pp. 38–50. filters,” J. Opt. Soc. Amer. A, Opt. Image Sci. Vis., vol. 2, no. 7, pp. 1160–
[10] J. F. Coughlin, B. Reimer, and B. Mehler, “Driver wellness, safety & the 1169, Jul. 1985.
development of an awarecar,” AgeLab, Mass Inst. Technol., Cambridge, [35] K. Pauwels and M. Van Hulle, “Realtime phase-based optical flow on
MA, 2009. the GPU,” in Proc. IEEE Conf. Comput. Vis. Pattern Recog., Workshop
[11] S. Hoch, M. Schweigert, F. Althoff, and G. Rigoll, “The BMW surf Comput. Vis. GPU, 2008, pp. 1–8.
project: A contribution to the research on cognitive vehicles,” in Proc. [36] T. Gautama and M. Van Hulle, “A phase-based approach to the estimation
Intell. Veh. Symp., 2007, pp. 692–697. of the optical flow field using spatial filtering,” IEEE Trans. Neural Netw.,
[12] E. D. Dickmanns and V. Graefe, “Dynamic monocular machine vision,” vol. 13, no. 5, pp. 1127–1136, Sep. 2002.
Mach. Vis. Appl., vol. 1, no. 4, pp. 223–240, 1988. [37] S. Sabatini, G. Gastaldi, F. Solari, K. Pauwels, M. Van Hulle, J. Diaz,
[13] M. A. Turk, D. G. Morgenthaler, K. D. Gremban, and M. Marra, “VITS- E. Ros, N. Pugeault, and N. Krüger, “A compact harmonic code for early
a vision system for autonomous land vehicle navigation,” IEEE Trans. vision based on anisotropic frequency channels,” Comput. Vis. Image
Pattern Anal. Mach. Intell., vol. 10, no. 3, pp. 342–361, May 1988. Understanding, vol. 114, no. 6, pp. 681–699, Jun. 2010.
[14] S. Thrun, M. Montemerlo, H. Dahlkamp, D. Stavens, A. Aron, J. Diebel, [38] B. Leibe, K. Schindler, N. Cornelis, and L. Van Gool, “Coupled ob-
P. Fong, J. Gale, M. Halpenny, G. Hoffmann, K. Lau, C. Oakley, ject detection and tracking from static cameras and moving vehicles,”
M. Palatucci, V. Pratt, P. Stang, S. Strohband, C. Dupont, IEEE Trans. Pattern Anal. Mach. Intell., vol. 30, no. 10, pp. 1683–1698,
L. E. Jendrossek, C. Koelen, C. Markey, C. Rummel, J. van Niekerk, Oct. 2008.
E. Jensen, P. Alessandrini, G. Bradski, B. Davies, S. Ettinger, A. Kaehler, [39] B. Leibe, N. Cornelis, K. Cornelis, and L. Van Gool, “Dynamic 3d
A. Nefian, P. Mahoney, “Stanley: The robot that won the darpa grand scene analysis from a moving vehicle,” in Proc. IEEE Conf. Comput. Vis.
challenge,” J. Robot. Syst., vol. 23, no. 9, pp. 661–692, Sep. 2006. Pattern Recog., Minneapolis, 2007, pp. 1–8.
[Online]. Available: http://dx.doi.org/10.1002/rob.v23:9 [40] K. Pauwels, N. Krueger, M. Lappe, F. Woergoetter, and M. van Hulle, “A
[15] C. Urmson, J. Anhalt, D. Bagnell, C. Baker, R. Bittner, J. Dolan, cortical architecture on parallel hardware for motion processing in real-
D. Duggins, D. Ferguson, T. Galatali, C. Geyer, M. Gittleman, time,” J. Vis., vol. 10, no. 10, pp. 1–21, Aug. 2010.
S. Harbaugh, M. Hebert, T. Howard, A. Kelly, D. Kohanbash, [41] T. Zhang and C. Tomasi, “On the consistency of instantaneous rigid
M. Likhachev, N. Miller, K. Peterson, R. Rajkumar, P. Rybski, B. Salesky, motion estimation,” Int. J. Comput. Vis., vol. 46, no. 1, pp. 51–79,
S. Scherer, Y. Woo-Seo, R. Simmons, S. Singh, J. Snider, A. Stentz, Jan. 2002.
W. Whittaker, and J. Ziglar, “Tartan racing: A multi modal approach to [42] F. Mosteller and J. Tukey, Data Analysis and Regression: A Second
the DARPA urban challenge,” Defense Advanced Res. Projects Agency, Course in Statistics. Reading, MA: Addison-Wesley, 1977.
Arlington, VA, DARPA Tech. Rep., 2007. [43] W. Thompson and T. Pong, “Detecting moving-objects,” Int. J. Comput.
[16] B. D. Argall, S. Chernova, M. Veloso, and B. Browning, “A survey of Vis., vol. 4, no. 1, pp. 39–57, 1990.
robot learning from demonstration,” Robot. Auton. Syst., vol. 57, no. 5, [44] M. Bertozzi and A. Broggi, “Real-time lane and obstacle detection on the
pp. 469–483, 2009. system,” in Proc. IEEE Intell. Veh., 1996, pp. 213–218.
MARKELI Ć et al.: DRIVING SCHOOL SYSTEM: LEARNING BASIC DRIVING SKILL FROM TEACHER IN REAL CAR 1145
[45] E. Dickmanns, Dynamic Vision for Perception and Control of Motion. Lars Baunegaard With Jensen received the B.Sc.
Berlin, Germany: Springer-Verlag, 2007. and M.Sc. degrees in computer systems engineering
[46] M. Aly, “Real time detection of lane markers in urban streets,” in Proc. in 2004 and 2007, respectively, from the University
IEEE Intell. Veh. Symp., Jun. 4–6, 2008, pp. 7–12. of Southern Denmark, Odense, Denmark, where he
[47] J. F. Canny, “A computational approach to edge detection,” IEEE Trans. is currently working toward the Ph.D. degree with
Pattern Anal. Mach. Intell., vol. PAMI-8, no. 6, pp. 679–698, Nov. 1986. the Mærsk McKinney Møller Institute.
[48] I. Sobel and G. Feldman, A 3 × 3 Isotropic Gradient Operator for Image His research interests include real-time computer
Processing. Hoboken, NJ: Wiley, 1973. vision and graphics processing unit computing.
[49] R. E. Kalman, “A new approach to linear filtering and prediction
problems,” Trans. ASME, J. Basic Eng., vol. 82, no. 1, pp. 35–45, 1960.
[50] I. Markeli, “Teaching a robot to drive—A skill learning inspired
approach,” Ph.D. dissertation, Georg-August Univ. Göttingen, Göttingen,
Germany, 2010.
[51] M. A. Fischler and R. C. Bolles, “Random sample consensus: A para-
digm for model fitting with applications to image analysis and automated
cartography,” Commun. ACM, vol. 24, no. 6, pp. 381–395, 1981.
[52] Eisat Website, 2010. [Online]. Available: http://www.mi.auckland.ac.nz/ Nikolay Chumerin received the M.Sc. degree in
EISATS mathematics and educational science and the Higher
[53] D. H. Douglas and T. K. Peucker, “Algorithms for the reduction of the Educational Certificate of teacher in mathematics
number of points required to represent a digitized line or its caricature,” and computer science from Brest State University,
Cartographica, Int. J. Geogr. Inf. Geovis., vol. 10, no. 2, pp. 112–122, Brest, Belarus, in 1999. He is currently working
Dec. 1973. toward the Ph.D. degree with the Laboratorium
[54] L. Bottou and V. Vapnik, “Local learning algorithms,” Neural Comput., voor Neuro- en Psychofysiologie Medical School,
vol. 4, no. 6, pp. 888–900, Nov. 1992. Katholieke Universiteit Leuven (K.U.Leuven),
[55] J. C. Santamaria, R. S. Sutton, and A. Ram, “Experiments with rein- Leuven, Belgium.
forcement learning in problems with continuous state and action spaces,” His research interests include biologicallyinspired
Adapt. Behav., vol. 6, pp. 163–217, 1997. computer vision, machine learning, and brain com-
[56] W. D. Smart and L. P. Kaelbling, “Practical reinforcement learning puter interfacing.
in continuous spaces,” in Proc. 17th Int. Conf. Mach. Learn., 2000,
pp. 903–910.
[57] Future and Emerging Technologies (FET), Newsletter, Jan. 2010. [Online].
Available: ftp://ftp.cordis.europa.eu/pub/fp7/ict/docs/fet/fet-nl-06_en.pdf
[58] TV Report About the Drivsco System (in German), 2009.
[Online]. Available: http://www1.ndr.de/mediathek/index.html?media= Aušra Vidugiriene received the B.Sc. and M.Sc.
ndsmag2340 degrees in computer science in 2003 and 2005, re-
spectively, working on language technologies, from
Vytautas Magnus University, Kaunas, Lithuania,
where she is currently working toward the Ph.D.
Irene Markelić received the B.Sc. and M.Sc. de- degree.
grees in computer science from the University of Her research interests include signal processing
Koblenz-Landau, Koblenz, Germany, in 2002 and and analysis of driver’s behavior.
2005, respectively, and the Ph.D. degree in computer
science from Georg-August-University Göttingen,
Göttingen, Germany, in 2010.
She is currently a Postdoctoral Researcher with
the Institute of Physics 3, Georg-August-University
Göttingen. Her research interests include computer
vision, developmental robotics, cognitive skill learn-
ing in humans and robots, and data mining.
Minija Tamosiunaite received the Engineer
Diploma from Kaunas University of Technology,
Kaunas, Lithuania, in 1991 and the Ph.D. degree
from Vytautas Magnus University, Kaunas, in 1997.
Anders Kjær-Nielsen received the B.Sc. and M.Sc. Her research interests include machine learning
degrees in computer systems engineering, in 2004 and applications in robotics.
and 2007, respectively, from the University of South-
ern Denmark, Odense, Denmark, where he is cur-
rently working toward the Ph.D. degree with the
Mærsk McKinney Møller Institute.
His research interests include real-time process-
ing, computer vision, embedded systems, and field-
programmable gate arrays.
Marc Van Hulle received the M.Sc. degree in elec- Florentin Wörgötter received the Ph.D. degree
trotechnical engineering (electronics) and the Ph.D. from the University of Essen, Essen, Germany,
degree in applied sciences from the Katholieke Uni- in 1988.
versiteit Leuven (K.U.Leuven), Leuven, Belgium, He worked experimentally on visual cortex before
the Doctor Technics degree from Queen Margrethe he turned to computational issues at the California
II of Denmark, Odense, in 2003 from the Danmarks Institute of Technology, Pasadena, from 1988 to
Tekniske Universitet (DTU), Lyngby, Denmark, and 1990. After 1990, he became a Researcher with the
the Honorary Doctoral degree from Brest State Uni- University of Bochum, Bochum, Germany, where he
versity, Brest, Belarus, in 2009. was concerned with experimental and computational
He is currently a Full Professor with K.U.Leuven neuroscience of the visual system. Between 2000 and
Medical School, where he heads the Computational 2005, he was a Professor of computational neuro-
Neuroscience Group of the Laboratorium voor Neuro- en Psychofysiologie. In science with the Department of Psychology, University of Stirling, Stirling,
1992, he was with the Brain and Cognitive Sciences Department, Massachusetts U.K., where his interests strongly turned toward “learning in neurons.” Since
Institute of Technology, Boston, as a Postdoctoral Scientist. His research in- July 2005, he has been leading the Department for Computational Neu-
terests include computational neuroscience, neural networks, computer vision, roscience, Bernstein Center, Georg-August-University Göttingen, Göttingen,
data mining, and signal processing. Germany. His research interests are information processing in closed-loop
perception-action systems, including aspects of sensory processing, motor
control, and learning/plasticity. These approaches are tested in walking and
Norbert Krüger received the M.Sc. degree from driving robotic implementations. His group has developed the RunBot, which
Ruhr-Universität Bochum, Bochum, Germany, and is a fast and adaptive biped walking robot.
the Ph.D. degree from the University of Bielefeld,
Bielefeld, Germany.
He is currently a Professor with Mærsk McKinney
Møller Institute, University of Southern Denmark,
Odense, Denmark. He is a partner in several
European Union and national projects PACO-PLUS,
Drivsco, NISA, and Handyman. He is leading the
Cognitive Vision Laboratory, which focuses on com-
puter vision and cognitive systems, particularly the
learning of object representations in the context of grasping. He has also been
working on computational neuroscience and machine learning.