Professional Documents
Culture Documents
ABSTRACT With the prosperity of artificial intelligence, more and more jobs will be replaced by robots. The
future of precision agriculture (PA) will rely on autonomous robots to perform various agricultural operations.
Real time kinematic (RTK) assisted global positioning systems (GPS) are able to provide very accurate
localization information with a detection error less than ±2 cm under ideal conditions. Autonomously driving
a robotic vehicle within a furrow requires relative localization of the vehicle with respect to the furrow
centerline. This relative location acquisition requires both the coordinates of the vehicle as well as all the
stalks of the crop rows on both sides of the furrow. This extensive number of coordinate acquisitions of
all the crop stalks demand onerous geographical survey of entire fields in advance. Additionally, real-time
RTK-GPS localization of moving vehicles may suffer from satellite occlusion. Hence, the above-mentioned
±2 cm accuracy is often significantly compromised in practice. Against this background, we propose sets
of computer vision algorithms to coordinate with a low-cost camera (50 US dollars), and a LiDAR sensor
(1500 US dollars) to detect the relative location of the vehicle in the furrow during early, and late growth
season respectively. Our solution package is superior than most current computer vision algorithms used for
PA, thanks to its improved features, such as a machine-learning enabled dynamic crop recognition threshold,
which adaptively adjusts its value according to the environmental changes like ambient light, and crop size.
Our in-field tests prove that our proposed algorithms approach the accuracy of an ideal RTK-GPS on cross-
track detection, and exceed the ideal RTK-GPS on heading detection. Moreover, our solution package neither
relies on satellite communication nor advance geographical surveys. Therefore, our low-complexity, and
low-cost solution package is a promising localization strategy as it is able to provide the same level of
accuracy as an ideal RTK-GPS, yet more consistently, and more reliably, as it requires no external conditions
or hassle of the work demanded by RTK-GPS.
INDEX TERMS Computer vision, machine learning, navigation, precision agriculture, crop row detection,
localization, LiDAR.
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
74 VOLUME 1, 2020
FIGURE 1. Rx Robotics autonomous horticultural rover, the ‘bikebot’. The
rover is currently an electrically powered wheeled vehicle, which will be
remodeled with tracks.
VOLUME 1, 2020 75
LEVOIR ET AL.: HIGH-ACCURACY ADAPTIVE LOW-COST LOCATION SENSING SUBSYSTEMS
Task Sequencer). It is able to combinatively consider the out the crop pixels and a black-and-white (BW) binary image
factor of current position, speed, heading, etc., to make a is consequentially generated [45]. The common furrow edge
projection of the road boundary and create a road map for formulation methods following segmentation approach are the
the vehicle. Equally prominent is the research conducted Hough transform [48] and its variations [42], [46], [47], where
for vehicles travelling on the ‘Autobahns’ [13], [14] and all a best line of fit will be generated penetrating as many crop
the derived work [15], [16], including the EUREKA-project pixels as possible. However, segmentation with a constant
Prometheus [17], [18], which aimed at improving the traffic threshold is not very robust against inevitable environmental
safety by developing a system to warn the drivers when they changes, such as crop size and light conditions. To combat this
are off-track. Other distinguished research work focusing on challenge, an opening operation may be conducted to remove
computer-vision based road following include [19], [27], [31], the small bright pixels in the foreground [42] before a constant
[32], [36]. threshold is applied.
Admittedly, computer vision has been extensively used in Another commonly used crop row recognition approach is
self-driving and driver assistance systems in transportation as using column-wise comparison to decide the location of the
stated above. Nevertheless, operation in precision agriculture crop edges in an image. When this method is applied, the
poses new challenges to localization of autonomous naviga- by-column change of a certain indicator, normally the green
tion for the following reasons: value or a linear combination of RGB values, is recorded.
r The furrow is much narrower than a highway road, hence One branch of this method is to sum the indicator per column
allowing for less trajectory oscillation. and locate the crop rows either by finding the peaks of the
r The agricultural vehicle is uniquely equipped with two achieved waveform [41], [43] or the median value of the row
tracks as to reduce pressure to the soil, while most vehi- histogram [44]. Another branch divides the whole image into
cles in transportation have wheeled structures. strips and finds the furrow edge points per strip followed by a
r The aforementioned two major differences evoke a dif- linear regression to figure out the complete furrow edge [40].
ferent navigation algorithm [5], which relies on high- Nevertheless, all these algorithms have assumed a known
precision detection of cross-track (lateral) distance and row spacing [51]. Winterhalter et al. [51] proposed a Pattern
heading values of the rover with respect to the fur- Hough and Pattern RANSAC based crop row detection and
row centerline to produce a narrowly bounded trajectory localization algorithm, which detect all the rows in 3D cam-
without damaging the crops. era or laser images simultaneously by exhaustively searching
r Unlike most roads, the furrows do not have painted through all the possible combinations of the normal distances
lanes or physically existent boundaries. The scattered and angles of all the supposedly parallel and equidistant crop
bottom of each crop stalk forms an imagined boundary rows. However, this method has relatively high computation
of the furrow. Detection of these imagined boundaries complexity due to the exhaustive searching process. Besides,
are interfered with the ever growing crop leaves and ever more advanced sensors such as stereo cameras [49], infrared
changing ambient light. cameras [42], [50], or 3D laser scanners [51] may also be
r Unlike in transportation, there are much more furrows used to provide more informative raw images to facilitate the
present in one picture in agricultural fields. The irrele- localization process.
vant furrows may lead to compromised localization ac- Standing on the shoulders of all the above-mentioned and
curacy, if all the furrows are not equidistant and parallel. many other giants, we propose an alternative approach to
r The challenges presented in the above two items calls for detect the lateral and heading values of the rover in a low
a more robust and effective crop row detection algorithm. complexity manner with low cost 2d camera ($50) and LiDAR
r When the crops are taller than a certain height, the com- ($1500). The major differences between our findings and the
puter vision based algorithm may not be efficacious to above-mentioned previous excellent achievements lie in the
detect the crop rows and a LiDAR will be used instead following aspects. First, during early growth stage, we set
of a camera. a dynamic threshold to segment the crops, which is able to
These new challenges posed by precision agriculture have change with environment. The dynamic threshold is derived
inspired a large amount of research in crop row detection and from two inputs that can be read directly from the images.
autonomous rover localization. In 1987, Searcy et al. [37] These two inputs are indicative of the two most influential
presented the primary steps a computer-vision based crop-row varying environmental factors, the crop volume and the am-
sensing subsystem needs to include and serves as the foun- bient light intensity respectively. The exact formula of the
dation followed by many other researchers [38]–[40]. Most threshold as a function of the two inputs is obtained via a
monocular based computer vision solutions contain the fol- heuristic machine-learning approach. We denote this process
lowing fundamental steps: crop recognition, furrow edge for- as the adaptive RGB filtering algorithm in our paper.
mulation, perspective (camera-view to top-view) conversion, Second, while most of the aforementioned papers presented
cross-track and heading calculation. Crop recognition can be results only when the rover is aligned with the furrow, we did
done either via pixel segmentation or column-wise compar- extensive field tests with a large variety of preset cross-track
ison. Pixel segmentation is a series of operations, where a and heading offsets ranging from −16 cm to +16 cm and
threshold of a color (normally green) index is set to filter −20◦ to +20◦ , respectively. Through the test, we found the
76 VOLUME 1, 2020
detection results to be increasingly inaccurate as the cross- formulation methods are adopted to process the exper-
track offset gets bigger. The worsened detection error in mis- imental data and the results are compared.
aligned positions has led us to find out the non-negligible in- The rest of this paper is organized in the following way:
terference imposed by the body of the tall crops. We quantify The algorithm for early growth season will be introduced
the negative effect as a function of an input from the image in Section II. The proposed triangle filtering, the adaptive
via heuristic machine learning approach and compensate for RGB filtering algorithm, the anti-overinflation correction will
the interference with the aid of the formula. This process is be introduced in Section II-B, II-C1 and II-E. Late growth
denoted as the anti overinflation process in our paper. localization approach will be introduced in Section III. The
Third, we fuse the information from the two track tachome- results of all six combinations of the algorithms introduced in
ters with the previous location detection result to estimate the this paper will be presented in Section IV. The paper will be
current location of the rover, and calculate the predicted pixel concluded in Section V.
position of the centerline. This will help us filter out the un-
wanted furrows, which the rover is not currently between. We II. EARLY GROWTH SUBSYSTEM
name this strategy the triangle filtering in our paper. Fourth, The crops’ early growth stage is defined as the time from when
rather than an ordinary linear regression approach, we develop seeds germinate from the ground till when the height of the
an inverted linear regression approach to more accurately crops are approximately 45 cm. During this stage, the color
obtain the furrow edge line. Last but not least, during the late of the crops is visually distinguishable from the dirt around it,
growth stage, the crop rows are detected through horizontal and the stocks of crops are not tall enough to reflect the beams
scanning starting from the estimated centerline of the furrow emanated from the light-reflective sensors such as radar, sonar
in the 2d point cloud obtained by the LiDAR sensor. or LiDAR installed on the rover. Hence, image-based sensors
All of the above-mentioned meticulous operations result like a camera have a better opportunity to capture the edges of
in a greatly improved localization precision. Our field tests the furrow.
indicate that our solution achieves an average cross-track and Every 0.02 seconds, an image of the field is taken with a
heading detection error of 0.09 cm and 0.43◦ with a standard camera facing the direction of the rover’s heading and located
deviation of 2.53 cm and 0.83◦ under all light and crop-size on top of the rover. To best capture the information, we tilt the
circumstances and all possible preset positions during early camera with such an angle that the center of the camera will
growth stage, and an average detection error of 0.83 cm and face the point on the ground that is 2.25 meters in front. This
0.37◦ with standard deviation of 1.38 cm and 0.94◦ for cross- set up will ensure enough crop pixels present in the obtained
track and heading respectively during late growth stage. picture, yet is not that long to undermine the assumption that
This paper presents the methodologies and results of the the furrow edges appearing in the images are two straight
research continued from the IEEE proceeding paper [52]. lines.
The major improvements from the proceeding paper to this The overall flow chart of the entire computer-vision based
treatise are reflected in the following aspects: algorithm for early growth season is presented in Figure 3.
r This paper introduces the adaptive RGB filtering, which More details of each process will be given in the following
quantitatively evaluates the dynamic threshold that subsections.
changes based on the crop volume and ambient light
condition. A. IMAGE CAPTURE AND UNDISTORTION
r Triangle filtering is developed and presented in this pa- All cameras are subject to symmetrical radial distortion and
per, which can remove the irrelevant interfering crop tangential distortion [21], [22] caused by deviation from recti-
rows. linear projection and non-parallel alignment between the lens
r Inverted linear regression and Deming regression are and the image recording component [23]. The effects of these
introduced to acquire the crop row edges. distortions can be clearly observed in the raw image of a
r Anti overinflation adjustment is introduced to compen- chessboard taken with the camera used by the same group of
sate for the detection error incurred by the crop height. authors in this treatise, as seen in Figure 2(a) of [52], while
r While the conference paper only focused on the early the corrected image of it is shown in Figure 2(b) of [52].
growth stage, we introduce a LiDAR based low- On the one hand, the adjustments needed to apply to any
complexity localization algorithm for late growth stage pixel in the raw image caused by radial distortion can be
in this treatise. formulated as [24]–[26]:
r While the conference paper only presented results from
Crx δ (K r 2 + K2 r 4 + · · · )
preliminary test done with tapes on floor imitating the Cr = = x 1 2 , (1)
Cry δy (K1 r + K2 r 4 + · · · )
crop rows, we did extensive field tests throughout the
early and late growth stage of the crops, during different where Crx and Cry are the amount of revision needed to apply
times of the day, with a large number of combinations of to pixel (xd , yd ) in the raw distorted image in both axes, to
preset cross-track and heading values. correct for the radial distortion.
r On top of that, six combinations of two different In the equation, the location of the pixel (xd , yd ) in the raw
crop recognition approaches and three furrow edge image is alternatively defined by δx and δy , the horizontal and
VOLUME 1, 2020 77
LEVOIR ET AL.: HIGH-ACCURACY ADAPTIVE LOW-COST LOCATION SENSING SUBSYSTEMS
78 VOLUME 1, 2020
facilitate, the location detection process. Hence, as shown in
Figure 3, an extra step called ‘triangle filtering’ is applied
to filter out the interfering crop rows before any approach is
invoked to identify the crop row pixels in the image.
To do this, a priori location information of the rover is
required to estimate the pixel location of the centerline in the
image. To be more exact, the pixel location of the centerline in
the image is determined by the cross-track and heading errors
of the furrow. Hence, we will estimate the a priori cross-track
error ěk and heading error ˇ e,k of the rover at the current
moment using dead reckoning method, based on the location
of the rover at the previous moment and measurements of the
track speeds at the current moment.
More quantitatively, the heading rate of the rover in ra-
dian/sec can be calculated as [5]:
FIGURE 5. Rectified image of Figure 4. The blue dashed straight lines were
˙ = VL,act − VR,act ,
(4)
manually added by the authors. The blue dashed straight lines delineating
W
the crop rows were created by directly connecting the lower and upper where VL,act and VR,act are the measurements of the actual left
ending points of each crop row in the image with a straight line.
and right track speed in m/s obtained with encoders, and W
is the width of the rover. Hence, the estimated heading of the
rover at the current moment can be calculated as
k = k−1 +
˙ · dt, (5)
where dt is the sampling period. According to the rover kine-
matics [5], the fraction of the rover speed on x and y Cartesian
axes can be computed as:
VL,act − VR,act
Vk =
2
Vx = Vk sin(k )
Vy = Vk cos(k ), (6)
where Vk is the computed rover speed along its heading at
the current moment. Therefore, the location coordinates of the
rover at the current moment (Xk , Yk ) can be derived as:
FIGURE 6. Cropped image of Figure 5. The blue dashed straight lines were
manually added by the authors. The blue dashed straight lines delineating Xk = Xk−1 + Vx · dt
the crop rows were created by directly connecting the lower and upper
ending points of each crop row in the image with a straight line. Yk = Yk−1 + Vy · dt (7)
As a result, the a priori estimation of the cross-track error
and heading error of the rover at the current moment can be
blackout areas on the edges of the image, as shown in Figure 5. concluded as:
To reduce the interference from these blackout areas to the
true information conveyed by the images, the images need to ěk = Xk − X0
be cropped as shown in Figure 6. ˇ e,k = k − row ,
(8)
B. TRIANGLE FILTERING AND CENTERLINE CALCULATION where X0 and row are the x coordinate and heading of the
Intrinsically, there are multiple crop rows in each distortion- furrow centerline. Here we assume that X0 = 0 and row = 0
corrected and cropped picture as shown in Figure 6. However, for simplicity. That is, we assume that the centerline of the
only the furrow traversed by the rover is of interest and all furrow is aligned with the y axis. This assumption doesn’t af-
the location detection algorithms introduced in this treatise fect the development and application of the location detection
are developed to prevent the rover from running over the crop algorithms introduced in this treatise as the detection results,
row on either side of the interested furrow. Our experiments namely the cross-track and heading errors are relative position
have indicated that the inclusion of crop rows other than these information of the rover with respect to the centerline of the
two rows of concern in the image interfere with, rather than furrow.
VOLUME 1, 2020 79
LEVOIR ET AL.: HIGH-ACCURACY ADAPTIVE LOW-COST LOCATION SENSING SUBSYSTEMS
80 VOLUME 1, 2020
FIGURE 9. An unprocessed raw image taken in the field, as a reference for
post-process pictures.
FIGURE 10. The image of the raw picture in Figure 9 after distortion
To further visualize the filtering output of applying Pi,g > γ correction, triangle filtering and RGB filtering.
to the image P , we turn the original picture P into a BW image
PBW,γ , with all the pixels in set P>γ marked in white and all
the pixels in set P<γ in black. More quantitatively,
BW mean value would be the optimal value for BW mean
PBW,γ = PW,γ PB,γ (13) threshold of the image. More quantitatively, the BW mean
value of a BW image achieved after applying filter Pi,g > γ
where is defined as
|P |
PW,γ = PW,γ ,i = 255 ∀i : Pi,g ≥ γ Pi ∈ P PBW,γ ,i
m(γ ) = i=1 (15)
|PBW,γ |
PB,γ = PB,γ ,i = 0 ∀i : Pi,g < γ Pi ∈ P
As the BW mean m can be any real number ∈ [0, 255],
i = 1, 2, . . . |P | (14) exhaustively searching for the optimal value of it can be time
consuming and therefore not considered in real-time robotic
Figure 9 is an unprocessed raw image of the corn field. After
systems. We hereby propose a low-complexity algorithm to
applying distortion correction, triangle filtering and Pi,g > γ
find the suboptimal value of the BW mean and set it as the
filtering to Figure 9, the achieved BW image PBW,γ is de-
mean threshold, which is denoted as m̃ in this treatise. The
picted in Figure 10. More exactly, all the pixels that have
two major factors that affect the value of m̃ are the volume
Pi,g > γ are represented as white pixels in Figure 10, and the
of the crops and the ambient brightness. With denoting the
rest are in black.
crop-volume factor as α and the brightness factor as β, we
The mean value of the BW image PBW,γ is called the BW
assume that the BW mean threshold of each picture can be
mean value and is denoted as m, and m ∈ [0, 255]. Ideally, ev-
represented by
ery single pixel of the crop should be highlighted in white and
all other pixels should be in black. In this case, the resulting m̃ = α × β. (16)
VOLUME 1, 2020 81
LEVOIR ET AL.: HIGH-ACCURACY ADAPTIVE LOW-COST LOCATION SENSING SUBSYSTEMS
82 VOLUME 1, 2020
FIGURE 16. The average crop-volume factor ᾱd on each testing day.
VOLUME 1, 2020 83
LEVOIR ET AL.: HIGH-ACCURACY ADAPTIVE LOW-COST LOCATION SENSING SUBSYSTEMS
to keep the range close to the value range used for non-linear
regression to avoid possible inaccuracies from extrapolation.
The lower boundary of β is set to be 1/5000 at ḡ = 135 for
a couple of reasons. First, (135, 1/5000) is the vertex of the
parabola shown in Figure 17. As stated before, the brightness
factor function β = f (ḡ) needs to be monotonically decreas-
ing. Hence, the upper bound of the value range of ḡ cannot
exceed the value of the independent variable at the vertex of
the parabola. Second, we could have created another equation
of β for ḡ > 135. Nevertheless, this will cause the β values
to be less than 1/5000 and further incur insufficient pixels
to be selected as crops. In addition to the aforementioned
boundaries set to β, we have also imposed minimum and
maximum values allowed for m̃ to be 1 and 50 respectively.
This is to further safeguard the system against selecting far
too few or too many plant pixels.
All in all, the brightness factor β can be defined piecewise
FIGURE 19. The resultant furrow edge lines after linear regression is
in terms of average green value ḡ as: applied to all the detected plant pixels in the image of Fig. 9. All plant
⎧ pixels on the left side are highlighted in red and the right side in blue. The
⎪
⎪
10
, if ḡ ∈ (0, 110]
⎪
⎪
resultant furrow edge lines are marked in white.
⎪
⎪ 5000
⎨ 1 (ḡ − 135)2
β= +1 if ḡ ∈ (110, 135] (24)
⎪
⎪ 5000 70
⎪
⎪ recalculated using the same methodologies for different types
⎪
⎪ 1
⎩ , if ḡ ∈ (135, 255) of crops through training process.
5000
Based on Eq. (20) and (24), the calculated BW mean thresh-
2) CANNY EDGE DETECTION
olds m̃d on each testing day are shown in Figure 15 as the
yellow line with dot markers. As can be seen from the figure, A comprehensive description of the Canny edge detection
the calculated BW mean threshold values are very close to algorithm was addressed in [52], which will not be repeated
their optimal counterparts. During the actual application of here. Fig. 18 shows the Canny edge detection result after
the adaptive RGB filtering to detect the crop pixels, the mean applying triangle filtering on the input image shown in Fig. 9.
threshold m̃ of each picture will be calculated using Equation
(16) with (17) and (24). D. CALCULATING THE FURROW LINES
One thing worthy of note is that, equations in Sec. II-C1 After all the plant pixels have been picked out, the system
are developed in the context of corn field. Nevertheless, the classifies the marked pixels into left and right sides based on
formats of all the equations and steps of the algorithms will each pixel’s location in relation to the assumed centerline cal-
maintain the same for different types of crops. However, the culated earlier (see Equation (9) and Section II-B). As shown
thresholds and constants used in various steps and equations, in Fig. 19 and 20, the left side and right side categories are
including γ0 , γh and the constants in Eq. (24), will need to be colored in red and blue for display. Then the left and right
84 VOLUME 1, 2020
following equations [54]:
N
(xi − x̄)2
sxx = i=1
N −1
N
(xi − x̄)(yi − ȳ)
sxy = i=1
N −1
N
(yi − ȳ)2
syy = i=1 , (27)
N −1
and the slope and intercept in Equation (25) can be obtained
as:
syy − sxx + (syy − sxx )2 + 4sxy
2
m=
2sxy
N
N
FIGURE 20. All the resultant Hough lines detected in the image of Fig. 9. yi − m i=1 xi
Hough lines on the left side are highlighted in red and the right side in b = i=1 (28)
blue. The final furrow edge lines are marked in white.
N
VOLUME 1, 2020 85
LEVOIR ET AL.: HIGH-ACCURACY ADAPTIVE LOW-COST LOCATION SENSING SUBSYSTEMS
FIGURE 23. This chart shows the cross-track inaccuracy caused by the
final furrow lines being drawn on top of the plants rather than at the base.
FIGURE 21. The black lines were drawn on this image manually. They The green line was created by linearly regressing all the blue data in the
show the ideal case of where the final furrow lines should be placed, at figure. The equation for the green line is included at the upper right corner
the base of the crops. The red lines show the actual output from the of the chart.
system, where the lines are drawn on top of the crops rather than the base.
VOLUME 1, 2020 87
LEVOIR ET AL.: HIGH-ACCURACY ADAPTIVE LOW-COST LOCATION SENSING SUBSYSTEMS
FIGURE 26. An example of the raw data collected by the LiDAR sensor FIGURE 28. The kept data points after the closest points on each side of
represented in Cartesian format. This set of data has a known cross-track the estimated centerline are found and the additional filtering process is
error of 0 cm and a heading error of 0 degree. applied to the data in Figure 27.
be quantified as:
x =m×y+b (30)
where:
b = −ěk and m = − tan ˇ e,k (31)
The raw Cartesian data are then broken up into discrete sets
based on their y value, where each set contains all the points
having their y values within a 1 mm range. For example, the
first set contains all points with y values in [0,1) mm and the
next set would contain all points with y values in [1,2) mm.
This continues until y = 6 m, resulting in 6000 sets of points.
For each set, the algorithm finds the closest points from the
estimated centerline in the horizontal (x) direction on both
sides. If either or both of the points are within 450 mm of
FIGURE 27. Another example of the raw data collected by the LiDAR
sensor represented in Cartesian format. This set of data has a known
the estimated centerline in the x direction, they are kept as
cross-track error of 8 cm and a heading error of 15 degrees. desired data points, with all other points in the set thrown out.
The value of the threshold, i.e. 450 mm, is an empirical value
determined by the width of the furrows, which is 30 inches
(762 mm) here. The threshold of 450 mm is set so that if a
crop rows and are reflected back from the objects behind them. data point is further than 450 mm away from the centerline, it
These undesired data points need to be filtered out to avoid is likely not from the crops of interest, but something behind
interfering in the process of determining the centerline of the them, such as the crops in another row. The value of this
interested furrow. threshold should be readjusted when the algorithm is applied
To ferret out the desired data points from the whole set, to a crop field having different furrow width. An example of
the first step is to calculate the estimated centerline based the retained data points can be seen in Figure 28. After this
on the estimated cross-track and heading errors ěk and ˇ e,k . additional filtering process is completed, a line of best fit is
The process of estimating the a priori cross-track and heading created on each side of the estimated centerline using orthogo-
errors is the same as used in early growth season, which were nal regression. Lastly, the slope and x-intercept of each furrow
detailed in Section II-B from Equation (4) to (8). The equation edge is utilized to calculate the final heading and cross-track
of the estimated centerline during late-growth season can then values.
88 VOLUME 1, 2020
FIGURE 29. Please note values have been normalized for easier comparison and display. A vertical value of 1 corresponds to the following values:
average absolute cross-track detection error is 2.73 cm, average absolute heading detection error is 0.86◦ , standard deviation of cross-track detection
error is 3.73 cm, and standard deviation of heading detection error is 0.92◦ .
FIGURE 30. Comparison of histograms of the cross-track and heading detection errors, when a camera and an RTK-GPS are respectively used as the
sensor during the early growth season. When the camera is used, the adaptive RGB filtering and inverted linear regression are jointly adopted as parts of
the computer vision algorithm.
IV. RESULTS algorithm introduced in [5] and the computer vision algo-
In this section, Figures 29 and 30 present the experimental rithms introduced in this paper are jointly applied. In our ex-
results during the early growth season, while Figures 31 and periment, a camera with a part number ELP-USB30W04MT-
32 present the results during the late growth season. More- RL36, a LiDAR with a part number UST-10LX and a GPS
over, Figure 33 displays the closed-loop route taken by the with a part number Emlid Reach M+ were used to generate
rover given an offtrack initial position, when the navigation the results.
VOLUME 1, 2020 89
LEVOIR ET AL.: HIGH-ACCURACY ADAPTIVE LOW-COST LOCATION SENSING SUBSYSTEMS
FIGURE 31. A 3D surface showing the (a) cross-track and (b) heading detection errors over all possible rover locations during late growth season.
FIGURE 32. Comparison of histograms of the cross-track and heading detection errors when a LiDAR and an RTK-GPS are respectively used as the sensor
during late growth season.
A. EARLY GROWTH SEASON regression does a better job when one side of the furrow is
According to the flowchart in Figure 3, for early growth sea- brighter that the other. In this scenario, more pixels will be
son, there are 2 × 3 possible ways to implement the computer- marked as plants on one side, causing almost all of the Hough
vision based algorithm to calculate the cross-track and head- lines to be drawn on that side, with only a few lines left on the
ing errors, with choices between ‘RGB filtering’ and ‘Canny other side. On the other hand, the creation of the line of best
edge detection,’ and choices among ‘inverted linear regres- fit by regression is not influenced by the brightness difference
sion,’ ‘orthogonal regression’ and ‘Hough line detection’. between two sides.
In order to know which combination will achieve the best Also, both forms of regression showed similar results, but
results, all six combinations were implemented to process 575 the inverted linear regression turned out to be slightly more
raw pictures we collected at the Molitor Bros Farm in Min- accurate than orthogonal regression. The inverted linear re-
nesota. These pictures were taken at different times on differ- gression approach only tries to minimize the least mean square
ent days with different weather and light conditions through- errors from the resultant furrow edge lines to the scattered
out corn’s early growth season. The normalized average de- pixel points in the horizontal direction (see Section II-D). Due
tection error and standard deviation of each combination is to the narrow width of the furrow, all the possible headings
presented in Figure 29. a rover can have without damaging the crops are limited
The results indicate that regression was generally more in (−25, 25) degrees. The narrow heading range causes all
accurate than Hough line fitting. This is due to the fact that furrow edge lines to span from bottom to top of the picture
90 VOLUME 1, 2020
FIGURE 33. Top view of the paths of the rover in the furrow when a low-cost camera, a LiDAR and an RTK-GPS are used as the sensor respectively. The
initial position of the rover is a cross-track error of 20 cm to the left of and a heading of 2◦ counter-clockwise from the centerline. Parameters of the
dynamic-inversion based navigation algorithm are as follows: the damping ratio ζ = 0.8 and the decay length Y0 = 1.5m. Please refer to [5] for more
details of the navigation algorithm. The horizontal and vertical axes in (a) do not have equal scales, while those in (b) have equal scales.
in vertical direction, while only taking up a short span in growth is seen from 2.53 cm to 1.377 cm, while standard
horizontal direction, as can be evidenced in Figures 19∼22. deviation of heading offset increases slightly from 0.83◦ to
Therefore, minimizing the least mean square error in the hor- 1.2056◦ . Average absolute cross-track detection error goes
izontal direction is more important to final detection results from 1.998 cm to 1.206 cm, while average absolute heading
of cross-track and heading locations, as compared to a nor- error goes from 0.7379◦ to 0.8631◦ compared to early growth.
mal distance between the scattered plant pixel points to the The system is also much simpler and more immune to the
achieved furrow edge lines. interference from ambient light and crop heights.
As can be seen in Figure 29, the adaptive RGB filtering
and inverted linear regression produced the best results, so we C. OPEN AND CLOSED LOOP PERFORMANCE
have included a histogram of the distribution of cross-track COMPARISONS
and heading detection errors achieved by RGB filtering and The open-loop performance of the proposed computer vision
inverted linear regression in Figure 30. algorithms are compared with RTK-GPS during both the early
One thing worthy to note is that the camera used to take and late growth seasons in Figures 30 and 32, while the RGB
all the pictures had a problem causing the images to appear filtering and inverted linear regression are jointly used for
grey and monotone. We recognized this after the first few early growth detection.
tests, but by the time the manufacturer got back to us with a From both figures we can see that the accuracy of the RTK-
replacement the early growth season was over. With a largely GPS is slightly better than both the low-cost camera ($50)
reduced color gradient, the proposed algorithms still provide and the LiDAR when the proposed computer vision algorithm
the sensing subsystem with a detection accuracy close to GPS. is used to detect the cross-track locations, as evidenced by a
It is hence safe to say that the accuracy will be enhanced, if a smaller standard deviation achieved by the RTK-GPS. How-
better camera is used. ever, using standard deviation as the metric, our computer vi-
sion algorithms outperform the RTK-GPS in terms of heading
B. LATE GROWTH SEASON detection.
It can be seen from the late-growth results in Figure 31 that the The detection accuracy of the RTK-GPS presented in
cross-track detection error is bounded within −2 cm to +2 cm Figures 30 and 32 is only achievable during ideal conditions,
for all the preset locations of the rover, except a very slanted when constant high-quality satellite connection and accurate
position at +16 cm and −20◦ , while the heading detection geographical survey of all the crop stalks in the field are avail-
error is bounded within −2◦ to +1.5◦ . An improvement in able. Unfortunately, these two assumptions are challenging
standard deviation of cross-track values compared to early to meet in current agriculture environments, and hence the
VOLUME 1, 2020 91
LEVOIR ET AL.: HIGH-ACCURACY ADAPTIVE LOW-COST LOCATION SENSING SUBSYSTEMS
accuracy of RTK-GPS promised in Figures 30 and 32 are hard requires onerous advance geographical survey, the proposed
to achieve. On the contrary, the proposed computer vision computer vision algorithms are expected to achieve the above-
algorithms are not dependent on satellite signals or onerous mentioned accuracy over all kinds of weather conditions with-
geographical survey. Therefore the accuracy of the proposed out extra work.
computer vision algorithms as presented in Figures 30 and 32
is easier to achieve and for a longer period of time.
ACKNOWLEDGMENT
Moreover, the gap of the detection accuracies between the
The authors would like to express their deep gratitude to
RTK-GPS and the proposed computer vision algorithms ob-
Mr. Brian Molitor and Molitor Brothers Farm (Cannon Falls,
served in Figures 30 and 32 is diminished in the closed-loop
Minnesota) for their enormous support to this project. Special
simulation results shown in Figure 33. It may be worthy of
thanks also go to Mr. Scott Morgan and Rx Robotics (Red-
note that the horizontal and vertical axis of Figure 33(a) are
wing, Minnesota) for their consistent financial support.
not equally scaled. The horizontal scale is much bigger than
the vertical scale, hence the horizontal fluctuation of each
path and the differences among the three paths are exag- REFERENCES
gerated in Figure 33(a). If the same scales are applied (see [1] A. B. McBratney, B. Whelan, and T. Ancev, “Future directions
of precision agriculture,” Precis. Agriculture, vol. 6, pp. 7–23,
Figure 33(b)), the differences become negligible and all three 2005.
paths are straightly aligned with the centerline of the furrow [2] A. B. McBratney and M. J. Pringle, “Estimating average and pro-
after a small vibration. This proves that the proposed computer portional variograms of soil properties and their potential use in pre-
cision agriculture,” Precis. Agriculture, vol. 1, no. 2, pp. 125–152,
vision algorithms, when utilized with a low-cost 2D camera Sep. 1999.
or a LiDAR during early or late growth season, are able to [3] J. A. Howard and C. W. Mitchell, Phytogeomorphology, Hoboken, NJ,
provide the rover with the same route as the RTK-GPS, yet at USA: Wiley, 1985.
[4] T. C. Kaspar, T. S. Colvin, B. Jaynes, D. L. Karlen, D. E. James, and
a much lower realization complexity. D. W Meek, “Relationship between six years of corn yields and terrain
attributes,” Precis. Agriculture, vol. 4, pp. 87–101, 2003.
V. CONCLUSION [5] J. Hammer and C. Xu, “Dynamic inversion based navigation algorithm
for furrow-traversing autonomous crop rover,” in Proc. IEEE Conf.
This paper continues the work published in [52] and intro- Control Technol. Appl., Aug. 19–21, 2019, pp. 53–58.
duces advanced computer vision algorithms to detect the lo- [6] G. N. DeSouza and A. C. Kak, “Vision for mobile robot navigation:
cation of the rover when it is travelling inside a furrow. The A survey,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 24, no. 2,
pp. 237–267, Feb. 2002.
proposed solution package utilizes a low-cost camera and a [7] D. A. Pomerleau, “ALVINN: An autonomous land vehicle in a neural
LiDAR sensor during early and late growth season respec- network,” Carnegie Mellon University, Pittsburgh, PA, USA, Tech. Rep.
tively. Each sensor works with its own set of computer vi- CMU-CS-89–107, 1989.
[8] C. Thorpe, T. Kanade, and S. A. Shafer, “Vision and navigation for the
sion algorithms. The proposed computer vision algorithms for Carnegie-Mellon Navlab,” in Proc. Image Understand Workshop, 1987,
early-growth season are able to adaptively adjust the threshold pp. 143–152.
to segment the crops in the picture according to the ambient [9] C. Thorpe, M. H. Herbert, T. Kanade, and S. A. Shafer, “Vision and
navigation for the Carnegie-Mellon Navlab,” IEEE Trans. Pattern Anal.
light intensity and crop size at the time when the picture is Mach. Intell., vol. 10, no. 3, pp. 362–372, May 1988.
taken. Six combinations of crop recognition and furrow edge [10] T. M. Jochem, D. A. Pomerleau, and C. E. Thorpe, “Vision-based neural
formulation algorithms are proposed and applied to process network road and intersection detection and traversal,” in Proc. IEEE
Conf. Intell. Robots Syst., Aug. 1995, vol. 3, pp. 344–349.
the pictures taken at Molitor Bros Farm for optimal selection. [11] D. A. Pomerleau, “Efficient training of artificial neural networks
The combination of the adaptive RGB filtering and inverted for autonomous navigation,” Neural Comput., vol. 3, pp. 88–97,
linear regression turns out to provide the most accurate de- 1991.
[12] M. A. Turk, D. G. Morgenthaler, K. D. Gremban, and M. Marra,
tection results. The proposed computer vision algorithm for “VITS–A vision system for autonomous land vehicle navigation,”
the late-growth season is low-complexity and independent of IEEE Trans. Pattern Anal. Mach. Intell., vol. 10, no. 3, pp. 342–361,
environmental changes too. May 1988.
[13] E. D. Dickmanns and A. Zapp, “A curvature-based scheme for improv-
The in-field tests at Molitor Bros Farm have shown that the ing road vehicle guidance by computer vision,” in Proc. SPIE–Int. Soc.
proposed computer vision algorithms are able to provide a Photo-Opt. Instrum. Engineers, Oct. 1986, vol. 727, pp. 161–168.
standard deviation of 2.51 cm and 0.83◦ for cross-track and [14] E. D. Dickmanns and A. Zapp, “Autonomous high speed road vehicle
guidance by computer vision,” in Proc. 10th World Congr. Autom. Con-
heading detection during early growth season, and a standard trol, 1987, vol. 1, pp. 221–226.
deviation of 1.37 cm and 0.93◦ during late growth season. [15] E. D. Dickmanns, “Computer vision and highway automation,” Veh.
These are very close to the accuracy achieved by the RTK- Syst. Dyn., vol. 31, no. 5, pp. 325–343, 1999.
[16] E. D. Dickmanns and B. Mysliwetz, “Recursive 3D road and relative
GPS under ideal conditions, with its standard deviations of egostate recognition,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 14,
0.88 cm and 1.51◦ . The closed-loop simulation further proves no. 2, pp. 199–213, Feb. 1992.
that there is no noteworthy difference in the path made by [17] V. Graefe and K. Kuhnert, “Vision-based autonomous road vehicles,”
in Vision-Based Veh. Guidance, I. Masaki, Ed., Berlin, Germany:
the rover in the furrow, no matter whether the proposed sens- Springer-Verlag, 1992, pp. 1–29.
ing systems or the RTK-GPS is employed. The advantage of [18] U. Regensburger and V. Graefe, “Visual recognition of obstacles on
the proposed computer vision algorithms over RTK-GPS is roads,” in Proc. IEEE Int. Conf. Intell. Robots Syst., 1994, pp. 980–987.
[19] A. M. Waxman et al., “A visual navigation system for autonomous
their reliability and facility. That is, while the RTK accuracy land vehicles,” IEEE Trans. Robot. Autom., vol. 3, no. 2, pp. 124–141,
will degrade significantly during reduced GPS coverage and Apr. 1987.
92 VOLUME 1, 2020
[20] S. K. Kenue, “LANELOK: Detection of lane boundaries and vehicle [44] D. C. Slaughter, P. Chen, and R. G. Curley, “Vision guided precision
tracking using image-processing techniques–Parts I and II: Template cultivation,” Precision Agriculture, vol. 1, no. 2, pp. 199–216, 1999.
matching algorithms,” in Proc. SPIE, Mobile Robots IV, vol. 1195, [45] J. A. Marchant, “Tracking of row structure in three crops using image
Mar. 1990, doi: 10.1117/12.969886. analysis,” Comput. Electron. Agriculture, vol. 15, no. 2, pp. 161–179,
[21] C. C. Slama, Manual of Photogrammetry. 4th ed., Falls Church, Vir- Jul. 1996.
ginia: American Society of Photogrammetry, 1980. [46] T. Bakker et al., “A vision based row detection system for sugar beet,”
[22] J. Heikkila and O. Silven, “A four-step camera calibration procedure Comput. Electron. Agriculture, vol. 60, no. 1, pp. 87–95, Jan. 2008.
with implicit image correction,” in Proc. IEEE Comput. Soc. Conf. [47] V. Leemans and M.-F. Destain, “Line cluster detection using a variant of
Comput. Vis. Pattern Recognit., 1997, pp. 1106–1112. the Hough transform for culture row localisation,” Image Vis. Comput.,
[23] P. van Walree, “Distortion,” Photographic optics, Jan. 2009. [Online]. vol. 24, no. 5, pp. 541–550, 2006.
Available: https://web.archive.org/web/20090129134205/toothwalker. [48] J. A. Marchant, and R. Brivot, “Real-time tracking of plant rows us-
org/optics/distortion.html. Accessed on: Feb. 2, 2009. ing a hough transform,” Real-Time Image, vol. 1, no. 5, pp. 363–371,
[24] J. P. de Villiers, F. W. Leuschner, and R. Geldenhuys, “Centi-pixel accu- Nov. 1995.
rate real-time inverse distortion correction,” in Proc. SPIE Optomecha- [49] M. Kise, Q. Zhang, and F. R. Más, “A stereovision-based crop row
tronic Technol., Nov. 17, 2008, Paper 726611, doi: 10.1117/12.804771. detection method for tractor-automated guidance,” Biosystems Eng.,
[25] D. C. Brown, “Decentering distortion of lenses,” Photogrammetric vol. 90, no. 4, pp. 357–367, 2005.
Eng., vol. 32, no. 3, pp. 444–462, May 1966. [50] A. Pérez, F. López, J. Benlloch, and S. Christensen, “Colour and shape
[26] A. E. Conrady, “Decentred lens-systems,” Monthly Notices Roy. Astro- analysis techniques for weed detection in cereal fields,” Comput. Elec-
nomical Soc., vol. 79, pp. 384–390, 1919. tron. Agriculture, vol. 25, no. 3, pp. 197–212, 2000.
[27] Q. Li, L. Chen, M. Li, S.-L. Shaw, and A. Nchter, “A sensor-fusion [51] W. Winterhalter, F. V. Fleckenstein, C. Dornhege, and W. Burgard,
drivable-region andlane-detection system for autonomous vehicle nav- “Crop row detection on tiny plants with the pattern Hough trans-
igation in challenging road scenarios,” IEEE Trans. Veh. Technol., form,” IEEE Robot. Autom. Lett., vol. 3, no. 4, pp. 3394–3401,
vol. 63, no. 2, pp. 540–555, Feb. 2014. Oct. 2018.
[28] Y. Wang, E. Teoh, and D. Shen, “Lane detection and tracking using [52] D. Langan, R. Vraa, and C. Xu, “Machine-vision based location de-
B-Snake,” Image Vis. Comput., vol. 22, no. 4, pp. 269–280, Apr. 2004. tection solutions for autonomous horticulture rover during early growth
[29] D. Liu, B. Sheng, and F. Hou, “From wireless positioning to mobile season,” in Proc. IEEE Int. Conf. Intell. Saf. Robot., Aug. 24–27, 2018,
positioning: An overview of recent advances,” IEEE Syst. J., vol. 8, pp. 429–434.
no. 4, pp. 1249–1259, Dec. 2014. [53] D. A. Freedman, Statistical Models: Theory and Practice, Cambridge,
[30] K. Chiang, H. Chang, C.-Y. Li, and Y.-W. Huang, “An artificial neu- U.K.: Cambridge Univ. Press, 2009.
ralnetwork embedded position and orientation determination algorithm [54] W. E. Deming, Statistical Adjustment of Data. Hoboken, NJ, USA:
forlow cost MEMS INS/GPS integrated sensors,” Sensors, vol. 9, no. 4, Wiley, 1985.
pp. 2586–2610, Apr. 2009.
[31] V. Indelmanpini, G. Rivlin, and H. Rotstein, “Real-time vision-aided
localization and navigation based on three-view geometry,” IEEE Trans.
Aerosp. Electron. Syst., vol. 48, no. 3, pp. 2239–2259, Jul. 2012.
[32] T. B. Karamat, R. G. Lins, S. N. Givigi, and A. Noureldin, “Novel EKF-
based vision/inertial system integration for improved navigation,” IEEE
Trans. Instrum. Meas., vol. 67, no. 1, pp. 116–125, Jan. 2018. SAMUEL J. LEVOIR received the B.S. degree in
[33] E. Royer, M. Lhuillier, M. Dhome, and T. Chateau, “Towards an alter- computer engineering from the University of St.
native GPS sensor in dense urban environment from visual memory,” in Thomas in St. Paul, Minnesota USA in 2019. Dur-
Proc. Brit. Mach. Vis. Conf., 2004, pp. 197–206. ing his time at St. Thomas, his research focused
[34] S. Segvic, A. Remazeilles, A. Diosi, and F. Chaumette, “A mappin- on computer vision algorithms and machine learn-
gand localization framework for scalable appearance-based naviga- ing techniques. These studies were applied to the
tion,” Comput. Vis. Image Understanding, vol. 113, no. 2, pp. 172–187, project written about in this paper of navigating a
Feb. 2009. rover through a crop field as well as a laser scan-
[35] A. J. Davison, I. D. Reid, N. D. Molton, and O. Stasse, ning system which created 3D models of braille
“MonoSLAM:Real-time single camera SLAM,” IEEE Trans. Pattern documents. He also researched the development
Anal. Mach. Intell., vol. 26, no. 6, pp. 1052–1067, Jun. 2007. of efficient power usage algorithms in low voltage
[36] A. Vu, A. Ramanandan, A. Chen, J. Farrell, and M. Barth, “Real-time applications with a focus on dynamic voltage and frequency scaling. After
computer vision/DGPS-aided inertial navigation system for lane-level graduating from St. Thomas, he has been employed at Open Systems Interna-
vehicle navigation,” IEEE Trans. Intell. Transp. Syst., vol. 13, no. 2, tional (OSI) as a Software Developer, writing code to help maintain the power
pp. 899–913, Jun. 2012. grid.
[37] J. Reid and S. Searcy, “Vision-based guidance of an agriculture tractor,”
IEEE Control Syst. Mag., vol. 7, no. 2, pp. 39–43, Apr. 1987.
[38] J. Billingsley and M. Schoenfisch, “The successful development of a
vision guidance system for agriculture,” Comput. Electron. Agriculture,
vol. 16, no. 2, pp. 147–163, 1997.
[39] N. D. Tillett, T. Hague, and S. J. Miles, “Inter-row vision guidance for
mechanical weed control in sugar beet,” Comput. Electron. Agriculture, PETER A. FARLEY received the B.S. degree in
vol. 33, no. 3, pp. 163–177, Mar. 2002. computer engineering from the University of St.
[40] H. T. Søgaard and H. J. Olsen, “Determination of crop rows by image Thomas, St. Paul, Minnesota, USA in 2020. His
analysis without segmentation,” Comput. Electron. Agriculture, vol. 38, research interests include embedded systems, au-
no. 2, pp. 141–158, Feb. 2003. tonomous navigation, and operating system de-
[41] A. English, P. Ross, D. Ball, and P. Corke,“Vision based guidance for sign. While an undergraduate at the University of
robot navigation in agriculture,” in Proc. IEEE Int. Conf. Robot. Autom., St. Thomas, he designed software for automatic
2014, pp. 1693–1698. guidance and data collection and analytic in the
[42] B. Åstrand and A.-J. Baerveldt, “A vision based row-following sys- field of precision agriculture. He also researched
tem for agricultural field machinery,” Mechatronics, vol. 15, no. 2, and designed a proof of concept system to de-
pp. 251–269, Mar. 2005. tect anomalies in HVAC systems. He is currently
[43] H. J. Olsen, “Determination of row position in small-grain crops by working as a Software Engineer at Xirgo Technologies, where his work
analysis of video images,” Comput. Electron. Agriculture, vol. 12, no. 2, focuses on embedded software development in the field of vehicle and freight
pp. 147–162, Mar. 1995. monitoring.
VOLUME 1, 2020 93
LEVOIR ET AL.: HIGH-ACCURACY ADAPTIVE LOW-COST LOCATION SENSING SUBSYSTEMS
TAO SUN received the dual-B.S. degrees in elec- CHONG XU (Senior Member IEEE) received the
tronic engineering and computer science from the B.S. degree in electrical engineering from the Bei-
Huazhong University of Science and Technology jing University of Posts and Telecommunications
(HUST), China in 2004 and the Ph.D. degree (BUPT), China in 2004, M.S. and the Ph.D. de-
in electronic engineering from the University of gree in electrical engineering from University of
Southampton, UK in 2008. In 2010, he joined Re- Southampton, UK in 2006 and 2009 respectively.
search Laboratory of Electronics (RLE) at the Mas- In 2010, she joined Information Technology Lab-
sachusetts Institute of Technology (MIT), Cam- oratory at the National Institute of Standards and
bridge, MA, USA as a Research Fellow. In 2012, Technology (NIST), Department of Commerce,
he was employed as a System Engineer at Oracle Gaithersburg, MD, USA as a Research Fellow. Her
Inc. USA in high performance ZFS storage appli- work at NIST includes researches on smart grid,
ances. In 2015, he joined Inter Systems Inc. USA as a Kernel Architect scheduling/routing algorithms design and MIMO systems. Since 2012, she
in distributed system development. He is currently a System Design and has been employed as a full-time Faculty Member of School of Electrical and
Management (SDM) Fellow at Computer Science & Artificial Intelligence Computer Engineering at the University of St. Thomas in St. Paul, Minnesota,
Lab (CSAIL) and Sloan School of Management, Massachusetts Institute USA. Dr. Xu’s research interests include machine learning, internet-of-things
of Technology (MIT), Cambridge, MA, USA. Dr. Sun’s research interests (IoT), wireless communication, smart grid, robotics and heuristic optimiza-
are in artificial intelligence (AI), internet-of-things (IOT), smart sensors and tion algorithms.
robotics. He has published over 50 papers and patents with over 1800 citations
and i10-index of 19.
94 VOLUME 1, 2020