Professional Documents
Culture Documents
Article
Study on Body Size Measurement Method of Goat and Cattle
under Different Background Based on Deep Learning
Keqiang Li 1,2,3,4,† and Guifa Teng 5,6, *
1 School of Mechanical and Electrical Engineering, Hebei Agricultural University, Baoding 071001, China;
lkq910802@163.com
2 School of Mathematics and Information Technology, Hebei Normal University of Science and Technology,
Qinhuangdao 066004, China
3 Hebei Key Laboratory of Specialty Animal Germplasm Resources Exploration and Innovation,
Hebei Normal University of Science and Technology, Qinhuangdao 066004, China
4 Hebei Agricultural Data Intelligent Perception and Application Technology Innovation Center,
Hebei Normal University of Science and Technology, Qinhuangdao 066004, China
5 School of Information Science and Technology, Hebei Agricultural University, Baoding 071001, China
6 Hebei Key Laboratory of Agricultural Big Data, Baoding 071001, China
* Correspondence: tguifa@126.com
† Affiliations 3 and 4 are under planning.
Abstract: The feasibility of using depth sensors to measure the body size of livestock has been
extensively tested. Most existing methods are only capable of measuring the body size of specific
livestock in a specific background. In this study, we proposed a unique method of livestock body size
measurement using deep learning. By training the data of cattle and goat with same feature points,
different animal sizes can be measured under different backgrounds. First, a novel penalty function
and an autoregressive model were introduced to reconstruct the depth image with super-resolution,
and the effect of distance and illumination on the depth image was reduced. Second, under the
U-Net neural network, the characteristics exhibited by the attention module and the DropBlock were
adopted to improve the robustness of the background and trunk segmentation. Lastly, this study
Citation: Li, K.; Teng, G. Study on initially exploited the idea of human joint point location to accurately locate the livestock body feature
Body Size Measurement Method of
points, and the livestock was accurately measured. According to the results, the average accuracy
Goat and Cattle under Different
of this method was 93.59%. The correct key points for detecting the points of withers, shoulder
Background Based on Deep Learning.
points, shallowest part of the chest, highest point of the hip bones and ischia tuberosity had the
Electronics 2022, 11, 993. https://
percentages of 96.7%, 89.3%, 95.6%, 90.5% and 94.5%, respectively. In addition, the mean relative
doi.org/10.3390/electronics11070993
errors of withers height, hip height, body length and chest depth were only 1.86%, 2.07%, 2.42% and
Academic Editor: Amir Mosavi 2.72%, respectively.
Received: 7 March 2022
Accepted: 21 March 2022 Keywords: body size measurement; body segmentation; feature point location
Published: 23 March 2022
the feature points of the livestock. The average video duration of the respective livestock
was 120 s. The resolution of color video and depth video reached 1280 × 720 pixels, and
the frame rate of video was set to 15 frames/s. To expand the training set of semantic
segmentation model, this study used crawler technology to collect considerable images of
cattle and goats. Figure 1 illustrates the captured image.
Figure 1. Collected image: (A) is the image collected with the depth camera; (B) is the image collected
with the crawler.
At the end, two frames of images were randomly selected from each second of the
video and combined with the network images to construct the final data set. The data set
contains 9895 color images of cows and 9577 color images of goat. The ratio of the training
set, validation set, and test set is 5:3:2.
This experiment does not involve a livestock slaughter experiment, and the data
collection process of related livestock conformed to the regulations of the animal Ethics
Committee of Hebei Agricultural University.
height, chest depth, hip height and body length should be measured [17,18], as shown in
Figure 2B.
Figure 2. (A) Manual measurement of cattle and goat body dimensions; (B) Illustration of four
measured body dimensions: Line (a,f) is withers height, line; (b,e) is body length, line; (c,h) is chest
depth line; (d,g) is hip height.
Above, a is the highest point of the withers, b represents the anterior extremity of the
hummers, c is the shallowest part of the chest, d expresses the highest point of the hip
bones, and e is the extremity of the ischium. The straight line a–f represents withers height,
the vertical distance from the highest point of the withers to the ground surface at the level
of the forelegs. The straight line c–h expresses chest depth. It is the vertical distance from
the back to the shallowest part of the chest. The straight line d–g refers to hip height. It is
the vertical distance from the highest point of the hip bones to the ground surface at the
level of the hind legs. The straight line b–e denotes the body length; it is the distance from
the anterior extremity of the humerus to the posterior internal extremity of the ischium.
the real depth image via the bicubic interpolation. “Ω” denotes the points in the depth
image, and N(i) represents the points beside the point i in the depth image.
2
D = argminDi ∑
Di − Di0
+ λ∑ Φ Di − ∑ Aij Di
2
Ω
i ∈Ω j ∈ N (i )
First, the penalty function Φ(x) was adopted to act on the regular term, as
expressed below:
x
Φ (x) = − x
e +x
Φ(x) denotes the M-estimator. As reported by Smoli [28], robust M-estimators could
eliminate marginal uncorrelated patch. Figure 3 plots the L2 penalty function curve and
the penalty function curve applied in this study. The results indicated that the penalty
function proposed in this study was capable of efficiently saturating in the iteration, making
Φ(x) an approximation of L0 to keep the sharp edges of the image. Compared with the
penalty function G(x) = 2 ∗ 1 − exp − x2 /2 applied in reference [29], the curve of Φ(x)
achieved a faster saturation rate, so it would be more sensitive to outliers (e.g., noises
and void).
Figure 3. The curves of different penalty functions: The penalty function Φ(x) proposed in this study
achieved the fastest saturation rate.
Second, to increase the reconstruction effect of color images with unclear edges, the
autoregressive coefficient Aij is expressed as:
1 D0 C I
Aij = A F Ai,j , Ai,j
s i,j
0
D was defined by the Gaussian
where s denotes the normalization factor, the depth term Ai,j
filter, and the color term and the infrared term were, respectively, set as:
2
C h − ρh
∑
h∈c i
B ◦ ρ i j
2
C
Ai,j = exp
σ12
2
I h h
∑h∈c
Bi
◦ ρi − ρ j
I 2
Ai,j = exp
σ12
Electronics 2022, 11, 993 6 of 22
where σ1 denotes the decay rate of the exponential function; ρih represents the mask operator
with the size ω ∗ ω centered on the pixel point within the color channel h. According to
the extracted operators, the bilateral filter BiC was adopted to adjust the distance and
intensity difference.
∑h∈c Cih − Cjh
!
2
k i − j k
BiC (i, j) = exp 2
exp−
σ22 σ22
The filters BiC employed a bilateral kernel to weigh the distance of the local patch.
Compared with the Gaussian kernel adopted by the standard filter, the color term and the
infrared term showed a strong response to the pixels x with similar results. Accordingly,
the regression prediction of image edge pixels was significant.
Lastly, the check function F ( x ) was introduced in the regular term, as expressed below:
(
C , LBP (C ) ≥ LBP ( I )
Ai,j
F(x) = I , LBP (C ) ≺ LBP ( I )
Ai,j
In this study, the LBP equation is written as follows: LBP(C ) represents the ρ sampling
point in the circular area; R is the radius of the circular area; G (m) expresses the gray value
of the central pixel, and G ( p) denotes the gray value of the ρ circular boundary pixel.
ρ
LBPP,R ( xc , yc ) = ∑ s(G(ρ) − G(m)) ∗ 2ρ
n =1
Thus, in the regular term, the LBP function was adopted to judge the edge features
of the color image and the infrared image, a high-quality mask operator was selected
to act on the super-resolution model of the depth image. The effect of illumination and
distance on the depth image was solved effectively. The texture copy artifacts and the
depth discontinuities decreased significantly. Figure 5 gives the super-resolution result of
depth image.
Electronics 2022, 11, 993 7 of 22
Figure 5. Depth image restoration: (A) is the original depth image; (B) represents the reconstructed
depth image. Texture copy artifacts and depth discontinuities are significantly reduced.
In the real-world experiment, since there were no standard depth maps as reference,
Peak Signal-to-Noise Ratio (PSNR) and Structural Similarity (SSIM) could not be used to
measure the quality of image reconstruction. In order to further verify the effectiveness
of this method, Mean Gradient (MG) and Laplace Operator (LO) [31] were selected as
evaluation indexes in this study. The higher MG and LO values were, the better image
quality was. The formula of MG was as follows:
1
MG = (M −1)( N −1)
∗
M −1 N −1
q
( F (i,j)− F (i +1,j))2 +( F (i,j)− F (i,j+1))2
∑ ∑ 2
i =1 J =1
In the formula, the gray value of row i and column j in the image was expressed as
F (i, j); M and N were the total number of rows and columns of the image. The encapsulation
method of Laplace Operator was provided in OpenCV, which could be called directly. The
formula of LO was as follows:
LO = ∑ ∑| G ( x, y)|
x y
First, the U-Net neural network employed ReLU as the activation function. Since
the negative semi-axis value of ReLU was always zero, when the model inputted the
minimum value, the neuron would die because it was not activated. To solve the gradient
disappearance attributed to the ReLU activation function, the ReLU activation function after
each convolutional layer and Batch Normalization layer in the U-Net neural network was
replaced with the Mish activation function. The function of Mish and ReLU are illustrated
in Figure 7.
Since the positive semi-axis derivative of the Mish was greater than 1, it avoided the
saturation because of the function capping, while preventing the gradient from disappear-
ing. The negative semi-axis of the Mish was not always zero, whereas it remained a small
value, which solved the problem of neuron death; moreover, Mish improved the sparse
expression of the network and addressed the gradient disappearance attributed to the lack
of training data and the increase in network layers. According to existing studies, a smooth
activation function can more effectively allow the information to penetrate the neural
network [32]. Better generalization results can be achieved since the Mish is smoother than
that of ReLU.
Electronics 2022, 11, 993 9 of 22
Electronics 2022, 11, x FOR PEER REVIEW 9 of 22
Second, this method introduced the attention mechanism to increase the accuracy
of semantic
suggested thatsegmentation. The attention
avoiding dimension mechanism
reduction helpsrefers to effective
learn a data processing
channelmethod
attention, and
in machine learning, originating from natural language processing. Relevant researchers
appropriate cross-channel interaction is capable of reducing model complexity while
suggested that avoiding dimension reduction helps learn effective channel attention, and
keeping performance
appropriate cross-channel[33].
interaction is capable of reducing model complexity while keep-
Accordingly,
ing performance the model could add the attention mechanism module after the respec-
[33].
tive down-sampled convolution
Accordingly, the model could addmodule. First,mechanism
the attention the binary adaptive
module mean
after the method was
respective
adopted to calculate
down-sampled and pool
convolution theFirst,
module. input thedata to adaptive
binary a size ofmean
1 × 1.method
After was
the global
adoptedaverage
to calculate
pooling, theand
1D pool the inputwith
convolution data weight of 1 × 1.characteristics
to a sizesharing After the globalwasaverage pooling,
exploited to capture
the 1D convolution with weight sharing characteristics was exploited to capture
the cross-channel interaction information. Next, the nonlinear relation was introduced the cross-
channel the
through interaction
sigmoidinformation. Next, theLastly,
activation function. nonlinear
the relation
result waswasmultiplied
introducedby through
elements and
the sigmoid activation function. Lastly, the result was multiplied by elements and then
then expanded into the same dimension as the input data, so the dimension of the data
expanded into the same dimension as the input data, so the dimension of the data processed
processed by themechanism
by the attention attention module
mechanism module
remained remained
unchanged. Byunchanged.
introducing theBy attention
introducing the
attention
module, themodule,
accuracythe accuracy
of semantic of semantic
segmentation segmentation
tasks tasks
was significantly was significantly
improved. Figure 8 im-
proved. Figure 8 presents the attention
presents the attention mechanism module. mechanism module.
1×1×C 1×1×C
Figure
Figure 8. ECA module.
ECA module.
Lastly, to achieve the desired segmentation effect more accurately and efficiently,
the method added the DropBlock2D module during each up sampling. The DropBlock
module [34] refers to a regularization module for the convolutional neural network. The
effect is illustrated in Figure 9. Such a module removed an area around a feature and forced
the network to learn other features of the object after multiple iterations of training.
Lastly, to achieve the desired segmentation effect more accurately and efficiently, the
method added the DropBlock2D module during each up sampling. The DropBlock mod-
ule [34] refers to a regularization module for the convolutional neural network. The effect
is illustrated in Figure 9. Such a module removed an area around a feature and forced the
network to learn other features of the object after multiple iterations of training.
After processing by the neural network model proposed in this study, the background
and body feature regions of cattle and goats could be effectively segmented. The segmenta-
tion effect is given in Figure 10.
Figure 9. DropBlock2D module.
After processing by the neural network model proposed in this study, the back-
ground and body feature regions of cattle and goats could be effectively segmented. The
segmentation effect is given in Figure 10.
Electronics 2022, 11, 993 10 of 22
To further verify the effectiveness of the method, LabelImg was used to manually
annotate the images to obtain the real segmentation. Fifty images of cattle and goats were
randomly adopted from the test set, respectively, and the accuracy, the sensitivity, the
specificity and the dice similarity coefficient (dice) [35,36] were employed to assess to the
effect of the body-part segmentation method.
TP + TN
Accuracy =
TP + FP + TN + FN
TP
Sensitivity =
TP + FN
TN
Speci f icity =
TN + FP
2 ∗ TP
dice =
( TP + FN ) + ( TP + FP)
where TP represents the number of body-part pixels correctly identified as body-part
segmentation method compared to the real segmentation by LabelImg. FP represents
the number of body-part pixels correctly identified as Body-part segmentation method
compared to the pixels as non-body parts identified by LabelImg. TN represents the
number of body-part pixels incorrectly identified as the body-part segmentation method
compared to the pixels as non-body parts identified by LabelImg. FN denotes the number
of body-part pixels incorrectly identified as body-part segmentation method compared to
the real segmentation by LabelImg.
Since Dice is sensitive to internal filling, Hausdorff distance (HD) [37,38] is introduced
to evaluate the segmented boundary.
HD(A, M) = max max[d( ai , M )], max d m j , A
i
j
Dense Block
The stacked hourglass network comprises residual blocks. In this study, dense
blocks [44] were selected to replace residual blocks. Compared with residual blocks, dense
block networks exhibited a simple structure and fewer parameters than ResNet. Second,
dense blocks could send information between any two layers, thereby enhancing the reuse
of features and alleviating the problems of gradient vanishing and model degradation.
In a dense block, any two layers were connected to each other. The input of each
layer in the network was the output of all previous layers, and the feature map learned
by this layer would act as the input of the later network layer. The dense block structure
is expressed as xn = hn ([ x0 , x1 , x2 . . . xn−1 ]), where xn denotes the input of the layer n, [ ]
represents splicing, combining all output feature graphs from x0 to xn−1 layers according
to channels, and hn expresses a composite function (i.e., Batch Norm, activation function
and a 3 ∗ 3 convolution).
The Huber loss is essentially an absolute error, only becoming a squared error when
the error was small. The hyper parameter δ was used to control when it became a quadratic
error. When the value of the Huber loss function was in the interval [0 − δ, 0 + δ], the
Huber loss function was equal to mean square error. When the value of Huber loss function
was in the interval [−∞, δ] or the interval [δ, + ∞], the Huber loss function was equated
with Mean Absolute Error. Accordingly, the Huber loss function combined the advantages
of MSE and MAE, and it would be more robust to abnormal points.
The multistage densely stacked hourglass network refers to a serial stack of hourglass
sub-networks; however, with the increase in the network depth, the parameters increased,
making it difficult to perform effective training updates. To solve such a problem, this
study adopted the relay supervision strategy in the model for the supervised training. The
experiments in this study were all performed on a three-stage stacked hourglass network.
Figure 12 gives the four-level densely stacked hourglass network.
Figure 12. The structure of the three-level densely stacked hourglass network.
After the original image was preprocessed, the scale information of the image was
preserved for the next layer of feature fusion. In addition, the preprocessed image was
input into the hourglass sub-networks for a down sampling operation. After three residual
modules and two convolution modules, the features were extracted step by step, and the
feature map of the layer was obtained through the up-sampling operation. At the same
time, after up sampling operation, the scale information of the upper layer was fused and
used as the input of the next hourglass sub-networks to extract the features.
This study adopted the PyTorch framework for training, the input image was a color
image with a resolution of 768 × 768, and the number of iterations was set to 16,000.
The training set, validation set, and test set were divided according to the proportion of
5:3:2. For every 10,000 steps, the learning rate declined to 10%. Moreover, to increase the
convergence speed, the normalization operation was added to the model.
Goat Cattle
MG LO MG LO
JBU 0.0027 22.646 0.0028 21.404
WLS 0.0029 21.967 0.0029 22.142
AR 0.0032 74.482 0.0032 86.814
Ours 0.0048 142.170 0.0050 154.748
The results showed that the depth image reconstruction effect of AR algorithm was
better than JBU algorithm and WLS algorithm, but there were still deep texture copy
artifacts in the reconstructed depth image. In this research, making full use of the texture
features of infrared image could greatly improve the reconstruction quality of depth images.
The effect of illumination and distance on the depth image was solved effectively. The
texture copy artifacts and the depth discontinuities decreased significantly. Compared with
other algorithms, our method improved both visual effect and image quality.
Table 2. Body-part detection performance parameters for side view of cattle (%) *.
Table 3. Body-part detection performance parameters for side view of goats (%) *.
Table 4. Hausdorff distance of body segmentation between cattle and goat (mm).
Evaluation used the percentage of correct key points (σ = 0.5) [39,51] metric which
reports the percentage of detections that fall within a normalized distance of the ground
truth. In this study, because some pictures did not collect the head information of livestock,
Electronics 2022, 11, 993 15 of 22
the body length of livestock was selected to calculate the normalized distance between the
predicted feature point and the real feature point.
This study used the self-built data set to train the multistage densely stacked hourglass
network. The evaluation reported that the model exhibited high universality and could
accurately locate the feature points of cattle and goats. Figure 14 presents the location effect.
Figure 14. (A): The blue points are the manually marked feature points and the orange points are the
feature points identified by the algorithm. From left to right and from top to bottom of the right half
of Figure 14A is an enlarged image of the corresponding positions of points a, d, e, b and c in (B):
Body size measurement results.
those of the surrounding pixel points, resulted the great error in body length measurement.
The mean relative errors of withers height, hip height, body length and chest depth were
0.68%, 0.68%, 1.39% and 1.60% for cattle, and 0.65%, 1.06%, 1.77% and 1.99% for goat,
respectively. The average relative error of the measurement results was between −2% to
2%; therefore, the results of the automatic measurement could be trustworthy in terms of
the measurement accuracy.
Figure 15. Comparison of manual measurement and the proposed method for the four different cattle
body parameters: (A) Withers height, (B) hip height, (C) body length and (D) chest depth.
Figure 16. Comparison of manual measurement and the proposed method for the four different goat
body parameters: (A) Withers height, (B) hip height, (C) body length and (D) chest depth.
Electronics 2022, 11, 993 17 of 22
Figure 17. Box plots of the absolute error: (A) Box plots of the absolute error of the measurement of
the cattle body size; (B) Box plots of the absolute error of the measurement of the goat body size.
Figure 18. Box plots of relative error: (A) Box plots of the relative error the measurement of the cattle
body size; (B) Box plots of the relative error the measurement of the goat body size.
body measurements was greater than 0.9, thereby showing good measurement repeatability.
The ICC of withers height and hip height of cattle and withers height of goats all exceeded
0.980, which demonstrated higher reliability than other. Using the automatic measurement
to measure body size of cattle, the value of CV ranged from 0.049% to 0.090% (variance
ranges from 6.91 cm to 10.83 cm). The range of CV for goat body measurements ranged
from 0.043% to 0.141% (variance ranges from 2.79 cm to 4.43 cm). According to Fischer,
measurement methods with the repeatability and reproducibility had a CV less than 4%, as
in the present study were considered promising [53].
ICC CV
Cattle Goat
Cattle Goat
Manual Automatic Manual Automatic
withers height 0.990 0.985 0.064% 0.065% 0.052% 0.054%
hip height 0.985 0.943 0.052% 0.049% 0.041% 0.043%
body length 0.973 0.890 0.068% 0.069% 0.056% 0.064%
chest depth 0.972 0.942 0.091% 0.090% 0.123% 0.141%
* ICC describes the reliability and repeatability of measurements for clusters. The value of ICC close to 1 indicates
high reliability and repeatability. CV is a normalized measure of the dis-persion of probability distribution, which
is defined as the ratio of standard deviation to average.
Table 5 lists the body size measurement result of cattle, better than that of goats.
Besides the effect of fur, another reason was that during the depth data collection, the
behaviors of goats were more active than cattle; therefore, the distance between livestock
and depth camera is not consistent. In many goat images, the area occupied by the torso is
smaller than that of cattle; thus, there were more holes and texture copying artifacts in the
depth images. The mentioned holes and texture replication artifacts would exist around
the feature points, and would also affect the background segmentation and feature point
extraction even after the depth image super-resolution reconstruction.
and repeatability of the measurement of with height, hip height body length and chest
depth were better.
This study used deep learning to estimate the four body measurements of cattle and
goat, and it achieved good results; it proved that the deep learning method could be
applied to livestock trunk segmentation, body feature point location, and livestock body
size measurement. The future research is mainly divided into two parts. First, due to the
limitation of the number of data set, the current research object was limited to cattle and
goat. In future research, more types of livestock will be collected and annotated to achieve
the body measurement of different kinds of livestock (even horses and donkeys) with the
same feature point. Second, only the side views of cattle and goat have been studied. When
the side view of livestock cannot be collected completely, the measurement accuracy will
be reduced; therefore, more angles of livestock images (top view or images from different
angles) will be added in future research to further study the normalization of livestock
posture and measure more livestock body sizes. Finally, as the preliminary research, only
image processing was supported at present. In future research, a video stream will be used
to replace images to complete the measurement of livestock.
Author Contributions: Methodology, K.L.; validation, K.L.; resources, G.T.; writing—original draft
preparation, K.L. All authors have read and agreed to the published version of the manuscript.
Funding: This work was supported by the National Natural Science Foundation of China (Grant No.
U20A20180), the Graduate Innovation Foundation of Hebei (Grant No. CXZZBS2021045).
Data Availability Statement: This experiment does not involve a livestock slaughter experiment,
and the data collection process of related livestock conformed to the regulations of the animal Ethics
Committee of Hebei Agricultural University. And this study not involving humans.
Conflicts of Interest: The authors declare that they have no known competing financial interests or
personal relationships that could have appeared to influence the work reported in this paper.
References
1. Thorup, V.M.; Edwards, D.; Friggens, N.C. On-farm estimation of energy balance in dairy cows using only frequent body weight
measurements and body condition score. J. Dairy Sci. 2012, 95, 1784–1793. [CrossRef] [PubMed]
2. Pezzuolo, A.; Guarino, M.; Sartori, L.; González, L.A.; Marinello, F. On-barn pig weight estimation based on body measurements
by a Kinect v1 depth camera. Comput. Electron. Agric. 2018, 148, 29–36. [CrossRef]
3. Kuzuhara, Y.; Kawamura, K.; Yoshitoshi, R.; Tamaki, T.; Sugai, S.; Ikegami, M.; Kurokawa, Y.; Obitsu, T.; Okita, M.; Sugino, T. A
preliminarily study for predicting body weight and milk properties in lactating Holstein cows using a three-dimensional camera
system. Comput. Electron. Agric. 2015, 111, 186–193. [CrossRef]
4. Menesatti, P.; Costa, C.; Antonucci, F.; Steri, R.; Pallottino, F.; Catillo, G. A low-cost stereovision system to estimate size and
weight of live sheep. Comput. Electron. Agric. 2014, 103, 33–38. [CrossRef]
5. Leonard, S.M.; Xin, H.; Brown-Brandl, T.M.; Ramirez, B.C. Development and application of an image acquisition system for
characterizing sow behaviors in farrowing stalls. Comput. Electron. Agric. 2019, 163, 104866. [CrossRef]
6. Brandl, N.; Rgensen, E. Determination of live weight of pigs from dimensions measured using image analysis. Comput. Electron.
Agric. 1996, 15, 57–72. [CrossRef]
7. Marchant, J.A.; Schofield, C.P.; White, R.P. Pig growth and conformation monitoring using image analysis. Anim. Sci. 1999, 68,
141–150. [CrossRef]
8. Tasdemir, S.; Urkmez, A.; Inal, S. Determination of body measurements on the Holstein cows using digital image analysis and
estimation of live weight with regression analysis. Comput. Electron. Agric. 2011, 76, 189–197. [CrossRef]
9. Ozkaya, S. Accuracy of body measurements using digital image analysis in female Holstein calves. Anim. Prod. Sci. 2012, 52,
917–920. [CrossRef]
10. Federico, P.; Roberto, S.; Paolo, M.; Francesca, A.; Corrado, C.; Simone, F.; Catillo, G. Comparison between manual and
stereovision body traits measurements of Lipizzan horses. Comput. Electron. Agric. 2015, 118, 408–413.
11. Shi, C.; Teng, G.H.; Li, Z. An approach of pig weight estimation using binocular stereo system based on LabVIEW. Comput.
Electron. Agric. 2016, 129, 37–43. [CrossRef]
12. Salau, J.; Haas, J.H.; Junge, W.; Thaller, G. A multi-Kinect cow scanning system: Calculating linear traits from manually marked
recordings of Holstein-Friesian dairy cows. Biosyst. Eng. 2017, 157, 92–98. [CrossRef]
13. Wang, K.; Guo, H.; Ma, Q.; Su, W.; Zhu, D.H. A portable and automatic Xtion-based measurement system for pig body size.
Comput. Electron. Agric. 2018, 148, 291–298. [CrossRef]
Electronics 2022, 11, 993 21 of 22
14. Shi, S.; Yin, L.; Liang, S.H.; Zhong, H.J.; Tian, X.H.; Liu, C.X.; Sun, A.; Liu, H.X. Research on 3D surface reconstruction and body
size measurement of pigs based on multi-view RGB-D cameras. Comput. Electron. Agric. 2020, 175, 105543. [CrossRef]
15. Ruchay, A.; Kober, V.; Dorofeev, K.; Kolpakov, V.; Miroshnikov, S. Accurate body measurement of live cattle using three depth
cameras and non-rigid 3-D shape recovery. Comput. Electron. Agric. 2020, 179, 105821. [CrossRef]
16. Lina, Z.A.; Pei, W.B.; Tana, W.C.; Xinhua, J.D.; Chuanzhong, X.E.; Yanhua, M.F. Algorithm of body dimension measurement and
its applications based on image analysis. Comput. Electron. Agric. 2018, 153, 33–45. [CrossRef]
17. Guo, H.; Ma, X.; Ma, Q.; Wang, K.; Su, W.; Zhu, D.H. LSSA_CAU: An interactive 3d point clouds analysis software for body
measurement of livestock with similar forms of cows or pigs. Comput. Electron. Agric. 2017, 138, 60–68. [CrossRef]
18. Guo, H.; Li, Z.B.; Ma, Q.; Zhu, D.H.; Su, W.; Wang, K.; Marinello, F. A bilateral symmetry based pose normalization framework
applied to livestock body measurement in point clouds. Comput. Electron. Agric. 2019, 160, 59–70. [CrossRef]
19. Zhao, K.X.; Li, G.Q.; He, D.J. Fine Segment Method of Cows’ Body Parts in Depth Images Based on Machine Learning. Nongye
Jixie Xuebao 2017, 48, 173–179.
20. Jiang, B.; Wu, Q.; Yin, X.Q.; Wu, D.H.; Song, H.B.; He, D.J. FLYOLOv3 deep learning for key parts of dairy cow body detection.
Comput. Electron. Agric. 2019, 166, 104982. [CrossRef]
21. Li, B.; Liu, L.S.; Shen, M.X.; Sun, Y.W.; Lu, M.Z. Group-housed pig detection in video surveillance of overhead views using
multi-feature template matching. Biosyst. Eng. 2019, 181, 28–39. [CrossRef]
22. Song, X.; Bokkers, E.; Mourik, S.V.; Koerkamp, P.; Tol, P. Automated body condition scoring of dairy cows using 3-dimensional
feature extraction from multiple body regions. J. Dairy Sci. 2019, 102, 4294–4308. [CrossRef] [PubMed]
23. Zhang, J.; Shan, S.G.; Kan, M.; Chen, X.L. Coarse-to-Fine Auto-Encoder Networks (CFAN) for Real-Time Face Alignment. In
Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; pp. 1–16.
24. Cozler, Y.L.; Allain, C.; Xavier, C.; Depuille, L.; Faverdin, P. Volume and surface area of Holstein dairy cows calculated from
complete 3D shapes acquired using a high-precision scanning system: Interest for body weight estimation. Comput. Electron.
Agric. 2019, 165, 104977. [CrossRef]
25. Wang, K.; Zhu, D.; Gu, O.H.; Ma, Q.; Su, W.; Su, Y. Automated calculation of heart girth measurement in pigs using body surface
point clouds. Comput. Electron. Agric. 2019, 156, 565–573. [CrossRef]
26. Weisheng, D.; Lei, Z.; Rastislav, L.; Guangming, S. Sparse representation based image interpolation with nonlocal autoregressive
modeling. IEEE Trans. Image Process. 2013, 22, 1382–1394.
27. Hornácek, M.; Rhemann, C.; Gelautz, M.; Rother, C. Depth Super Resolution by Rigid Body Self-Similarity in 3D. In Proceedings
of the 2013 IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 1123–1130.
[CrossRef]
28. Smoli, A.; Ohm, J.R. Robust Global Motion Estimation Using A Simplified M-Estimator Approach. In Proceedings of the 2000
International Conference on Image Processing, Vancouver, BC, Canada, 10–13 September 2000; pp. 868–871.
29. Zhu, R.; Yu, S.J.; Xu, X.Y.; Yu, L. Dynamic Guidance for Depth Map Restoration. In Proceedings of the 2019 IEEE 21st International
Workshop on Multimedia Signal Processing (MMSP), Kuala Lumpur, Malaysia, 27–29 September 2019; pp. 1–6. [CrossRef]
30. Ojala, T.; Pietikainen, M.; Maenpaa, T. Multiresolution Gray-Scale and Rotation Invariant Texture Classification with Local Binary
Patterns. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 971–987. [CrossRef]
31. Said, P.; Domenec, P.; Miguel, A.G. Analysis of focus measure operators for shape-from-focus. Pattern Recognit. 2013, 46,
1415–1432.
32. Misra, D. Mish: A Self Regularized Non-Monotonic Neural Activation Function. In Proceedings of the British Machine Vision
Conference, Cardiff, UK, 9–12 September 2019.
33. Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Zuo, W.; Hu, Q. ECA-Net: Efficient Channel Attention for Deep Convolutional Neural Networks.
In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 14–19
June 2020; pp. 11531–11539.
34. Ghiasi, G.; Lin, T.Y.; Le, Q.V. Dropblock: A regularization method for convolutional networks. arXiv 2018, arXiv:1810.12890.
35. Shibata, E.; Takao, H.; Amemiya, S.; Ohtomo, K. 3D-Printed Visceral Aneurysm Models Based on CT Data for Simulations of
Endovascular Embolization: Evaluation of Size and Shape Accuracy. Am. J. Roentgenol. 2017, 209, 243–247. [CrossRef]
36. Carass, A.; Roy, S.; Gherman, A.; Reinhold, J.C.; Jesson, A.; Arbel, T.; Maier, O.; Handels, H.; Ghafoorian, M.; Platel, B.; et al.
Evaluating White Matter Lesion Segmentations with Refined Sørensen-Dice Analysis. Sci. Rep. 2020, 10, 8242. [CrossRef]
37. Abdel Aziz, T.; Allan, H. An Efficient Algorithm for Calculating the Exact Hausdorff Distance. IEEE Trans. Pattern Anal. Mach.
Intell. 2015, 37, 2153–2163.
38. Karimi, D.; Salcudean, S.E. Reducing the Hausdorff Distance in Medical Image Segmentation with Convolutional Neural
Networks. IEEE Trans. Med. Imaging 2020, 39, 499–513. [CrossRef] [PubMed]
39. Newell, A.; Yang, K.; Deng, J. Stacked Hourglass Networks for Human Pose Estimation. In Proceedings of the European
Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016; pp. 483–499.
40. Chu, X.; Yang, W.; Ouyang, W.; Ma, C.; Yuille, A.L.; Wang, X.G. Multi-Context Attention for Human Pose Estimation. In
Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017;
pp. 5669–5678.
41. Cao, Z.; Simon, T.; Wei, S.E.; Sheikh, Y. Realtime Multi-person 2D Pose Estimation Using Part Affinity Fields. In Proceedings of
the 2017 IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1302–1310.
Electronics 2022, 11, 993 22 of 22
42. Sun, Y.; Wang, X.G.; Tang, X.O. Deep Convolutional Network Cascade for Facial Point Detection. In Proceedings of the 2013 IEEE
Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 3476–3483.
43. Ranjan, R.; Patel, V.M.; Chellappa, R. HyperFace: A Deep Multi-task Learning Framework for Face Detection, Landmark
Localization, Pose Estimation, and Gender Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 41, 121–135. [CrossRef]
[PubMed]
44. Huang, G.; Liu, Z.; Laurens, V.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of the 2017 IEEE
Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2261–2269.
45. Hua, G.; Li, L.; Liu, S. Multipath affinage stacked-hourglass networks for human pose estimation. Front. Comput. Sci. 2020, 14,
155–165. [CrossRef]
46. Bao, W.; Yang, Y.; Liang, D.; Zhu, M. Multi-Residual Module Stacked Hourglass Networks for Human Pose Estimation. J. Beijing
Inst. Technol. 2020, 29, 110–119.
47. Donner, A.; Koval, J.J. The estimation of intraclass correlation in the analysis of family data. Biometrics 1980, 36, 19–25. [CrossRef]
48. Johannes, K.; Michael, F.C.; Dani, L.; Matt, U. Joint bilateral upsampling. ACM Trans. Graph. 2007, 26, 96.1–96.4.
49. Camplani, M.; Mantecón, T.; Salgado, L. Depth-Color Fusion Strategy for 3-D Scene Modeling With Kinect. IEEE Trans. Cybern.
2013, 43, 1560–1571. [CrossRef]
50. Yang, J.; Ye, X.; Li, K.; Hou, C.; Wang, Y. Color-guided depth recovery from RGB-D data using an adaptive autoregressive model.
IEEE Trans. Image Process. 2014, 23, 3443–3458. [CrossRef]
51. Enwei, S.; Ramakrishna, V.; Kanade, T.; Sheikh, Y. Convolutional pose machines. In Proceedings of the IEEE Computer Society,
Las Vegas, NV, USA, 27–30 June 2016; pp. 4724–4732.
52. Cozler, Y.L.; Allain, C.; Caillot, A.; Delouard, J.M.; Delattre, L.; Luginbuhl, T.; Faverdin, P. High-precision scanning system
for complete 3D cow body shape imaging and analysis of morphological traits. Comput. Electron. Agric. 2019, 157, 447–453.
[CrossRef]
53. Fischer, A.; Luginbühl, T.; Delattre, L.; Delouard, J.M.; Faverdin, P. Rear shape in 3 dimensions summarized by principal
component analysis is a good predictor of body condition score in Holstein dairy cows. J. Dairy Sci. 2015, 98, 4465–4476.
[CrossRef] [PubMed]