Professional Documents
Culture Documents
III. METHODOLOGY
II. LITERATURE SURVEY
Birdvocalization
Rahul P measures The complete
Audio based
Tivareka introduce identification process
bird species
r, automated involves collection of
recognition
methods deals recorded vocalizations
using naive
Vinayak with powerful of different species
bayes
D. audio signal which is lengthy and in
algorithm
Chavan processing efficient
techniques.
Accuracy is
close to the
Bird species state-of-the-art A common problem
identification and has an when training neural
John
using advantage over networks is often the
Martinss
convolutional raw spectral data lack of available training
on
neural when data. data augmentation
networks computational techniques.
resources are
limited.
2
978-1-7281-4141-1/$31.00 ©2020 IEEE
2020 International Conference on Emerging Trends in Information Technology and Engineering (ic-ETITE)
3
978-1-7281-4141-1/$31.00 ©2020 IEEE
2020 International Conference on Emerging Trends in Information Technology and Engineering (ic-ETITE)
Figure 3. Shown the process by which the bird is Amazon Mechanical Turk is focused on the idea that
detected from the image. It consists of following steps people can do far more work than machines, such as
recognizing objects in a picture or video, de-
• First upload the image and then consider the duplicating data, transcribing audio recordings or
various alignments such as head, body, color, researching data details [10]. Tasks like this have
beak and whole image from that file. historically been done by recruiting and undoing a
• Each alignment is given to extract functions broad temporary workforce (which is time consuming,
from multiple layers of the network via a expensive and difficult to scale).
deep network of conventions.
• The image will be taken into account after its
representation.
• Finally, the result of the classification (i.e. the
characteristics are aggregated and transferred
to the classifier) will be generated and the
bird species will be found
4
978-1-7281-4141-1/$31.00 ©2020 IEEE
2020 International Conference on Emerging Trends in Information Technology and Engineering (ic-ETITE)
The average accuracy of all 7 classes is 92.90% which Conversions, Thus each GPU executes all of the CNN
is shown in the bar map. (Specific Bird classification operations on a separate data set. For each duplicate,
accuracy) measured gradients are averaged before weight changes occur
After analyzing the score sheet, it is observed that the which is the graph's synchronization stage.
accuracy obtained is less when using a single
parameter. But if a mixed (figure 4) approach is used V. CONCLUSION
which Increase accuracy by observing parameters such
This research examined the use of deep learning
as pose, wings, color, beak, head, etc.
algorithms to address the problem of identification of
We test our experiments in a GPU cluster with 2
bird species. The framework used Deep learning
NVIDIA GeForce GTX 680 dual GPU chips, 2 Intel
algorithm (Unsupervised Learning) for image
Xeon E5-2630 8-core processors and 128 GB of RAM
classification on the dataset (Caltech-UCSD Birds
fitted with each node. A 6 GB / s InfiniBand network
200). It consists of 200 categories or 11,788 photos.
performs inter-node communication. TensorFlow 2,
The CUB-200-2011 dataset consisted of mixed
operating on CUDA 7.5 and using cuDNN 5.1.3
parameters such as pose, wings, color, beak, arms, etc.
primitives for improved performance implements on
The initial study of the 200 images of Caltech-UCSD
the DCNN architectures with their training [2].
Birds for these species could then be carried out using
the findings as test images. The experiment was able to
achieve 90.93 percent by fine-tuning the images in
Tensorflow and trained on the GoogleNet dataset and
on the Caltech-UCSD Birds 200 dataset, which
included 11,788 images, the classification average
accuracy. Future research will concentrate on
developing Architecture for Real Time Smartphone.
REFERENCES
5
978-1-7281-4141-1/$31.00 ©2020 IEEE
2020 International Conference on Emerging Trends in Information Technology and Engineering (ic-ETITE)
[7] Wah, C., Van Horn, G., Branson, S., Maji, S., Perona, P.,
Belongie, S., “Similarity Comparisons for Interactive Fine-Grained
Categorization”, IEEE Conference on Computer Vision and Pattern
Recognition (CVPR). 2014.
[11] http://www.vision.caltech.edu/visipedi a
/CUB-200.html
6
978-1-7281-4141-1/$31.00 ©2020 IEEE