Professional Documents
Culture Documents
A PROJECT REPORT
(19EC74C)
Submitted by
BANUPRIYA.R (1911005)
SUVALAKSHMI.S (1911019)
UMA RAJA SELVI.M (1911021)
of
BACHELOR OF ENGINEERING
in
NOVEMBER 2022
BONAFIDE CERTIFICATE
SIGNATURE SIGNATURE
Dr.V.R.S.MANI, Dr.A.SHENBAGAVALLI,M.E.,Ph.D.,
ii
ACKNOWLEDGEMENT
iii
TABLE OF CONTENTS
viii
ABSTRACT
v
LIST OF FIGURES
vi
LIST OF TABLES
vii
LIST OF SYMBOLS, ABBREVATIONS
AND NOMENCLATURE
1 INTRODUCTION 1
2 LITERATURE SURVEY 2
3 PROPOSED WORK 9
3.1 PROBLEM STATEMENT 9
3.2 OBJECTIVE 11
3.3 METHODOLOGY 13
24
4 RESULTS AND OUTPUT
26
5 CONCLUSION
27
REFERENCES
iv
LIST OF FIGURES
v
LIST OF TABLES
vi
LIST OF SYMBOLS, ABBREVIATIONS AND NOMENCLATURE:
vii
ABSTRACT
viii
CHAPTER-1
INTRODUCTION
Oceans are like the lifeblood of Mother Nature, holding 97% of the
earth's water. They produce more than half of the oxygen and absorb most of
the carbon from our environment. Deep learning, the state of art machine
many underwater objects. To make sense of texts, images, sounds etc., deep
learning transforms input data through more layers than shallow learning
image) and create models to learn these representations from large-scale data.
A survey on the current deep learning approaches for various marine object
1
CHAPTER-2
LITERATURE SURVEY
2
2. TITLE OF THE PAPER: Human-Visual-System-Inspired Underwater
Image Quality Measures
Underwater images suffer from blurring effects, low contrast, and grayed
out colors due to the absorption and scattering effects under the water.Many
image enhancement algorithms for improving the visual quality of underwater
images have been developed. Unfortunately, no well-accepted objective
measure exists that can evaluate the quality of underwater images similar to
human perception. The UIQM comprises three underwater image attribute
measures: the underwater image colorfulness measure (UICM), the underwater
image sharpness measure (UISM), and the underwater image contrast measure
(UIConM).These measures are also used on the AirAsia 8501 wreckage images
to show their importance in practical applications.
3
3.TITLE OF THE PAPER: Investigation of Vision-based Underwater
Object Detection with Multiple Datasets
AUTHOR: Dario Lodi Rizzini, Fabjan Kallasi, Fabio Oleari and Stefano
Caselli
4
4. TITLE OF THE PAPER: Comprehensive Survey on Underwater
Object Detection and Tracking
AUTHOR: Girish Gaude1, Samarth Borkar
PUBLISHED YEAR&JOURNAL: 30/Nov/2018 & International Journal
of Computer Sciences and Engineering
5
5. TITLE OF THE PAPER: Underwater Object Detection and
Reconstruction Based on Active Single-Pixel Imaging and Super-
Resolution Convolutional Neural Network
Various approaches have been developed for imaging objects under low-
light conditions and for improving the quality of underwater images.
Polarimetric imaging systems were used to reduce the backscattering effect and
produced a good quality image. Although polarization correction can improve
the renovation quality, it also blocks part of the light incident on the detector,
thereby reducing the signal to noise ratio and complicating the problem of
imaging underwater. This system can overcome the limitations of obtaining
good-quality images in low-light conditions. The proposed SPI system
employed a combination of the CS algorithm and a neural network to enhance
the quality of the images to reduce sampling time and postprocessing of data.
Referring to the network structure of an CS-SRCNN, the employed network
algorithm successfully deduced the HR images from being trained with
multiple measurement data.
6
6. TITLE OF THE PAPER: Underwater Image Processing and Object
Detection Based on Deep CNN Method
AUTHOR: Fenglei Han,Jingzheng Yao , ,Haitao Zhu,and Chunhui Wang
PUBLISHED YEAR & JOURNAL: 22 May 2020
7
7. TITLE OF THE PAPER: Hyperspectral imaging for underwater object
detection
Various approaches have been developed for imaging objects under low-
light conditions and for improving the quality of underwater images.
Polarimetric imaging systems were used to reduce the backscattering effect and
produced a good quality image. Although polarization correction can improve
the renovation quality, it also blocks part of the light incident on the detector,
thereby reducing the signal to noise ratio and complicating the problem of
imaging underwater. This system can overcome the limitations of obtaining
good-quality images in low-light conditions. The proposed SPI system
employed a combination of the CS algorithm and a neural network to enhance
the quality of the images to reduce sampling time and postprocessing of data.
Referring to the network structure of an CS-SRCNN, the employed network
algorithm successfully deduced the HR images from being trained with
multiple measurement data.
8
CHAPTER-3
PROPOSED WORK
3.1 PROBLEM STATEMENT:
ANCHOR BOXES
Multiple RoIs may be predicted at each position and are described
relative to reference anchor boxes. The shapes and sizes of these anchor boxes
are carefully chosen to span a range of different scales and aspect ratios. This
allows various types of objects to be detected with the hopes that the bounding
box coordinates need not be adjusted much during the localization task.
9
MULTIPLE FEATURE MAPS
LIMITED DATA
CLASS IMBALANCE
10
3.2 OBJECTIVE:
MISSED OBJECTS
DETECTION OF CROWDEDOBJECTS
The latest addition to the CNN based object detection technique after R-CNN
and Fast R-CNN is named Faster R-CNN. Fast R-CNN is faster than R-CNN
but was not able to achieve real-time detection for video data using deep
neural networks. Both R-CNN and Fast R-CNN are based on the concept of
calculating region proposals which act as the tailback. Finding regions of
interest in both R-CNN and Fast R-CNN depends on selective search method
that uses the greedy algorithm to merge super pixels based on low-level
features, which is a slow operation
Though the initial Faster R-CNN was based on VGG16 convolutional neural
network, the detection accuracy have been improved due to the extensive
experimentation and crafting of the architecture of the CNN’s and by
increasing the width and height of the network. On the series of improvement
for the object detection network, the new network codenamed ‘inception’ has
gained the best and highest accuracy in ILSVRC 2014.
11
Fig 3.1 The naive version of the Inception module
12
3.3 METHODOLOGY
Data collection
camera with the lens was encapsulated inside a compatible underwater housing.
Data labelling
All the images from both real and laboratory environments are then
For training the detector, we have divided the whole ECUHO-1 dataset
into the training set and testing set. ECUHO-1 training set consists of 2,160
13
Fig 3.3 Block diagram of the proposed methodology
14
3.4 Performance Evaluation
Once the training was completed, the inception V2 based Faster R-CNN
object detector was tested and evaluated using COCO detection metrics
Classifier localization loss, (c) RPN localization loss, (d) RPN objectness
15
Fig 3.5 Loss curves of training the Inception V2 based faster R-
localization loss, (c) RPN localization loss, (d) RPN objectness loss, (e)
Total loss, (f) clone loss, and (g) time required per step
16
RESULT AND DISCUSSION:
YOLO and Faster RCNN both share some similarities. They both uses a
anchor box based network structure, both uses bounding both regression.
Things that differs YOLO from Faster RCNN is that it makes classification and
bounding box regression at the same time. it make sense that YOLO wanted a
more elegant way to do regression and classification.
Faster RCNN on the other hand, do detect small objects well since it has
nine anchors in a single grid, however it fails to do real-time detection with its
two step architecture.
18
3.5 YOLO METHODOLOGY
A. Network architecture:
There are prior three versions of YOLO before YOLOv3. YOLOv1 was
the first implementation of single stage detector concept which uses reduction
layers of dimension 1x1 followed by convolutional layer of dimension 3x3 and
uses batch normalization and leaky ReLU activation function. The network
consists of 24 convolutional layer which extracts features and the two fully
connected (FC) layer that predicts bounding boxes and its class probability. The
final output obtained is a 7x7x30 tensor consisting of bounding boxes. This
model is trained to detect 49 objects, but produces high value of error in
localizing them.
19
Bounding box
Both the previous YOLO models can detect less than 20 classes, hence a more
advanced model YOLO9000 was developed which can detect and classify more
objects and classes. These models were then improved by adding more features
like residual blocks, skip connections and up-sampling and named as YOLOv3
which utilizes a 53 layered network which is trained on ImageNet database.
YOLOv3 uses a single pipeline for feature extraction and hence the
whole image is passed on to the convolutional network and produces a square
output called grid on to which the bounding boxes are anchored. The grid cell
and anchor share a common centroid. The YOLO algorithm predicts location
offset against anchor box: tx, ty, tw, th, objectness scores, and class probability.
Objectness-score gives the confidence of object presence in the bounding box
and the and class probability defines the class which it belongs perfectly or not.
The predictions correspond to the bounding box coordinates with (Cx, Cy)
being the upper-left corner and Ph and Pw being the width and height and as
depicted in the Figure and calculated as given in (1), (2), (3), (4)
C. Class prediction:
Predictions are made by three different scales; 13x13, 26x26 and 52x52.
Features are extracted using feature pyramid network followed by darknet53.
The last stage of prediction is a 3-d tensor encoding the bounding box,
confidence value that gives objectness score, and probability of the object being
in a particular class.
21
Fig 3.10 Darknet53 Architecture
Fig 3.11 Sample F4K dataset image showing single and multi-objects
22
Fig 3.12 Object Detection Methodology and bounding box with
objectness score
E. Feature extractor
23
CHAPTER-4
{
"predictions":
[
{
"x": 486,
"y": 377.5,
"width": 52,
"height": 69,
"class": "echinus",
"confidence": 0.854
},
{
24
"x": 300,
"y": 478,
"width": 32,
"height": 62,
"class": "echinus",
"confidence": 0.842
},
{
"x": 393.5,
"y": 128,
"width": 27,
"height": 48,
"class": "echinus",
"confidence": 0.777
},
{
"x": 380,
"y": 490.5,
"width": 52,
"height": 43,
"class": "echinus",
"confidence": 0.686
},
{
"x": 413,
"y": 92,
"width": 16,
"height": 30,
"class": "echinus",
"confidence": 0.598
}
]
}
25
CHAPTER-5
CONCLUSION
26
REFERENCES:
[1] Beijbom, O., Edmunds, P. J., Kline, D.I., Mitchell, B.G., Kriegman, D.:
Automated annotation of Coral Reef Survey Images. In: IEEE Computer
Society Conference on Computer Vision and Pattern Recognition, pp. 1170–
1177. IEEE Press, Rhode Island (2012)
[2] Campos, M.M., Codina, G.O., Amengual, L.R., Julia, M.M.: Texture
Analysis of Seabed
Images: Quantifying the Presence of Posidonia Oceanica at Palma Bay. In:
OCEANS 2013
- MTS/IEEE Bergen, pp. 1-6. IEEE Press, Bergen (2013)
[3] Erftemeijer, P.L.A., Lewis III, R.R.R.: Environmental Impacts of
Dredging on Seagrasses: A review. Marine Pollution Bulletin 52, 1553–1572
(2006)
[4] Hedley, J.D., Roelfsema, C.M., Chollett, I., Harborne, A.R., Heron,
S.F., Weeks, S.J., Mumby, P.J.: Remote Sensing of Coral Reefs for Monitoring
and Management : A Review. Remote Sens. 8(118), (2016)
[5] Hinton, G.E., Osindero, S.: A fast Learning Algorithm for Deep Belief
Nets. Neural Comput. 18(7), 1527-1554 (2006)
[6] Krizhevsky, A., Sutskever, I., Hinton, G.E.: ImageNet Classification with
Deep Convolutional Neural Networks. In: Advances in Neural Information
Processing Systems 25 (NIPS) (2012)
[7] Lavery, P.S., McMahon, K., Weyers, J., Boyce, M.C., Oldham, C.E.:
Release of Dissolved Organic Carbon from Seagrass Wrack and its
Implications for Trophic Connectivity. Marine Ecology Progress Series 494,
121–133 (2013)
[8] Li, X., Shang, M., Qin, H., Chen, L.: Fast Accurate Fish Detection and
Recognition of Underwater Images with Fast R-CNN. In: OCEANS 2015 -
MTS/IEEE Washington, pp. 1-5.
27
IEEE Press, Washington, DC (2015)
[9] Lu, H., Li, Y., Zhang, Y., Chen, M., Serikawa, S., Kim, H.: Underwater
Optical Image Processing: A Comprehensive Review. Computer Vision and
Pattern Recognition. (2017)
[10] Mahmood, A., Bennamoun, M., An, S., Sohel, F., Boussaid, F., Hovey,
R., Kendrick, G., Fisher, R.B.: Coral Classification with Hybrid Feature
Representations. In: IEEE Inter1national Conference on Image Processing, pp.
519-523. IEEE Press, Arizona (2016)
[11] Oguslu, E., Erkanli, S., Victoria, J., Bissett, W.P., Zimmerman, R.C.,
Li, J.: Detection of Seagrass Scars using Sparse Coding and Morphological
Filter. In: Proc. SPIE 9240, Remote Sensing of the ocean, Sea Ice, Coastal
Waters, and Large Water Regions, pp. 92400G, Amsterdam (2014)
[12] Qin, H., Li, X., Yang, Z., Shang, M.: When Underwater Imagery
Analysis Meets Deep Learning : A Solution at the Age of Big Visual Data in
OCEANS 2015 – MTS/IEEE Washington, pp. 1-5. IEEE Press, Washington,
DC (2015)
[13] Ravanbakhsh, M., Shortis, M., Shafait, F., Mian, A., Harvey, E., Seager,
J.: Automated Fish Detection in Underwater Images UsingShape‐ Based
Level Sets. The Photogrammetric Record 30(149), 46-62 (2015)
[14] Shiela, M.M.A., Soriano, M., Saloma, C.: Classification of Coral Reef
Images from Underwater Video using Neural Networks. Optics express 13(22),
8766–8771 (2005)
[15] Spampinato, C., Chen-Burger, Y.H., Nadarajan, G., Fisher, R.B.:
Detecting, Tracking and Counting Fish in Low Quality Unconstrained
Underwater Videos. In: 3rd International Conference on Computer Vision
Theory and Applications (VISAPP), pp. 514-519. Funchal, Madeira (2008)
[16] Stokes, M.D., Deane, G.B.: Automated Processing of Coral Reef Benthic
Images. Limnol. Oceanogr.: Methods. 7, 157-168 (2009)
28
[17] Teng, M.Y., Mehrubeoglu, M., King, S.A., Cammarata, K., Simons, J.:
Investigation of Epifauna Coverage on Seagrass Blades Using Spatial and
Spectral Analysis of Hyperspectral images. In: 5th Workshop on Hyperspectral
Image and Signal Processing: Evolution in Remote Sensing, pp. 25-28.
Gainsville, Florida (2013)
[18] Vanderklift, M., Bearham, D., Haywood, M., Lozano-Montes, H.,
Mccallum, R., Mclaughlin, J., Lavery, P.: Natural Dynamics: Understanding
Natural Dynamics of Seagrasses in North-Western Australia. Theme 5 Final
Report of the Western Australian Marine Science Institution (WAMSI)
Dredging Science Node. 39 (2016)
[19] Deng, L., Yu, d.: Methods and Applications. Foundations and Trends.
Signal Processing. 7, 197-387 (2013)
[20] Schmidhuber, J.: Deep Learning in Neural Networks: An Overview.
Neural Networks. 61, 85-117 (2015)
[21] Song, H. A., Lee, S.Y.: Hierarchical Representation Using NMF. In:
ICONIP 2013. LNCS, vol. 8226, pp. 466-473. Springer, Heidelberg (2013)
[22] Liu, M.: AU-aware Deep Networks for facial expression recognition. In:
10th IEEE International Conference and Workshops on Automatic Face and
Gesture Recognition, pp. 1-6, IEEE Press, Shanghai (2013)
[23] Ciresan, D., Meier, U., Schmidhuber, J: Multi-column deep neural
networks for image classification. Computer Vision and Pattern Recognition.
3642-3649 (2012)
[24] Elawady, M.,E.: Sparsem: Coral Classification Using Deep
Convolutional Neural Networks. Msc Thesis, Hariot-Watt University (2014)
[25] Villon, S., Chaumont, M., Subsol, G., Villeger, S., Claverie, T.: Coral
Reef Fish Detection and Recognition in Underwater Videos by Supervised
Machine Learning: Comparison between Deep Learning and HOG+SVM
Methods. In: 17 International Conference on Advanced Concepts for Intelligent
29
Vision Systems, Springer (2016)
[26] Py, Q., Hong, H., Zhongzhi, S.: Plankton Classification with Deep
Convolutional Neural Network. In: IEEE Information Technology,
Networking, Electronic and Automation Control Conference, pp. 132-136.
Chongqin (2016)
[27] Lee, H., Park, M., Kim, J.: Plankton classification on imbalanced large
scale database via convolutional neural networks with transfer learning. In:
IEEE International Conference on Image Processing (ICIP), pp. 3713-3717.
Phoenix (2016)
[28] Dai, J., Wang, R., Zheng, H., Ji, G., Qiao, X.: ZooplanktoNet: Deep
Convolutional Network for Zooplankton Classification. In: OCEANS, pp. 1-6.
Shanghai (2016)
[29] Hemminga, M.A., Duarte, C.M.: Seagrass Ecology. Cambridge
University Press, Cambridge (2004)
[30] National Data Science Bowl, https://www.kaggle.com/c/datasciencebowl
[31] Dieleman, S.: Classifying Planktons with Deep Neural Networks,
http://ben-
anne.github.io/2015/03/17/plankton.html
30