You are on page 1of 8

Algae Detection Using Computer Vision and Deep Learning

Arabinda Samantaray1 , Baijian Yang2 , J. Eric Dietz2 , and Byung-Cheol Min1

Abstract—A disconcerting ramification of water pollution


caused by burgeoning populations, rapid industrialization and
modernization of agriculture, has been the exponential increase
in the incidence of algal growth across the globe. Harmful algal
blooms (HABs) have devastated fisheries, contaminated drinking
water and killed livestock, resulting in economic losses to the tune
of millions of dollars. Therefore, it is important to constantly
monitor water bodies and identify any algae build-up so that
arXiv:1811.10847v1 [cs.CV] 27 Nov 2018

prompt action against its accumulation can be taken and the


harmful consequences can be avoided. In this paper, we propose
a computer vision system based on deep learning for algae
monitoring. The proposed system is fast, accurate and cheap,
and it can be installed on any robotic platforms such as USVs
and UAVs for autonomous algae monitoring. The experimental
results demonstrate that the proposed system can detect algae
Fig. 1. Concept of a real-time algae monitoring using a mobile platform. Our
in distinct environments regardless of the underlying hardware proposed computer system can be installed on any robotic platforms such as
with high accuracy and in real time. USVs and UAVs.
Index Terms—Deep Learning, Computer Vision, Robotic Water
Monitoring, Algae

I. INTRODUCTION communities that depend on these water bodies for potable


water, recreation and a flourishing fishing and tourism industry
Algae are primarily aquatic, uni or multi-cellular organisms
[8].
that contain chlorophyll [1]. In a healthy aquatic environment,
Hence, it is extremely important that ponds, lakes and rivers
algae play the role of primary producers and are critical
are constantly monitored and prompt action is taken against
in preserving the food chain. Algae also benefit humans by
any abnormal algae buildup. However, current methodologies
reducing the level of greenhouse gases in the atmosphere by
for algae monitoring require significant manpower, making
fixing large quantities of CO2 in oceans [2], serving as a
them economically unfeasible (e.g., it was projected that a
source of energy in the form of bio fuels [3] and acting as
budget of $3.5 million was required to monitor the 100 largest
a cheap but highly effective means for waste-water treatment
lakes in Oklahoma once a month [9]). Alternatively, methods
[4].
may depend upon the hardware available, which can be in the
Under congenial ecological conditions, however, the rate of
form of drones [10], robotic fish [11], underwater cameras [12]
algae proliferation can increase exponentially, and large algal
or satellites [13], which limits their applicability in varying
colonies, sometimes covering an area of many square kilo-
environmental, topological and socio-economic conditions.
meters, can be formed. Such colonies are known as harmful
algal blooms (HABs). They have been found to be responsible The primary contributions of our research include the fol-
for releasing paralytic, neurotoxic, diarrhetic, amnesic and lowing:
azaspiracid toxins causing shellfish poisoning [5] in bodies • Development of a computer vision algorithm based on
of water, leading to the deaths of fishes, sea mammals, birds deep learning that can detect and locate algae accurately
and even humans [6][7]. in water bodies, irrespective of the variations in camera
In addition to releasing toxins, prolonged periods of algal parameters, environmental conditions, orientation of the
cover on water surfaces reduces sunlight penetration, thus captured image or the presence of significant background
preventing underwater aquatic flora from conducting photo- clutter in the form of trees, vegetation, buildings etc.
synthesis and negatively affecting the aquatic ecosystem. Sub- • Our proposed system can be installed on mobile platforms
sequently, when HABs decay, they consume large quantities of including unmanned aerial vehicles (UAVs), unmanned
oxygen, causing anoxia and death of marine mammals, fishes surface vehicles (USVs), and airplanes. when they are
and crustaceans among others [5]. used for algae monitoring, because the system can detect
The destruction of aquatic environments by HABs has a algae in real time.
significant negative impact on quality of life for many local • The proposed computer vision system is cheap as it does
1
not require any specialized hardware.
A. Samantaray and B.-C. Min are with SMART Lab, Department of
Computer and Information Technology, Purdue University, West Lafayette, This paper is organized as follows. In Section II, we describe
IN 47907, USA. e-mails: samantar@purdue.edu,minb@purdue.edu the current algae monitoring systems and their shortcomings.
2 B. Yang and J. E. Dietz are with Department of Computer and Information
Technology, Purdue University, West Lafayette, IN 47907, USA. e-mails: We introduce the proposed algae detection system using com-
byang@purdue.edu,jedietz@purdue.edu puter vision and deep learning in Section III. In Section IV,
we present the evaluation procedure and the results obtained,
followed by conclusions and future works in Section V.

II. RELATED WORK


Eutrophication by agricultural runoff [14], dumping of un-
treated industrial and household effluents [15] and transporta-
tion of foreign algae species in ballast water [16][17], coupled
with natural phenomena such as storms, tsunamis, currents
[18] and global warming [19], have been cited as the root
(a) Ground view (b) Aerial view
cause for the rising incidence of HABs in water bodies across
the globe. This rising incidence has increased the relevance of Fig. 2. Different images of algae filled pond taken from a ground view and
effective algae monitoring methodologies in the modern world. an aerial view.
Algae monitoring techniques can be roughly classified into 3
categories: in-situ sampling, computer vision-based techniques have on their algae monitoring system, considering that their
and hyperspectral remote sensing using satellite or aircraft. detector is only trained on iconic images of algae, grass and
A. In-situ Sampling water.
Even in paper [22], the authors only considered images that
In-situ sampling is done by performing on-site sampling and were taken from the ground and did not describe the speed
transporting those samples to laboratories for further evalua- with which their vision system could detect algae. Hence, it is
tion. Although onsite sampling is done at regular intervals, difficult to ascertain whether this method could be used from
this methodology is extremely time and labor intensive. Also, mobile platforms such as UAVs, USVs and airplanes, which
the possibility of contaminated samples negatively affecting operate in different environments.
observations is high [20].
C. Satellite Remote Sensing
B. Computer Vision based Techniques
The use of satellite-based remote sensing for algae monitor-
The distinctive green or greenish-blue color, characteristic ing makes use of the increased diffused reflectance caused by
of algae has been used to develop computer vision-based the presence of algal pigment in a body of water [13]. Different
algae monitoring systems. However, such traditional com- categories of algorithms such as reflectance classification
puter vision pipelines do not have high repeatability because algorithms, reflectance band-ratio algorithms and spectral band
they are significantly dependent on the effectiveness of the difference algorithms take the spectral data as input to detect
feature detectors or the segmentation procedure, which can the presence of algae in bodies of water [23].
be rendered ineffective by various environmental conditions However, although these algorithms have been successful
such as fluctuating illumination, occlusion or the presence of in monitoring algal blooms in the open ocean, they have been
comparable objects in the background [21]. unproductive in coastal water and in bodies of water with
The use of color-based segmentation, species specific fea- significant human activity because the reflectance spectrum
tures and an underwater camera in study [12] implies that becomes distorted in the presence of organic material and
the algae monitoring system cannot be applied on a different suspended particles. Also, issues such as unavailability of real-
hardware platform such as an UAV, USV, or smartphone and is time data, irregular site revisit times, low resolution of publicly
susceptible to occlusion and variations in illumination, which available satellite products such as LANDSAT or MODIS
are regular occurrences in an outdoor environment. Similarly, (> 30m) and exorbitant costs of proprietary systems such as
the novel approach of combining the use of a smartphone QuickBird [24] make the use of satellite imagery for a general
camera and the inertial sensors present on a robotic fish to purpose algae monitoring system difficult.
detect the shoreline and hence perform image segmentation to Hence, an economically feasible algae monitoring system
detect algal blooms as presented in the paper [11] is optimized that is robust to changes in image parameters (e.g., image
for their specific hardware platform (i.e., robotic fish) and is size, resolution, orientation), can be easily calibrated depend-
susceptible to variable illumination, occlusion and shadows ing on the environmental conditions and algal species under
cast by surrounding vegetation. consideration and is able to work from a wide variety of
Similar limitations can also be observed in the work of platforms (e.g., UAVs, USVs, airplane, and even smartphones)
[10], which made use of a local binary pattern (LBP) texture would significantly facilitate administrators and civilians in
detector to detect algae. However, in their work, they have only monitoring bodies of water for algal blooms. The development
indicated the success of their methodology by using a UAV of such a computer vision system is the major contribution of
platform and have not described their system’s performance in this paper.
case the images were taken from different orientation or higher
elevation. Also,they have not presented any results describing III. M ETHODOLOGY
the impact that the presence of comparable objects such as Current computer vision-based algae monitoring techniques
trees, plants and seaweeds in the image’s background would can be classified into three categories. The first comprises
Algorithm (Transfer learning
Dataset Development Labelling Images + Models + Training) Results
1. Generating Region
Proposals

Labeled
Raw Data Data Prediction Ground view
l
Algae? Yes

Algae? No

3. Classify
Regions
Bounding box 2. Compute CNN Features
Aerial view

Fig. 3. The procedure undertaken to develop the proposed computer vision system for algae detection.

methods that use UAVs, satellites or airplanes for algae Subsequently, the images and their corresponding annota-
monitoring. The second class of monitoring techniques makes tions should be randomly assigned to training, validation and
use of water based robots, USVs and ships. A third approach testing datasets. Typically, we can assign 70% of images to
that has been proposed is to build a community-based al- the training set, 20% of images to the validation set and 10%
gae monitoring system composed of citizen scientists using to the test set.
ground-based cameras such as smartphones [22]. However,
all these techniques capture images from different orientations B. Object Detection Algorithm
and use computer vision pipelines that exploit the advantages Initially, the general lack of algae images on the Internet
provided by their respective hardware platforms, making it made it inconceivable to apply deep learning algorithms due
highly improbable that the vision system developed for aquatic to the unavailability of a large amounts of annotated data,akin
platforms could work with images taken from the aerial to conventional datasets such as COCO [25] or PASCAL-VOC
view or vice-versa because the images would be significantly [26]. However, a new paradigm of machine learning known
different from each other as shown in Figure 2. as Transfer Learning enabled us to overcame the limitations
However, using only one kind of platform severely restricts posed by the size of our dataset.
our ability to monitor bodies of water of different sizes 1) Transfer Learning: Transfer learning is a new paradigm
effectively. A ground-based algae monitoring system would in machine learning that enables us to use different domains,
be incapable of detecting the growth of algal blooms at the distributions, and tasks in training and testing datasets [27].
center of large lake, while aerial or aquatic robot platform This implies that, Convolutional Neural Network (CNN),
based methods would restrict algae monitoring capabilities to which has been trained on a large labeled dataset from a
only communities that possess that particular hardware. Hence, different domain can be applied for feature detection in a
a computer vision system that enables algae monitoring to completely new domain. This is made possible by the fact
be executed by different platforms, as depicted in Figure 1, that the lower layers of the network are capable of detecting
would democratize the process of algae monitoring and make general features such as edges, blobs etc., that comprise an
it available to communities all across the globe. image, while higher layers capture domain specific features.
The following subsections describe the steps shown in This enables a pre-trained network to be learn from a much
Figure 3, which reflects the steps we have undertaken to smaller dataset, to recognize a completely new class of objects,
develop our computer vision system. as only the latter layers of the network have to be trained [28].
2) Deep Learning Models: An algae monitoring system,
A. Dataset Development that conceivably could be used from mobile platforms such
The first step in the application of machine learning al- as USVs, UAVs, airplanes etc., has to detect and locate algae
gorithms is the preparation of a dataset. Because there is at near real time speeds with high accuracy. As Faster R-
no publicly available dataset containing images of algae in CNN, Single Shot Detector (SSD), and Region-based Fully
an outdoor environment, a new benchmark dataset must be Convolutional Networks (R-FCN) models have shown near
developed. Each image in the dataset should be labeled with real-time object detection on conventional datasets such as
annotation software to generate files containing the coordinates COCO and PASCAL-VOC with very high accuracy, they are
of the bounding boxes, indicating the location of algae in the applicable to our envisaged scenario. The working of these 3
image. To ensure that the complexity of the images in the models is described in the following subsections.
dataset is equivalent to that of the outdoor environment, it is • Faster R-CNN: The Faster R-CNN [29] was developed to
recommended that a significant number of background objects remove the limitations of R-CNN [30] and Fast R-CNN
such as trees, grass, plants, roads and buildings be present in [31] models. It makes use of a region proposal network
the images, rather than just having the region of interest, in to generate box proposals from an intermediate feature
this case algae, as the main part of the image. map. These box proposals are then applied on the same
feature map to extract proposal regions, on which class Algorithm 1 Training a neural network to detect algae
predictions and bounding box improvements are made. Input: A file containing pre-trained weights of the model;
• R-FCN: R-FCN network has a shared convolution archi- Label map; Configuration file for the pre-trained model;
tecture that makes the use of position sensitive score maps Output: A file containing the weights of the newly trained
to compromise between translation invariance for image model
classification and translation variance for object detection 1: repeat
[32]. 2: Prepare an annotated dataset and split it into training,
• Single Shot Detector: The single shot multibox detector validation and testing dataset
was recently developed in [33]. It is a single layer feed 3: Convert the dataset annotations into appropriate input
forward convolution network that makes the use of pre- format
defined anchor boxes, to train the network and make 4: Fine tune the hyperparameters of the neural network
predictions about the class within the anchor box and 5: Monitor the training loss and mean average precision
the offset by which the anchor needs to be shifted to fit on validation dataset
the ground-truth box. 6: If mAP graph converges stop training and observe the
final validation mAP
3) Learning from Custom Dataset: The procedures fol-
7: until validation mAP > satisfactory mAP
lowed to enable the chosen deep learning models to detect
8: Obtain the mAP of the trained network on the test dataset
algae are as follows:
9: Deploy the model into production
a) Convert data: The input data, i.e., the images and bound- 10: Set a confidence threshold and visualize the results in the
ing box coordinates, need to be converted into appropriate image
formats depending upon how the data is being ingested
and the deep learning framework being utilized. In case of
Tensorflow [34], the input data should be converted into
tfrecords while for MXNET [35], recordIO file should be validation dataset. Observing the different metric values
used. This ensures faster processing as compared to when such as training loss, mean average precision (mAP) on
the data is read directly from disk. the validation dataset enables us to infer:
b) Data Augmentation: To mitigate the effects of having
• When the model has stopped learning, so that training
a small dataset and to replicate external environmental
can be stopped
conditions such as variable illumination, fluctuating con-
• Whether subsequent iterations of training should be
trast, and blurring, the dataset should be augmented by
conducted, by optimizing hyperparameters or using a
applying transformations such as randomly changing the
larger dataset
brightness, contrast, hue, color, and saturation.
c) Label Maps: Machine learning algorithms cannot work f) Evaluating the trained model on the test dataset and
with categorical features (e.g., class labels for object of deploying into production: The trained model would be
interest in images), and hence categorical features should evaluated on the test set to examine its classification
be related with a numerical value. This is usually done by and detection accuracy on a completely new set of
using a label map. Since our dataset annotations belong data, allowing us to infer whether the chosen model is
to only one class (i.e., algae), in our label map, the algae applicable to be used as an algae monitoring system.
class is related to the ID 1. g) Visualizing results generated by neural network: For each
d) Fine tune hyperparameters in configuration files: Pre- frame/image the neural network populates the following
trained neural networks are made available with 2 com- arrays:
ponents, which are the weights of the model and the • Boxes – this array contains the normalized coordinates
configuration file that determines the meta-architecture of for each predicted bounding box.
the model, feature extractor present in the model, training • Scores – this array contains the confidence scores for

parameters and the evaluation metrics. To customize the each of the predicted boxes.
pre-trained networks for our dataset, a decaying learning • Classes – this array contains the class label for each

rate of 10% every 5000 steps is applied, and we change of the predicted boxes.
the final layer to reflect that there is only one class of • Number of detections – This array contains the value
objects in our dataset. regarding total number of detections made per image.
e) Begin training and monitor evaluation metrics: After cus- From these individual arrays, we create a list of all the
tomizing the configuration files, the training of the neural predicted bounding boxes having a confidence score of
network models can be initialized, which would fine-tune higher than 50%. Each list item contains the class label,
the weights in the latter layers of the neural networks normalized box coordinates and the confidence scores for
enabling them to detect algae. During the training of the each bounding box as shown in Figure 4.
model, after a specified number of steps, an intermediate By applying the following equations for each bounding
trained model would be stored and evaluated on the box, we convert the normalized coordinates into image
TABLE I
D ISTRIBUTION OF IMAGES ACROSS THE TRAINING , VALIDATION , AND
TEST SETS (D.= D ETECTION )

Training Validation Testing (D.)


Ground Images 277 79 41
Aerial Images 150 43 20

(a) Results (b) Resultant Image


testing dataset containing images of water bodies having
algae (52 images) and not having algae (48 images).
Fig. 4. Results generated by the model and visualizing them on the respective • Test set for detection – on this dataset, we had images
image.
containing ground truth boxes around algal bloom patches
and the mAP of each network was calculated on this
coordinates, dataset
The division of the entire dataset into training, validation,
Coordinatek = Boxji · ImageW idth (1) and testing dataset is presented in Table I.
2) Model Training Environment: We utilized Tensorflow
where k ∈ (left,right,top,bottom), i is an index of boxes,
Object Detection API to train our chosen models to detect
j ∈ (0, 1, 2, 3), and ImageW idth is a width of the
algae. Tensorflow Object Detection API is an open source
image. Subsequently, these image coordinates can then
framework based on the tensorflow library, and it provides a
be used to visualize the results of the predicted boxes as
well structured environment for developing, training, testing,
shown in Figure 4.
and deploying deep learning models.
Algorithm 1 summarizes all the steps from (a)–(g), involved 3) Hardware for Evaluation: For evaluating our proposed
in developing deep learning models to detect algae. computer vision system, we use a HP Pavilion laptop, having
a Intel(R) Core(TM) i7-6500U CPU and NVIDIA GeFORCE
IV. E VALUATION AND R ESULTS
940MX GPU.
A. Preparation for Evaluation
B. Results and Analysis
Since our objective was to develop a computer vision system
that can detect and locate algae in water bodies at a fast speed, In this section, we will describe the results obtained by the
we focused on the following three different evaluation metrics: 3 different neural networks on each evaluation criterion:
1) Classification: A reliable algae monitoring system
• Precision and Recall – to evaluate the accuracy of our
should have high Recall( if algae is present, then the system
system in detecting whether a given water body contains
should detect it) and high Precision(if the presence of algae is
algae or not
predicted, then algae should be present in the water body). To
• Mean Average Precision – to evaluate how accurately our
evaluate the competence of our system in detecting algae, we
system can locate an algal bloom in a water body
performed a binary classification between two sets of images,
• Speed – to evaluate the speed at which each neural
one having water bodies containing algae and others having
network detects algae, so as to validate our approach’s
water bodies not containing algae. The trained version of each
appropriateness to be used in mobile platforms such as
model was tested on the aforementioned images and their
USVs, UAVs, airplanes etc.
results can be seen from the Table II (See column 2–4). Since
1) Dataset Development: For purpose of this research, we there were only two class of images, we can observe that SSD
developed a dataset containing images of algae in pools, lakes, performed very poorly and its ability to detect algae was akin
ponds etc., taken from ground and aerial vehicles. We also to that of random selection which implies that it is not at all a
collected some images from aerial vehicles of water bodies suitable model for algae detection. However, while Faster R-
not containing algal blooms. The dataset we developed had 4 CNN and R-FCN had nearly similar Precision values, the high
categories: Recall value of R-FCN indicates that it is highly robust and
• Training dataset – this dataset was used to train each of nearly always detects an algae bloom if it is actually present.
our neural networks 2) Detection: The second objective of our algae monitoring
• Validation dataset – this data was used to validate the per- system was to locate the exact position of algal blooms in an
formance of our training, by evaluating the intermediate image. Accurate algae localization on an image would enable
neural networks. Based upon the results on the validation us to generate precise world coordinates of where the algae
set, we decided whether to continue training, fine-tune blooms are present, by applying Camera Calibration. The
hyper-parameters or stop training. availability of such fine-grained algae monitoring would be a
• Test set for classification – to ensure that our trained neu- great asset to the local administrators in their battle against the
ral networks were only detecting water bodies containing algae menace. To evaluate the algae detection accuracy of the
algae, and not water bodies in general, we developed a 3 trained models we used the mean average precision (mAP)
TABLE II
E VALUATING THE ACCURACY OF CLASSIFICATION (ACCURACY, R ECALL , AND P RECISION ), THE ACCURACY OF DETECTION (VALIDATION M AP AND
T EST M AP), AND THE SPEED OF THE DETECTION (FPS (GPU) AND FPS (CPU)) OF EACH MODEL (VALID . = VALIDATION )

Accuracy Precision Recall Valid. mAP Test mAP FPS (GPU) FPS (CPU)

Faster R-CNN 72% 74% 71.15% 38.68% 23.46% 2.83 0.62


R-FCN 82% 78.33% 90.38% 38.14% 21.44% 2.72 0.58
SSD 50% 52.08% 48.07% 23.72% 17.29% 19.37 5.16

metric. The mAP is obtained by integrating the precision recall However, the detection speeds for each model, particularly
curve [36], when implemented on a GPU, are satisfactory for being used
Z 1
as algal monitoring systems as algal blooms grow in static or
mAP = p(x)dx (1)
0
slow moving water bodies.
where p(x) is the precision-recall curve. To determine the pre- C. Discussion
cision recall curve, the true positive and false positive values
Despite using state-of-art object detection models, we ob-
of the prediction can be computed by using the Intersection
served that there were a few occasions in which either algae
over Union (IoU) criterion,
was not detected or the surrounding vegetation was detected as
Boxpred ∩ Boxgt algae, which could be a topic for further study. Some examples
IOU = (2) are presented in Figure 7. The most likely cause for such
Boxpred ∪ Boxgt
incidences is that, the only defining characteristic of algae is
where Boxpred and Boxgt are the areas included in the its green color which is the most common color found in an
predicted and ground truth bounding box, respectively. Then, outdoor environment. However, a larger dataset enabling the
a threshold for IoU is designated (e.g., 0.5), and if the IoU neural network to learn more intricate features would be able
exceeds the threshold, the detection is marked as correct de- to detect algae with more consistency.
tection. Multiple detections of the same object are considered
as one correct detection, while others are considered as false V. C ONCLUSION AND F UTURE W ORK
detections. Post, obtaining the true positive and false positive In this work, our objective was to develop a computer vision
values, a precision-recall curve is generated, based on which system that can detect and locate algal blooms in real time,
the mAP can be calculated. from cameras placed on aerial, aquatic or ground based plat-
The mAP values obtained by evaluating all the 3 trained forms. In order to attain our objective we compared state-of-art
models on the validation and test dataset are presented in Table object detectors such as Faster R-CNN, R-FCN, and SSD.Our
II (See column 5 and 6). From the table, we can observe final conclusion was that an algae monitoring system based on
that all the 3 models show reasonably acceptable accuracy the R-FCN model would be highly robust, accurate and fast
in algae bloom localization. In particular, it reveals that both to enable effective, real time algae monitoring. We expect that
Faster R-CNN and R-FCN have higher detection accuracy such a system will significantly facilitate researchers, local
than SSD on both the test and validation dataset. Also, we administrators and civilians in monitoring water bodies and
can observe that the mAP on the test data set is lower than promptly curbing any excessive algal growth.
the mAP on the validation dataset. We believe, this resulted Future works will be focused on improving the current
from the fact that we had chosen the most complex images performances by developing a larger dataset and implementing
in our entire dataset with significant amounts of background field tests. In addition, we will use this computer vision system
clutter (i.e., trees, buildings, roads etc.,) to be a part of our to facilitate the working of a multi-robot ecosystem composed
test dataset. Nevertheless, despite the complexity inherent to of autonomous UAVs and USVs which would be responsible
our test dataset we observe from Figure 5 and Figure 6 that for monitoring water bodies and removal of HABs as shown
algal blooms are well detected regardless of orientation and in Figure 1.
location from where the image was taken.
3) Speed of Detection: Since we envision that our algae VI. ACKNOWLEDGEMENTS
monitoring system can be applied to fast moving mobile The author would like to acknowledge the help and support
platforms such as UAV, USV, airplanes etc., it is necessary provided by the SMART Lab members, notably Wonse Jo
that our system can perform algae detection and localization and Dr. Ramviyas Parasuraman in the development of this
in real time. The results regarding the speed with which each computer vision system.
model can detect algae in an input image is shown in Table II
(See the last two columns). In contrast to the classification and R EFERENCES
detection accuracy, SSD outperforms the other two models in [1] “General algae information,” http://www.ecy.wa.gov/programs/wq/
terms of a detection speed both when using a CPU and GPU. plants/algae/lakes/AlgaeInformation.html, accessed: 2017-09-30.
Algae: 99%
Algae: 96%
Algae: 92% Algae: 96%
Algae: 83% Algae: 99%
Algae: 99% Algae: 99% Algae: 96% Algae: 99% Algae: 96%
Algae: 92% Algae: 96% Algae: 96%
Algae: 92% Algae: 96% Algae: 33% Algae: 53% Algae: 99%
Algae: 99% Algae: 83%
Algae: 99%
Algae: 99% Algae: 83% Algae: 59% Algae: 99% Algae:
Algae: 99% Algae: Algae: 99% Algae: 98% Algae: 66% Algae: 96%
96%
Algae: Algae: 33%
33% Algae:
Algae: 53%
Algae:
Algae: 99%
99%
Algae: 99%
99% 53%
Algae:
Algae: 59%
59% Algae: Algae: 98% Algae:
Algae: 98% Algae: 66%
66%
Algae: 99% Algae: 99%
99%
Algae: 99%
Algae: 99% Algae: 99%

(a) Algae Faster


detection
R‐CNN by R-CNN
Faster R‐CNN
Algae: 99%
Algae: 99% Algae: 65%
Algae: 99% Algae: 65%
Algae: 99% Algae: Algae:
65% 92%
Algae: 99% Algae: 76% Algae: 99% Algae: 80% Algae: 92%
Algae: 99% Algae: 92% Algae: 99% Algae: 99% Algae: 54%
Algae: 76% Algae: 99% Algae: 80%
Algae: 76% Algae: 99% Algae: 80% Algae: 82% Algae: 88% Algae: 99%
99% Algae: 99%
Algae:99% Algae: 54%
80% Algae: 54%
Algae: Algae:
Algae: 82%
82% Algae: 88%
Algae: 88% Algae: 80%
80%
Algae: 99% Algae: Algae:
Algae:
Algae: 99%
99%

RFCN

RFCN
(b) Algae RFCN
detection by R-FCN
Algae: 99% Algae: 83%
Algae: 99% Algae: 82%
Algae: 53%
Algae: 66% Algae: 99% Algae: 83%
Algae: 99% Algae: 99% Algae: 88%
Algae: 83% Algae: 82%
Algae: 99% Algae: 53% Algae: 82%
Algae: 53%
Algae: 66% Algae: 88%
Algae: 66% Algae: 88%

SSD

SSD
(c) Algae detection by SSD
SSD
Algae: 99%
Algae: 66%
Fig. 5. Images that show detection results for each model from aAlgae:
Algae: 99%
ground
99% view drawn in green.
Algae: 88% Algae: 99%
Algae: 94%
Algae: 99%
Algae: 99%
Algae: 99% Algae: 99%
Algae: 66% Algae: 91% Algae: 99% Algae:Algae:
99% 99%
Algae: 88% Algae: Algae: 58% Algae: 99%
Algae: 99% Algae: 99% Algae: 99%99%
Algae: 94% Algae: 88% Algae: 91%
Algae: 99% Algae: 99% Algae: 99%
Algae: 91% Algae: 99% Algae: 99%
Algae: 99% Algae: 99%
Algae: 99% Algae: 99% Algae: 99% Algae: 58%
Algae: 99% Algae: 58% Algae: 99%
Algae: 99% Algae: 91%
Algae: 99% Algae: 99%
Algae: 91%
Algae: 99%

(a) Algae detection by R-CNN Algae: 97%


Algae: 99% Algae: 99%
Algae: 99% Algae: 95% Algae: 97%
97% Algae: 98%
Algae: 99% Algae: 99% Algae:
Algae: 99% Algae: 99% Algae: 99%
Algae: 99%
Algae: 99% Algae: 95% Algae: 98%
Algae: 99% Algae: 99% Algae: 99% Algae: 99% Algae: 95% Algae: 99% Algae: 98%
Algae: 99% Algae: 99% Algae: 86%
Algae: 99% Algae: 99% Algae: 52%
Algae: 99%
Algae: 99% Algae: 86%
Algae: 99% Algae: 52%
Algae: 99% Algae: 99% Algae: 86%
Algae: 99% Algae: 52%
Algae: 99%
Algae: 99%
Algae: 99% Algae: 99%
Algae: 99%

(b) Algae detection by R-FCN

Algae: 98% Algae: 99%


Algae: 99% Algae: 55%
Algae: 98% Algae: 91% Algae: 99%
Algae: 98% Algae: 71% Algae: 99% Algae: 99% Algae: 55%
Algae: 87% Algae: 91% Algae: 99% Algae: 55%
Algae: 71% Algae: 91%
Algae: 87% Algae: 71% Algae: 99% Algae: 63%
Algae: 87% Algae: 91%
Algae: 99% Algae: 63%
Algae: 99% Algae: 63%
Algae: 91%
Algae: 91%

(c) Algae detection by SSD

Fig. 6. Images that show detection results for each model from an aerial view drawn in green.

[2] “Ocean’s living carbon pumps: When viruses attack giant algal [5] D. M. Anderson, “Approaches to monitoring, control and management
blooms, global carbon cycles are affected,” https://www.sciencedaily. of harmful algal blooms,” Ocean and coastal management, vol. 52, no. 7,
com/releases/2014/10/141021101510.htm, accessed: 2017-10-12. pp. 342–347, 2009.
[3] M. Y. Menetrez, “An overview of algae biofuel production and potential [6] S. R. Bhat and S. G. P. Matondkar, “Algal blooms in the seas around
environmental impact,” Environmental science & technology, vol. 46, india networking for research and outreach,” Current Science, vol. 87,
no. 13, pp. 7073–7085, 2012. no. 8, pp. 1079–1083, 2004.
[4] H. Fallowfield and M. Garrett, “The treatment of wastes by algal [7] T. Okaichi, Sustainable Development in the Seto Inland Sea, Japan.
culture,” Journal of Applied Microbiology, vol. 59, no. s14, 1985. Terra Scientific Publishing Company, 1997.
techniques for the detection, mapping and analysis of phytoplankton
blooms in coastal and open oceans,” Progress in oceanography, vol.
123, pp. 123–144, 2014.
[24] K. F. Flynn and S. C. Chapra, “Remote sensing of submerged aquatic
vegetation in a shallow non-turbid river using an unmanned aerial
vehicle,” Remote Sensing, vol. 6, no. 12, pp. 12 815–12 836, 2014.
[25] T.-Y. Lin, M. Maire, S. Belongie, J. Hays, P. Perona, D. Ramanan,
P. Dollár, and C. L. Zitnick, “Microsoft coco: Common objects in
context,” in European conference on computer vision. Springer, 2014,
(a) Incorrect detection of surrounding vegetation as algae pp. 740–755.
[26] M. Everingham, L. Van Gool, C. K. Williams, J. Winn, and A. Zisser-
man, “The pascal visual object classes (voc) challenge,” International
journal of computer vision, vol. 88, no. 2, pp. 303–338, 2010.
[27] S. J. Pan and Q. Yang, “A survey on transfer learning,” IEEE Trans-
actions on knowledge and data engineering, vol. 22, no. 10, pp. 1345–
1359, 2010.
[28] H.-C. Shin, H. R. Roth, M. Gao, L. Lu, Z. Xu, I. Nogues, J. Yao,
D. Mollura, and R. M. Summers, “Deep convolutional neural networks
for computer-aided detection: Cnn architectures, dataset characteristics
and transfer learning,” IEEE transactions on medical imaging, vol. 35,
no. 5, pp. 1285–1298, 2016.
(b) Inabilty to detect algae from ground and aerial view [29] S. Ren, K. He, R. Girshick, and J. Sun, “Faster r-cnn: towards real-time
object detection with region proposal networks,” IEEE transactions on
Fig. 7. Images that show incorrect or inability to detect algae. pattern analysis and machine intelligence, vol. 39, no. 6, pp. 1137–1149,
2017.
[30] R. Girshick, J. Donahue, T. Darrell, and J. Malik, “Rich feature
hierarchies for accurate object detection and semantic segmentation,”
[8] U. S. E. P. Agency, “A compilation of cost data associated with the in Proceedings of the IEEE conference on computer vision and pattern
impacts and control of nutrient pollution,” Tech. Rep., may 2015. recognition, 2014, pp. 580–587.
[9] K. D. Hambright, X. Xiao, and A. R. Dzialowski, “Remote sensing of [31] R. Girshick, “Fast r-cnn,” arXiv preprint arXiv:1504.08083, 2015.
water quality and harmful algae in oklahoma lakes [extension from 2013 [32] J. Dai, Y. Li, K. He, and J. Sun, “R-fcn: Object detection via region-
(final report)],” Oklahoma Water Resources Research Institute, 2015. based fully convolutional networks,” in Advances in neural information
[10] S. Jung, D. Kim, K. Kim, and H. Myung, “Image-based algal blooms processing systems, 2016, pp. 379–387.
detection using local binary pattern,” 2016. [33] W. Liu, D. Anguelov, D. Erhan, C. Szegedy, S. Reed, C.-Y. Fu, and A. C.
[11] Y. Wang, R. Tan, G. Xing, J. Wang, X. Tan, and X. Liu, “Energy-efficient Berg, “Ssd: Single shot multibox detector,” in European conference on
aquatic environment monitoring using smartphone-based robots,” ACM computer vision. Springer, 2016, pp. 21–37.
Transactions on Sensor Networks, vol. 12, no. 3, 2016. [34] J. Huang, V. Rathod, D. Chow, C. Sun, M. Zhu, M. Tang, A. Korattikara,
[12] C. S. Tan, P. Y. Lau, S.-M. Phang, and T. J. Low, “A framework for the A. Fathi, I. Fischer, Z. Wojna, Y. Song, S. Guadarrama, J. Uijlings,
automatic identification of algae (neomeris vanbosseae ma howe): U 3 s,” V. Kovalevskyi, and K. Murphy, “Tensorflow object detection api,” https:
in Computer and Information Sciences (ICCOINS), 2014 International //github.com/tensorflow/models/blob/master/research/object detection.
Conference on. IEEE, 2014, pp. 1–6. [35] “Large scale image classification,” https://mxnet.apache.org/tutorials/
[13] G. A. Carvalho, P. J. Minnett, L. E. Fleming, V. F. Banzon, and vision/large scale classification.html, accessed: 2018-02-21.
W. Baringer, “Satellite remote sensing of harmful algal blooms: A [36] S. Tang and Y. Yuan, “Object detection based on convolutional neural
new multi-algorithm method for detecting the florida red tide (karenia network.”
brevis),” Harmful algae, vol. 9, no. 5, pp. 440–448, 2010.
[14] L. Bennett, “Algae, Cyanobacteria Blooms, and Climate Change,” Cli-
mate Institute, Tech. Rep., April 2017.
[15] Y.H.Cheung and M.H.Wong, “Properties of animal manures and sewage
sludges and their utilisation for algal growth,” Agricultural Wastes,
vol. 3, no. 2, pp. 109–122, 1981.
[16] G. Hallegraeff and C. Bolch, “Transport of diatom and dinoflagellate
resting spores in ships’ ballast water: implications for plankton biogeog-
raphy and aquaculture,” Journal of Plankton Research, vol. 14, no. 8,
p. 10671084, 1992.
[17] ——, “Transport of toxic dinoflagellate cysts via ships’ ballast water,”
Marine Pollution Bulletin, vol. 22, no. 1, pp. 27–30, 1991.
[18] K. G. Sellner, G. J. Doucette, and G. J. Kirkpatrick, “Harmful algal
blooms: causes, impacts and detection,” Journal of Industrial Microbi-
ology and Biotechnology, vol. 30, no. 7, pp. 383–406, 2003.
[19] L. Peperzak, “Climate change and harmful algal blooms in the north
sea,” Acta Oecologica, vol. 24, no. 1, pp. 139–144, May 2003.
[20] H. B. Glasgow, J. M. Burkholder, R. E. Reed, A. J. Lewitus, and J. E.
Kleinman, “Real-time remote monitoring of water quality: a review
of current applications, and advancements in sensor, telemetry, and
computing technologies,” Journal of Experimental Marine Biology and
Ecology, vol. 300, no. 1-2, pp. 409–448, 2004.
[21] Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner, “Gradient-based learning
applied to document recognition,” Proceedings of the IEEE, vol. 86,
no. 11, pp. 2278–2324, 1998.
[22] A. C. Kumar and S. M. Bhandarkar, “A deep learning paradigm for
detection of harmful algal blooms,” in Applications of Computer Vision
(WACV), 2017 IEEE Winter Conference on. IEEE, 2017, pp. 743–751.
[23] D. Blondeau-Patissier, J. F. Gower, A. G. Dekker, S. R. Phinn, and V. E.
Brando, “A review of ocean color remote sensing methods and statistical

You might also like