You are on page 1of 11

Engineering Science and Technology, an International Journal 46 (2023) 101500

Contents lists available at ScienceDirect

Engineering Science and Technology, an International Journal


journal homepage: www.elsevier.com/locate/jestch

Full length article

Automated bacteria genera classification using histogram-oriented


optimized capsule network
Jitendra P. Chaudhari a , Hiren Mewada b ,∗, Amit V. Patel a , Keyur Mahant a
a
CHARUSAT Space Research and Technology Center and V. T. Patel Department of Electronics & Communication, Chandubhai S. Patel Institute of Technology,
Faculty of Technology & Engineering, Charotar University of Science and Technology, Changa, Gujarat, India
b
Electrical Engineering Department, Prince Mohammad Bin Fahd University, Al Khobar, Kingdom of Saudi Arabia

ARTICLE INFO ABSTRACT

Keywords: Identifying the nature and type of bacteria is essential in diagnosing various fatal diseases and their treatments.
Bacteria classification Biologists classify bacteria using morphological characterization from microscopic images according to their
Machine learning color and shape information. Therefore, an automated bacterial recognition and classification approach is
Deep learning
required compared to a challenging and time-consuming manual process. Much research has been carried
Capsule Network
out on bacteria classification using machine learning algorithms. However, the major weakness of these
Optimization
conventional machine learning algorithms is that they cannot differentiate pose and deformation, have
significant trainable parameters, require extensive training time, and use trial-and-error-based hyperparameter
selection. In addition, the choice of optimization function to reduce the loss is also equally important. This
paper presents an efficient capsule network that encodes information from orientation as an alternative to
these machine learning models. The proposed model is designed with histogram-based feature sets requiring
minimal parameters. Various optimization algorithms are tested to find an appropriate optimizer. 33 categories
of bacteria species are classified using the proposed method. A comprehensive analysis of popular gradient
descent optimizers is presented with a capsule network to strengthen testing and validation and benchmark
the performance. The extensive empirical study on DiBAS datasets demonstrates that the proposed network
performs 95.08% efficiency among various machine learning algorithms, including KNN, Decision Trees, Naïve
Bayes and SVM. Furthermore, the proposed model achieves better accuracy with the least training parameters
of 6.9 million.

1. Introduction their shape and other morphological information. Fig. 1(a–b) depicts
the classification of bacteria as positive gram and negative gram based
Research in bacteria classification would directly help save millions on the Gram stain test. Fig. 1(c–d) shows the type of bacteria based on
of people’s lives. It would also be helpful in the detection of their their shapes, such as cocci and bacilli.
diseases at an early stage. Bacterial species are crucial in the field of The laboratory staff must work with this complicated process to
microorganisms, including medicine, biochemistry, veterinary science,
classify these bacteria. Several steps in this complex process are rightly
farming, and the food industry. Bacteria classification is essential in
expressed in [2]. Microscopic observation is used to classify them based
medical science for identifying and treating bacterial diseases such
as cholera, pneumonia, gonorrhea, meningitis, and bacterial dysen- on specific characteristics in the final stage. A few features that help
tery. Their cellular envelopes have a distinctive shape that gives them to discriminate are color and shape. However, certain bacterial species
virulence and protects them from different chemicals. For scientists are morphologically complex, with cells of various sizes and shapes.
and doctors, understanding these cellular processes and mitigating the Therefore, manual classification is challenging and time-consuming
effects of these resistance properties poses a significant challenge [1]. since they need to know multiple bacterial characteristics. On the other
The bacterial recognition process requires the staining of samples and hand, the convolutional neural network (CNN) is the ultimate solution
dedicated instruments to analyze the characteristics of stained samples. in image-based classification with its deep and self-learning abilities.
Generally, bacteria can be classified in two ways: (a) using their cell CNN’s self-learning and automated learning processes determine the
wall composition and reaction to the Gram-stain test, and (b) using

∗ Corresponding author.
E-mail addresses: jitendrapchaudhari.ec@charusat.ac.in (J.P. Chaudhari), hmewada@pmu.edu.sa (H. Mewada), amitvpatel.ec@charusat.ac.in (A.V. Patel),
keyurmahant.ec@charusat.ac.in (K. Mahant).

https://doi.org/10.1016/j.jestch.2023.101500
Received 9 January 2023; Received in revised form 17 May 2023; Accepted 25 July 2023
Available online 17 August 2023
2215-0986/© 2023 The Authors. Published by Elsevier B.V. on behalf of Karabuk University This is an open access article under the CC BY-NC-ND license
(http://creativecommons.org/licenses/by-nc-nd/4.0/).
J.P. Chaudhari et al. Engineering Science and Technology, an International Journal 46 (2023) 101500

model is evaluated and compared with other machine learning-based


classifiers, including K-Nearest Neighborhood (KNN), Decision Trees
(DT), Naïve Bayes, and Support Vector Machine (SVM).
Overall main contributions of this paper are as follows:

• Over the year, learnable features replaced the handcrafted fea-


tures in the convolutional neural network. Nonetheless,
histograms play an essential role in the compact representa-
tion of statistical information. Therefore, we proposed an image
histogram-trained capsule network for bacterial colony classifica-
tion.
• Minimizing the training parameters is essential in designing a
deep neural network (DNN). Significant training parameters may
lead to overfitting and require more drop-out layers. Therefore,
the capsule network is implemented, requiring fewer trainable
parameters.
• Optimization helps to tune the weights and learning rate in DNN.
Furthermore, the selection optimization model based on the quan-
titative evaluation helps improve network performance. There-
fore, the effect of optimization model selection on classification
accuracy is presented in the paper.
Fig. 1. Bacteria species classification based on reaction to the Gram-stain test and • The proposed algorithm is validated using the DIBAS dataset
shape. consisting of 33 classes of the bacterial colony. After validation,
a comparative analysis is presented with widely used machine
learning algorithms, including KNN, decision tree, Naïve Bayes,
parameters that distinguish the patterns. Unfortunately, it increases SVM, and CNN.
CNN’s risk of overfitting or having too many parameters. Another • Finally, the proposed network is compared with other CNN net-
disadvantage of CNN is the requirement of large datasets to train works using trainable parameters and accuracy.
a deeper network with significant trainable parameters. In addition,
bacteria with large cell colonies lag in high-level features associated The remainder of this paper is organized as follows: Section 2
with adapting the convolutional networks. presents a study of recent literature on bacterial species classifica-
Bacterial images can be differentiated using their pose and texture tion. Section 3 explains the proposed approach with mathematical
pattern. CNN uses spatial and textural features independent of the expressions to classify bacteria into species. Section 4 contains an ex-
object’s pose or position in the image and can be present in different perimental examination of the proposed model. Finally, the conclusion
orientations or scales. Spatial features, conversely, are highly depen- and future work are expressed in Section 5.
dent on the object’s pose and orientation in the image and can vary
significantly as the object is rotated or translated. A feature detector 2. Related work
that relies solely on spatial information may not be able to distinguish
between different orientations of the bacteria and may confuse the As mentioned previously, shape and color information are the main
texture with changes in pose. Thus, distinguishing the pose, texture, or characteristics of bacteria. Wide varieties of image classification al-
deformation within an image using CNN is challenging, and only spatial gorithms are available. These algorithms can be categorized as tra-
features might not be adequate for texture classification [3]. The pool- ditional or machine learning approaches. Further machine learning
ing operation causes the removal of some essential features. In a capsule approaches can be classified as statistical machine learning [9,10] and
network, the group of neurons is expressed as a capsule that encodes the neural network-based [11] machine learning models. In traditional ap-
geographical information and likelihood of finding an object [4]. On proaches, microbiologists classify bacterial colonies using their shape,
the other hand, the vector-based datastore capability of the capsule net- size, and color information. Molecular and biochemical testing is used
work [3] provides deeper object information. In addition, the routing to classify the species with huge color and shape similarities. These
mechanism enables each capsule to keep information from the previous methods are sensitive to the operator’s skills and expertise. Therefore,
capsule layers and try to find an agreement to classify an image. machine learning approaches are widely adopted due to their promising
Thus, their memorizing capability of orientation and components’ lo- results without human intervention. This section presents machine
cations motivated us to use them for Bacteria classification. Numerous learning approaches.
medical imaging applications have recently utilized capsule networks,
such as biomedical image classification [5,6], bioinformatics [7], and 2.1. Statistical machine learning models:
hyperspectral image classification [8].
This paper presents an alternate solution for CNN to mitigate the Statistical machine learning algorithms are a three-step automated
bacteria classification problem. Usually, a capsule network is trained tool applying (1) segmentation: Gaussian mixture model (GMM), fuzzy
using images (in the spatial domain), leading to many feature sets and C-mean, and K-means clustering; (2) feature extraction: GMM, principle
trainable parameters. In addition, for better training of the network, component analysis, independent component analysis, texture features
it needs many images; therefore, an image augmentation process is using wavelet, local binary pattern, etc.; and (3) classification using a
required. In contrast, the proposed model uses histogram-based reduced supervised or unsupervised classifier, i.e., KNN, Adaboost, multiplayer
feature sets in CNN training. Images are preprocessed to remove noise, perceptron network, etc.
and histogram-oriented features are obtained from the clean images. A comprehensive study on machine learning to classify bacteria
These features are used in capsule network training without any image over the last decade was presented in [12]. All models studied in this
augmentation. Later, a capsule network requiring fewer parameters is paper used different datasets, so a standard comparison is not shown
proposed for the classification process. For validation, the proposed in this paper. In [13], authors obtained features using the Gabor filter
network is tested using different optimizers. Furthermore, the proposed from the segmented images, and KNN was used for classification. Later,

2
J.P. Chaudhari et al. Engineering Science and Technology, an International Journal 46 (2023) 101500

they integrated the texture features using wavelet transform and color size in independent Staphylococcus aureus cultures by analyzing image
information using a few statistical features to improve classification sequences.
accuracy [6,14]. Finally, the feature set was reduced using principal A thin-film transistor sensor obtained the bacterial colony images
component analysis (PCA), and a multi-layer perceptron network was in [30]. Later, two CNNs were trained using spatiotemporal features
used as a classifier. The algorithm was tested on UTMC.V1.DB and for early detection and classification. Khalifa et al. [31] focused on
UTMC.V2.DB datasets and obtained an accuracy of 80.8% and 80.1%, the data augmentation process to solve the overfitting problem of
respectively. A similar three-step approach was presented in [15]. They CNN. They increased the sampled images of the DIBaS dataset using
look for the primary two challenges, e.g., segmentation of the bacterial the augmentation process. The model was trained using 6600 images
region due to a cluttered environment and occlusion. After handling and validated using 5940 images. The network achieved 98.22% ac-
these challenges, they extracted morphotype features and classified curacy. The authors in [32] presented foodborne pathogenic bacteria
them using adaptive boosting. Sajedi et al. [16] extracted features using recognition. Using principal component analysis (PCA), they reduced
the Gabor transform, and XGBoost was used to classify myxobacterial the feature sets and proposed a stacking algorithm for classification.
images. Their model accuracy was limited to 90.28% for three classes. Further improvement in the model’s performance was obtained using
Authors in [17] identified tuberculosis, distinguished between sin- a grid search and K-fold validation. The results were compared with
gle and multiple tuberculosis bacteria and non-bacteria, and compared KNN and SVM. The stacking algorithm achieved 95.73% accuracy
back-propagation and KNN methods. Bacteria length and endpoint were using SVM. A foodborne bacteria classification from hyperspectral mi-
used as features of tuberculosis bacteria. An accuracy 93.22% was croscopic images using deep learning was presented in [33]. They
obtained for backpropagation networks, while 94.22% accuracy was compared the 1D-CNN, deep residual network (ResNet), and long short-
obtained for KNN. Borowa et al. [18] presented a multistage clas- term memory (LSTM) network using 5000 images. 1D-CNN performed
sification model using ResNET and attention-based multiple instance best among all with 96.2% accuracy. The studies’ summary is tabulated
learning (AbMIL). First, they used ResNet to categorize image patches in Table 1.
as darker, rounded, or larger. Later, the weighted combination of this The study propagates that statistical approaches have limited accu-
patch representation was classified using AbMIL. Finally, they classified racy of up to 80%. On the other hand, a supervised CNN algorithm
the bacteria colony into three classes with an accuracy of 64.8%. integrating bacterial features outperformed traditional or statistical
techniques. In addition, the image augmentation process in CNN has
improved the classification rate, obtaining more than 90% accuracy.
2.2. Artificial neural network-based algorithms:
However, using spatial features obtained from CNN makes it challeng-
ing for texture pattern classification. Further, a system is required to
Zachariou et al. [19] used morphological operations to detect tu-
evaluate the quality of the image-augmented dataset. Also, a stand-
berculosis bacteria and a multi-stage deep CNN to count them. This
alone CNN could obtain better accuracy at the cost of high training
method lacks the classification of these bacteria and needs manual
times. Therefore, there is a need to design a CNN that can use texture
microscopy to get a bacteria region. Kim et al. [20] initially used
pose information, achieve better accuracy without the image augmen-
bacteria’s interacting components to obtain varied viscosity and surface
tation process, and offer the least number of trainable parameters,
tension and classified it as one of the six bacterial species with an
reducing overall training time. Thus, the proposed method presents the
accuracy of 85%. Jeong et al. [21] analyzed infrared (IR) spectroscopy
capsule network as an alternate solution to this heterogeneous CNN
spectra to identify the bacteria using wavelet features. A visual inspec-
approach, offering fewer training parameters.
tion tool was designed considering the 32 IR spectra of E. coli and
the 40 IR spectra of Pseudomonas aeruginosa. It has been observed 3. The proposed method
that logistic regression has a better accuracy of 92.5% compared to
various machine-learning approaches. In [22], Zielinski et al. used a This section presents the proposed classification approach using a
computationally complex process. They integrated the SIFT features capsule network. Initially, the selection of the capsule network and its
with convolutional features obtained from the last layer of CNN, and architecture are explained. Subsequently, histogram features and the
its fisher vector (FV) was obtained. Later, FV was concatenated with network training process are described. Finally, the optimization func-
features of CNN. The SVM was trained using these features and suc- tion is essential to reducing the loss in the training phase. Therefore,
ceeded in achieving 97% accuracy. A bacterial strain classification this section also presents various optimization models suitable for the
using biological microscope images was proposed by Trivedi et al. [23]. neural network.
They augmented image datasets to increase the number of images in
training datasets, and CNN was developed and optimized to classify 3.1. Capsule network:
bacterial strains.
A classification of vaginal bacteria was presented from the micro- The proposed method is based on capsule networks, which can
scopic imagery in [24]. Initially, bacterial segmentation was done, and differentiate pose and deformation compared to conventional CNNs.
features obtained from these segmented images were used to train SVM. The traditional CNN uses a max-pooling operation to remove crucial
Later, CNN was compared with the SVM classifier, and they observed attributes from the images. In contrast, the capsule network uses the
that the 3-class SVM outperformed the CNN model. Anna Plichta [25] squash function, which uses all the information from the previous layer
used seven features, including shape, color, size, number of clusters, in the classification. Therefore, the capsule network is more effective
cluster shape, density and cell distribution. A weighted classifier was than conventional CNNs and is more reliable when the images vary in
used for the recognition. The presented algorithm obtained 90.45% shape or pose. Capsule networks fetch spatial information and related
accuracy with the DIBaS dataset. Chopra and Verma [26] developed features. This information is used to overcome the loss of features in
a hybrid method to train the CNN using a genetic algorithm and clas- pooling operations. The output of the capsule is a vector with direction.
sify bacteria into four species. Several regularization techniques were For example, changing the orientation of the image will change the
introduced to tackle the overfitting problem [27]. Bacanin et al. [28] vector in the same direction. The result of the neuron is a scalar. It
presented a sine-cosine-based swarm intelligence model to regularize does not vary with a change in direction. Fig. 2 depicts the capsule
the dropout and successfully tested it on MNIST datasets. Further test- network’s architecture.
ing of the complex image classification model is required. An algorithm The capsule network starts with a standard convolution layer having
of bacterial growth estimation over time is presented in [29]. Trained a size equal to the input image/signal size. An appropriate kernel at
CNNs were utilized to estimate bacterial colonies’ growth rate and each layer extracts the features from the image. For example, each

3
J.P. Chaudhari et al. Engineering Science and Technology, an International Journal 46 (2023) 101500

Table 1
Summary of findings from the literature.
Ref Dataset and Algorithm Finding
[13] Database: UTMC sample image Texture information is used. Accuracy is limited to 65.76%.
Features: Gabor feature Further improvement in accuracy is required.
Classifier: KNN
[14] Database: UTMC sample image Spectral features are used.
Features: Wavelet features and color information Overall accuracy on the dataset is limited to 87.81%.
The feature set reduction using PCA. Classifier: multi-layer perceptron
network
[15] Database: DIBAS The average accuracy is 79.10% for 9 classes.
Features: Morphological features using unsupervised segmentation However, for more than 10 bacterial classification,
Classifier: Adaboost algorithm fails with an accuracy of 54.2%.
[17] Database: tuberculosis bacteria The main limitation is that morphological traits are not always clear.
Features: ROI extracted using Morphological operation Two class classification is presented.
Classifier: KNN Received 94.92% accuracy for binary classification.
[19] Database: tuberculosis bacteria This method lacks the classification of these bacteria
Features: morphological features and needs manual microscopy to get a region of bacteria.
Classifier: a multi-stage deep CNN for bacteria detection
[18] Database: 3 clones of Klebsiella pneumonia Novel approach of image patch categorization using shape is presented.
Features: Patch shape obtained using ResNET
Classifier: Attention-based multiple instant learning The classification accuracy is limited to 64.8%.
The clustering of clones can help to improve classification accuracy.
[20] Database: Authors generated dataset The process of obtaining viscosity and surface tension needs a complex
Features: viscosity and surface tension chemical process.
Classifier: SVM classifier They classified it into six categories with an accuracy of 85%.
[21] Dataset: Bacterial images obtained using Fiber-Optic Fourier Transform A binary classification was proposed. Achieved 92.5% accuracy.
Infrared (FO-FTIR) Sensing Algorithm validation is required for multi-class classification.
Feature: Continuous wavelet transform-based features
Classifiers: MLP, SVM and LR
[22] Features: SIFT-based texture features, CNN’s spatial features and Fisher High classification accuracy is achieved.
vector calculation Feature size is large due to SIFT and CNN.
Classifier: Concatenation of CNN, FVFC and SVM The use of CNN for feature extraction and integrating multiple
classifiers makes the system computationally intensive.
[24] Database: Florescent microscopic images SVM outperformed CNN for 3-class classification.
Images are enhanced using dual-scale Laplacian of Gaussian
Features: morphological features
Classifier: Naïve classifier, SVM and CNN
[25] Database: DIBAS 20 classes were used for classification.
Features: shape, color, size, number of clusters, cluster shape, density The claimed accuracy is 90.45%.
and cell distribution The classifiers’ details are not provided.
Classifier: Weighted average of Seven classifiers
[26] Features: obtained using autoencoder feature extraction. Hyperparameter optimized using PSO.
Classifier: CNN Training time is reduced in GA optimization.
Genetic algorithm used to learn the best hyperparameter for CNN. Categorized data in three classes with 94.9% accuracy.
[31] Database: DIBAS A large number of training and testing datasets, i.e. 6600 images after
Features: Convolutional features augmentation, were used.
Classifier: CNN The model succeeded with an accuracy of 98.22% with a large training
Data augmentation is used in the training network. time.
[34] Database: DIBAS Atrous convolution is used in VGG16 instead of spatial convolution.
Features: Convolutional features Achieve an accuracy of 93.38%.
Classifier: pre-trained VGG16 network Large learnable parameters i.e. 134.4M
[35] Database: DIBAS Transfer-learning approach to training the networks and hence less
Features: Convolutional features training time
Classifier: Shufflenet and the squeezenet neural network The network has the least number of parameters; however, its
Data augmentation using zooming and doubling to improve the precision and recall are very low.
network accuracy.
[36] Database: DIBAS Depth-wide convolution operations were used in feature extraction.
Features: Convolutional features Accuracy is 86%.
Convolutional features without image augmentation
Classifier: CNN

bacteria has a different pose or shape. The first layer extracts features size, hue, angle, thickness, etc. Thus, this capsule gives the position
representing different parts of the shape, for example, curves in each and orientation of each part in 2D space. The length of this small
shape. The layer parameters, i.e., stride and spatial dimension, are deck gives the likelihood of entity presence and parameter information
changed accordingly. Further non-linearity is induced in the feature encoded in its orientation [37]. The functionality of the capsule layer
map using a rectified linear unit function (ReLU). The number of is illustrated in Fig. 3, where 𝑢1 , 𝑢2 , and 𝑢3 represent the capsule, 𝑊
convolution layers is determined with a specific application in hand. is the transformation matrix, 𝑉 gives the predicted features, and 𝐶 is
The obtained scaler output from the convolution layers is stacked into the assigned weight. These primary capsules create secondary capsules,
tiny decks in the capsule layer, where a small subset of each deck is representing the probability that the input bacterial image belongs to
referred to as a capsule. This capsule stores valuable information like each possible class. The network’s final output is a set of class capsules,

4
J.P. Chaudhari et al. Engineering Science and Technology, an International Journal 46 (2023) 101500

Fig. 2. Architecture overview of the proposed capsule network-based method.

operation can be expressed as:



𝑠𝑗 = 𝑐𝑖,𝑗 𝑣𝑗,𝑖 (2)
𝑖

Squashing Function: The information is condensed in this last


component. The role of the squashing function is to take all the
information and convert it into a vector that is less than or equal
to 1. It also maintains the direction of the vector. Step, Sigmoid,
Tanh, and ReLu are the variants of the squashing function. ReLU
is not saturated and computationally simple; it has been used in
the proposed design. The squashing operation is obtained using
Fig. 3. Capsule layer functionality using matrix multiplication and squashing.
the following equation:
‖𝑠𝑗 ‖2 𝑠𝑗
𝑣𝑗 = (3)
each containing a probability distribution over the possible classes. The 1 + ‖𝑠𝑗 ‖2 ‖𝑠𝑗 ‖
bacterial class with the highest probability is the predicted class for the 𝑣𝑗 ≈ ‖𝑠𝑗 ‖𝑠𝑗 , for 𝑠𝑗 is small
input image. ‖𝑠𝑗 ‖ (4)
The capsule network has four significant processes which make it 𝑣𝑗 ≈ , for 𝑠𝑗 is large
𝑠𝑗
distinct from other conventional CNN:
c. Dynamic routing algorithm: The dynamic routing algorithm per-
a. Matrix Multiplication: The matrix multiplication is applied to un- mits the various components to transfer information. It is an iter-
derstand the spatial part of the image. Here, the transformation ative process where higher-level capsules get input from lower-
matrix is multiplied by the vector obtained from the preceding level capsules. This routing algorithm adjusts the weights 𝑐𝑖,𝑗
convolution layer producing the higher-level feature vector v. in the scalar weighting function. In Eq. (1) a, 𝑣𝑗,𝑖 expresses the
The matrix multiplication operation can be expressed as: prediction from capsule 𝑖 on output capsule 𝑗. The 𝑣𝑗 expresses
the similarity between the two capsules. Let the similarity vector
𝑣𝑖,𝑗 = 𝑤𝑖,𝑗 𝑢𝑖 (1) is expressed using 𝑏𝑖,𝑗 as:

b. Scalar weighting of the input: It is responsible for computing 𝑏𝑖,𝑗 = 𝑣𝑗,𝑖 𝑣𝑗 (5)
a higher-level capsule that would receive the current capsule
The weight coefficient 𝑐𝑖,𝑗 is computed using the softmax of 𝑏𝑖,𝑗
output. If all outputs 𝑣1 , 𝑣2 , 𝑣3 have the same prediction, then
as
a weighted combination of feature vectors 𝑣1 , 𝑣2 , 𝑣3 increases ∑ 𝑏
the probability of predicting the following layer features. This 𝑐𝑖,𝑗 = 𝑒𝑏𝑖,𝑗 ∕ 𝑒 𝑖,𝑗 (6)

5
J.P. Chaudhari et al. Engineering Science and Technology, an International Journal 46 (2023) 101500

Fig. 4. 9-patches, 196-bins histogram of the sample image.

The architecture of capsule networks consists of six layers. The first


three layers are the encoders. Encoders convert the input image into
a vector. The last three layers are called decoders. Finally, decoders
reconstruct the image using vectors. The configuration of the capsule
network for the bacterial classification requires input data size, the total
number of classes, and the routing iteration.

3.2. Histogram-oriented feature sets:

Creating a large dataset for training the neural network is time-


consuming. And designing and training a network without a good
number of datasets is the most challenging task. Augmentation is
widely employed to increase the dataset’s size. However, the quality of
augmented datasets heavily relies on the original datasets. Therefore,
the augmentation process may need a system to evaluate the aug-
mented datasets. In the proposed network, we used histogram features
with the capsule network. The histogram of color helps the neural
network exploit color information in an image. This also reduces the
size of the features in the capsule network’s first layer.
In the image, pixel color is represented using a 3D vector, where
Fig. 5. Flowchart of the proposed method for bacteria classification (Solid line
each vector gives red, green, and blue colors. The hue-saturation-value
represents network training and dotted lines represent network testing).
(HSV) space separates color components, making the histogram less
susceptible to color variation. Therefore, an RGB color space with
three histograms for each color component is proposed. We used a
histogram of nine patches with 196 bins in the experiment. These optimization; (ii) optimization in the training process, i.e., optimizing
generated patches are shown in Fig. 4. The histogram bins used in the learning rate, loss function, batch size, epoch, and activation function;
experiment are 196 bins (whole image), 2 × 98 (horizontal patches), and (iii) optimizing the regularization hyper-parameters, i.e., L1, L2,
2 × 98 (vertical patches), and 4 × 49 (horizontal and vertical corner and drop-out regularization values. The optimum selection of learning
patches). Thus, 784 histograms are obtained as feature vectors from rates provides faster convergence. And it tunes the weights in the
the image and are used to train the capsule network. backpropagation process, improving the neural network’s performance.
Therefore, various optimizers have been tested with the proposed cap-
The process using this capsule network is as follows:
sule network to minimize losses. The summary of the primarily used
• Pre-processing: In this step, bacteria images are preprocessed for optimizer with CNN networks is as follows:
noise removal using median filtering. Median filters are applied Loss reduction can be achieved by optimizing the network. The
to each image channel separately. Then the three image channels optimizer adjusts or tunes the weights and learning rate to reduce loss.
are combined again. However, many optimization functions keep the learning rate constant.
• Features Extraction: In the feature extraction phase, histogram- Therefore, various optimizers have been tested with the proposed cap-
based features from the noise-free images are extracted. Nine sule network to minimize losses further. The summary of the primarily
patches are obtained, and their 196 bins are concatenated to used optimizer with CNN networks is as follows:
create the feature vector. Finally, this feature vector is resized to (a) Stochastic Gradient Decent (SGD): This model updates the net-
an image, and a 2D capsule network is used. work’s parameter after loss computation on each training sample and
• Classification: Bacteria images are classified into various types offers less convergence time. It adapts both the learning rate component
using a capsule network. The capsule network is trained using (𝛼) and gradient of the learning rate (𝜕𝐿∕𝜕𝑤) to update the weight
the histogram features of patches. Different optimization func- parameters as expressed in the equation below. But there is a high
tions are also tested during training to see their performance. fluctuation in the parameters; hence, it has a high variance.
Finally, the proposed capsule network architecture’s performance 𝜕𝐿
𝑤𝑡+1 = 𝑤𝑡 − 𝛼 (7)
is compared with KNN, decision trees, Naïve Bayes, SVM, and 𝜕𝑤
deep learning methods. (b) AdaGrad: This second-order optimizer changes the learning rate
for each parameter at every step. Thus, it avoids manual tuning of the
The proposed method for bacteria classification is depicted in Fig. 5.
learning rate. The weight update equation can be expressed as follows:

𝛼 𝜕𝐿
3.3. Network optimization: 𝑤𝑡+1 = 𝑤𝑡 − √ (8)
𝑣𝑡 + 𝜀 𝜕𝑤𝑡
Optimizing the network can reduce the loss. There are mainly three However, its learning rate is a monotonically decreasing function,
categories of CNN hyper-parameter optimization. (i) network structure and the second-order derivative is computationally intensive.

6
J.P. Chaudhari et al. Engineering Science and Technology, an International Journal 46 (2023) 101500

(c) AdaDelta: AdaDelta solves the decaying problem of learning rate 4. Experiments and results
by accumulating all past squares of gradients over a fixed window
size. The equation below replaces the learning rate component with 4.1. Experiment setup:
the exponential moving average of the square delta (𝐷). However, it is
computationally expensive. The proposed experiment integrates color information into the cap-
√ sule network for classification. The proposed method is tested on the
𝐷𝑡+1 + 𝜀 𝜕𝐿
𝑤𝑡+1 = 𝑤𝑡 − √ (9) DIBaS dataset [22] with 660 (33 bacteria species 𝑥 20 images per
𝑣𝑡 + 𝜀 𝜕𝑤𝑡
species) images. The image resolution is 2048 × 1532 pixels. The sam-
Where, ple images from the dataset are shown in Fig. 6. Parameters are tuned to
[ ] obtain better accuracy. The bacteria images obtained from the dataset
𝐷𝑡 = 𝛽𝐷𝑡−1 + (1 − 𝛽) 𝛥𝑤𝑡 (10)
are preprocessed for noise removal. Median filtering is applied for
[ ] each channel separately. Then, the noise-free image is merged. During
𝜕𝐿
𝑣𝑡 = 𝛽𝑣𝑡−1 + (1 − 𝛽) (11) the feature extraction phase, histogram-based color feature extraction
𝜕𝑤𝑡
is performed. As explained in Section 3, nine patches representing
𝛥𝑤 = 𝑤𝑡+1 − 𝑤𝑡 (12) global and local histograms of color images are used as a feature
vector. Eventually, because of the patch generation, the experiment
(d) Root Mean Square Propagation (RMSP): RMSP is an adaptive
does not require augmentation to overcome the overfitting problem due
learning approach for improving AdaGrad. A cumulative sum of a
to limited datasets. These features are used for training the system. The
gradient in AdaGrad may become a stuck model when close to con-
train–test split algorithm randomly splits the test and train data. We
vergence. RMSP overcame this problem using exponential moving and
have used 70% data for training and 30% of the data for testing the
avoided the aggressive decay in the learning rate. Update the weight
network during our experimentation. Various classification algorithms
as follows:
are trained, including KNN, decision trees, Naïve Bayes (NB), SVM, and
𝛼 𝜕𝐿 conventional CNN. The performance of the trained system is evaluated
𝑤𝑡+1 = 𝑤𝑡 − √ (13)
𝑣𝑡 + 𝜀 𝜕𝑤𝑡 using 30% test data in terms of accuracy, precision, recall and f-
(e) Adaptive moment estimation (Adam) combines the gradient measure. The hyperparameters of each classifier are listed in Table 2.
descent function with momentum and RMSP. Thus, using these two The hyper-parameters are tuned using a grid search algorithm, and a
methods’ advantages optimizes gradient descent more. In addition, it 10 K-fold cross-validation is used to validate the models.
offers fast convergence and resolves the learning rate. The weight can A conventional CNN network is designed with four hidden layers
be updated by using and a max-pooling operation. The last two layers are dense, and the
𝛼 overall trainable parameters for this CNN are 27727. The structure of
𝑤𝑡+1 = 𝑤𝑡 − √ 𝑚
̂𝑡 (14) the capsule network used for the proposed experiment is presented in
𝑣̂𝑡 + 𝜀
Table 3.
where,
𝑚𝑡 𝑣𝑡 4.2. Results and discussion:
̂𝑡 =
𝑚 and 𝑣̂𝑡 = (15)
1 − 𝛽1𝑡 1 − 𝛽2𝑡
(f) AdaMax: Adam’s generalization is AdaMax, where the l2-norm The capsule network is trained with 100 iterations. Training can run
of Adam is changed to 𝑙∞ , and sometimes, it can produce better well if we also use validation data. The learning rate value is initialized
optimization than Adam. Therefore, the weight is updated as follows: to 1e-3, which adjusts the speed of the parameter to be updated. The
𝛼 experiment is performed with 30 epochs; images are shuffled in each
𝑤𝑡+1 = 𝑤𝑡 − 𝑚 ̂ (16) epoch, and validation data is used with a validation frequency 30. The
𝑣𝑡 𝑡
model is further evaluated to find the best optimizer. All quantitative
where,
parameters are assessed and listed in Table 3. Informally, accuracy
𝑚𝑡
̂𝑡 =
𝑚 (17) refers to the percentage of correct predictions made by our model. The
1 − 𝛽1𝑡 following is the formal concept of accuracy:
𝜕𝐿
𝑚𝑡 = 𝛽1 𝑚𝑡−1 + (1 − 𝛽1 ) (18) Number of correct prediction
𝜕𝑤𝑡 Accuracy =
( ) Total number of prediction (21)
| 𝜕𝐿 |
𝑣𝑡 = 𝑚𝑎𝑥 𝛽2 𝑣𝑡−1 , || |
| (19) =
𝑇𝑃 + 𝑇𝑁
| 𝜕𝑤𝑡 | 𝑇𝑃 + 𝑇𝑁 + 𝐹𝑃 + 𝐹𝑁
(g) NAdam: The inclusion of Nesterov momentum in ADAM instead Where true positive (TP) is the correct prediction of bacteria, false
of RMSP creates a NAdan optimizer. Two problems in ADAM sometimes negative (FN) gives sensitivity to the model, false positive (FP) pro-
cause worse optimization. Firstly, the stochastic gradient update is not vides specificity to the model, and true negative (TN) gives a correct
within the historical gradient span. Secondly, magnitude parameters prediction of the negative bacteria. Using these parameters, precision,
cause variation in network function; however, the magnitude of Adam recall, and F-Score, exactness and completeness can be measured using
parameters is unaffected by the gradient descaling. Therefore, tweaking the Eqs. (22) and (24).
Adam is proposed [38] considering the 𝑙2 -norm of the gradient vector 𝑇𝑃
instead of the magnitude of the gradient vector. The NAdam optimizer Precision = (22)
𝑇𝑃 + 𝐹𝑃
is expressed as follows:
( ) Recall =
𝑇𝑃
(23)
𝛼 1 − 𝛽1 𝜕𝐿 𝑇𝑃 + 𝐹𝑁
𝑤𝑡+1 = 𝑤𝑡 − √ ̂𝑡 +
𝛽1 𝑚 (20)
𝑣̂ + 𝜀
𝑡
1 − 𝛽 𝑡 𝜕𝑤𝑡 1 Precision x Recall
F-Score = 2 ∗ (24)
where 𝑚𝑡 is expressed in Eq. (18) and 𝑣𝑡 is obtained using Eq. (11). Precision + Recall
Optimizers mentioned above have been widely used in many CNN- Better optimizers are primarily concerned with being quicker and
based applications. The selection of a particular optimizer can lead to more effective, but they are also frequently recognized for being more
better loss reduction. Therefore, the proposed model is tested with these generalizable (less overfitting) than others. The selected optimizer has
optimizers and finds the best fit for classification. a significant impact on how well the model works.

7
J.P. Chaudhari et al. Engineering Science and Technology, an International Journal 46 (2023) 101500

Fig. 6. Sample images from DIBaS dataset used in experiments.

Table 2
Hyper-parameters for classifiers.
KNN: LDA:
N 5 No of components 2
Nayeve Base: SVM:
Variance 1e−9 Kernel RBF
Capsule:
Batch size 100 Epoch 50
Learning rate 0.001 Routing 3
Node in primary caps 256 Dimension of each capsule 8
Routing time 5 Batch Size 100

Table 3
Layer structure of the proposed capsule.
Layer Output Shape Parameters Connected to
input_1 (Input Layer) (None, 28, 28, 1) 0
Convolution layer (None, 20, 20, 256) 20992 Input Layer
Primarycap_conv2D (None, 6, 6, 256) 5308672 Convolution layer
primarycap_reshape (None, 1152, 8) 0 Primarycap_conv2D
primarycap_squash (None, 1152, 8) 0 primarycap_reshape
Secondarycap_conv2D (None, 2, 16) 294912 primarycap_squash
input_2 (InputLayer) (None, 2) 0
Mask (None, 32) 0 Secondarycap_conv2Dinput_2 (InputLayer)
capsnet (Length) (None, 32) 0 Secondarycap_conv2D
decoder (Sequential) (None, 28, 28, 1) 1345808 Mask
Total parameters 6,970,384
Trainable parameters 6,970,384
Non-trainable parameters 0

however, NADAM trained the model faster than other optimizers. On


the other hand, SGD and Adadelta have the lowest performance com-
pared to other optimization functions. The obtained training loss and
testing loss graph for ADAM is presented in Fig. 8.
A confusion matrix obtained from a proposed network gives valu-
able insights into the model’s performance in classifying images into
their respective categories. A five-fold confusion matrix in Fig. 9 shows
the model’s performance for each class. A few types are not classified,
for example, Staphylococcus aureus and Staphylococcus saprophyticus,
because of their color, shape, and orientation similarity.
The precision vs recall curve (PR curve) can provide insights into
the performance of the CNN for different classification thresholds.
A high precision value indicates that the CNN correctly classifies a
high proportion of predicted positive cases. In contrast, a high recall
value suggests that CNN correctly classifies a high proportion of actual
positive cases. Fig. 10 shows the PR curve for validation datasets.
Further analysis using precision, recall, F1-score, and accuracy for
Fig. 7. Accuracy analysis over the epochs of various optimizer.
each optimizer is presented in Table 4. ADAM provides the best perfor-
mance among all optimization models. Therefore, ADAM is used with
the capsule network to obtain a better classification rate.
Fig. 7 shows the convergence of the various optimizers over the Further, various classifiers, i.e., KNN, decision trees, Naïve Bayes,
epoch. The ADAM takes time to converge but obtains the best accuracy and SVM, have been considered to validate the model. The histogram-
among all. The performance of NADAM, RMSP, and ADAMAX is close; based features are obtained from the noise removal images in all these

8
J.P. Chaudhari et al. Engineering Science and Technology, an International Journal 46 (2023) 101500

Table 4
Performance evaluation of capsule network with different optimizer.
Optimizers Precision Recall FMeasure Accuracy
Adadelta 0.59106 0.49751 0.44701 0.49751
Adagrad 0.734236 0.676617 0.650898 0.676617
Adam 0.959886 0.960796 0.942739 0.950796
Adamax 0.774883 0.716418 0.704564 0.716418
Nadam 0.764819 0.721393 0.695114 0.721393
RMSP 0.751382 0.726368 0.708243 0.7263
SGD 0.131922 0.447761 0.195412 0.44776

Table 5
Performance measurement of various machine learning algorithms.
Parameter Precision Recall F Measure
KNN 0.78312 0.731343 0.717629
DT 0.660978 0.616915 0.614203
Fig. 8. Training and Testing loss of proposed model using ADAM. NB 0.652992 0.59204 0.586787
SVM 0.722511 0.686567 0.66932
Deep Learning CNN 0.709105 0.656716 0.640989
Capsule networks 0. 959886 0. 960796 0. 942739

classifiers. Table 5 presents the performance using precision, recall,


and F-measure. As depicted in Table 5, capsule network-based classi-
fication performance is better than other supervised machine learning
techniques.
Further, a comparison of the proposed model with literature hav-
ing a CNN-based bacterial species classification model is presented
in Table 6. The results reported in Table 6 show that the proposed
capsule network generates consistent results with one or more param-
eters better than other implemented methods. In [34], Patel performed
atrous convolution using VGG 16 to achieve an accuracy of 93.38%.
Fig. 9. Confusion matrix for 33 classes. In contrast, Garcia et al. [35] used a transfer-learning approach to
train the shufflenet and the squeezenet neural networks. Their model
has the least parameters; however, its precision and recall are very
low. In [39], Talo used the 5-fold cross-compilation technique and
achieved an accuracy of 99.10%. In [36], CNN was modified using a
depth-wide convolution operation and validated using augmentation
and without augmentation. They succeeded in obtaining 86% accuracy
without image augmentation. In [22], the authors extracted features
using a Fisher vector, and SVM was used for classification. They re-
ceived 83.4% accuracy among all classes. A conventional convolution
operation is replaced with depth-wise convolution in CNN, and species
classification is presented with an overall accuracy of 86%. The number
of parameters for each model is also listed in Table 3. It suggests that
the proposed model successfully maintains an accuracy of 95.08% and
requires few training parameters, i.e., 6.9 M, compared to other models.

5. Conclusions and future works

In the medical domain, diseases are predicted by analyzing patient


samples. Identification of appropriate bacteria types and microorgan-
isms plays an important role. In earlier days, this was done manually
using microscopic image observation. But the manual process is time-
consuming and complex. In this paper, we have proposed an automated
methodology for classifying bacteria images into the appropriate type.
Thirty-three bacteria species were used for analysis and classifica-
tion. In this study, capsule network-based bacteria classification using
histogram-based features is employed to classify bacteria species into
33 classes. This model has reached an average classification accuracy of
95.07%, and the total training parameters are 6.9M. The performance
of the proposed system is compared with that of various classifiers,
Fig. 10. Precision vs. Recall curve. including KNN, decision trees, Naïve Bayes, SVM, and deep learning.
The proposed model uses histogram features in classification. The
similarity in shape and orientation in a few classes has a similar
classifiers. It has been observed that a histogram-oriented capsule net- histogram; therefore, the model fails to classify such classes. So global
work provides maximum accuracy compared to other machine-learning and local histogram concatenated feature sets can be used in the future

9
J.P. Chaudhari et al. Engineering Science and Technology, an International Journal 46 (2023) 101500

Table 6
A quantitative comparison with various machine learning frameworks without image augmentation.
CNN Methods No of images Parameters Accuracy Precision Recall F Measure
Atrous convolution based VGG 16 [34] 660 134.5 M 0.9338 0.93 0.9410 0.9338
Squeeznet_1_0 [35] 669 0.751 M 0.8895 0.4366 0.4487 0.4047
Shufflenet_V2 [35] 669 5.543 M 0.6486 0.6827 0.6486 0.6266
FisherVector SIFT + SVM [22] 660 58.42 M 0.834 – – –
Depthwise Separable CNN [36] 689 3.23M 0.864 0.8899 0.8317 85.98
Proposed Capsule network 689 6.9M 0.9508 0.959 0.960 0.942

to improve performance. Further used datasets have a limited set of [15] Y. Song, L. He, F. Zhou, S. Chen, D. Ni, B. Lei, T. Wang, Segmentation,
images for each class, which may weaken the model in the training splitting, and classification of overlapping bacteria in microscope images for
automatic bacterial vaginosis diagnosis, IEEE J. Biomed. Health Inf. 21 (4) (2016)
process. A further extension of datasets with appropriate preprocessing
1095–1104.
algorithms needs to be investigated. Thus, the main focus is to improve [16] H. Sajedi, F. Mohammadipanah, A. Pashaei, Image-processing based taxon-
the system’s performance by considering different combinations of omy analysis of bacterial macromorphology using machine-learning models,
features, reductions in execution time, and extensions of datasets. Multimedia Tools Appl. 79 (2020) 32711–32730.
[17] L.N. Sahenda, M.H. Purnomo, I.K.E. Purnama, I.D.G.H. Wisana, Comparison of
tuberculosis bacteria classification from digital image of sputum smears, in:
CRediT authorship contribution statement 2018 International Conference on Computer Engineering, Network and Intelligent
Multimedia, CENIM, IEEE, 2018, pp. 20–24.
Jitendra P. Chaudhari: Conceptualization of this study, Methodol- [18] A. Borowa, D. Rymarczyk, D. Ochońska, M. Brzychczy-Włoch, B. Zieliński, Deep
learning classification of bacteria clones explained by persistence homology, in:
ogy. Hiren Mewada: Software, Writing – original draft. Amit V. Patel:
2021 International Joint Conference on Neural Networks, IJCNN, IEEE, 2021,
Software, Writing. Keyur Mahant: Modification for the final layout. pp. 1–8.
[19] M. Zachariou, O. Arandjelović, W. Sabiiti, B. Mtafya, D. Sloan, Tuberculosis
Declaration of competing interest bacteria detection and counting in fluorescence microscopy images using a
multi-stage deep learning pipeline, Information 13 (2) (2022) 96.
[20] S. Kim, A.S. Day, J.-Y. Yoon, Machine learning classification of bacterial species
The authors declare that they have no known competing finan- using mix-and-match reagents on paper microfluidic chips and smartphone-based
cial interests or personal relationships that could have appeared to capillary flow analysis, Anal. Bioanal. Chem. 1 (2022) 1–10.
influence the work reported in this paper. [21] S.-Y. Ji, D.H. Jeong, M. Hassan, I.K. Ilev, Signature infrared bacteria spectra
analyzed by an advanced integrative computational approach developed for
identifying bacteria similarity, IEEE J. Sel. Top. Quantum Electron. 25 (1) (2018)
References 1–8.
[22] B. Zieliński, A. Plichta, K. Misztal, P. Spurek, M. Brzychczy-Włoch, D. Ochońska,
[1] J.E. Sykes, Gram-negative Bacterial Infections, Elsevier Inc, 2014. Deep learning approach to bacterial colony classification, PLoS One 12 (9) (2017)
[2] B.D. Satoto, M.I. Utoyo, R. Rulaningtyas, E.B. Koendhori, Custom convolu- e0184554.
tional neural network with data augmentation and bayesian optimization for [23] S. Trivedi, N. Patel, N. Faruqui, Bacterial strain classification using convolutional
gram-negative bacteria classification, Int. J. Intell. Eng. Syst. 13 (5) (2020) neural network for automatic bacterial disease diagnosis, in: 2023 13th Interna-
524–538. tional Conference on Cloud Computing, Data Science & Engineering, Confluence,
[3] S. Sabour, N. Frosst, G.E. Hinton, Dynamic routing between capsules, in: IEEE, 2023, pp. 325–332.
Advances in Neural Information Processing Systems, Vol. 30, 2017, pp. 1–11. [24] Y. Yu, J. Wang, Automated enumeration and classification of bacteria in
[4] G.E. Hinton, S. Sabour, N. Frosst, Matrix capsules with EM routing, in: fluorescent microscopy imagery, in: 2018 IEEE Life Sciences Conference, LSC,
International Conference on Learning Representations, 2018, pp. 1–15. IEEE, 2018, pp. 57–60.
[5] G.U. Nneji, J. Deng, H.N. Monday, M.A. Hossin, S. Obiora, S. Nahar, J. [25] A. Plichta, Recognition of species and genera of bacteria by means of the product
Cai, COVID-19 identification from low-quality computed tomography using a of weights of the classifiers, Int. J. Appl. Math. Comput. Sci. 30 (3) (2020)
modified enhanced super-resolution generative adversarial network plus and 463–473.
siamese capsule network, in: Healthcare, Vol. 10, No. 2, MDPI, 2022, p. 403. [26] C. Chopra, R. Verma, Novel methods based on cnn for improved bacteria
[6] H.K. Mewada, A.V. Patel, M. Hassaballah, M.H. Alkinani, K. Mahant, Spectral– classification, in: Proceedings of Fifth International Congress on Information and
spatial features integrated convolution neural network for breast cancer Communication Technology, Springer, 2021, pp. 1–16.
classification, Sensors 20 (17) (2020) 4747. [27] G.H. De Rosa, J.P. Papa, X.-S. Yang, Handling dropout probability estimation
[7] J. Khanal, H. Tayara, Q. Zou, K. To Chong, DeepCap-Kcr: accurate identification in convolution neural networks using meta-heuristics, Soft Comput. 22 (2018)
and investigation of protein lysine crotonylation sites based on capsule network, 6147–6156.
Brief. Bioinform. 23 (1) (2022) bbab492. [28] N. Bacanin, M. Zivkovic, F. Al-Turjman, K. Venkatachalam, P. Trojovskỳ, I.
[8] M.E. Paoletti, S. Moreno-Álvarez, J.M. Haut, Multiple attention-guided capsule Strumberger, T. Bezdan, Hybridized sine cosine algorithm with convolutional
networks for hyperspectral image classification, IEEE Trans. Geosci. Remote Sens. neural networks dropout regularization application, Sci. Rep. 12 (1) (2022) 1–20.
60 (2022) 1–20, http://dx.doi.org/10.1109/TGRS.2021.3135506. [29] S.Á. Nagy, L. Makrai, I. Csabai, D. Tôzsér, G. Szita, N. Solymosi, Bacterial Colony
[9] S. Trattner, H. Greenspan, G. Tepper, S. Abboud, Automatic identification of Size Growth Estimation by Deep Learning, Cold Spring Harbor Laboratory, 2023,
bacterial types using statistical imaging methods, IEEE Trans. Med. Imaging 23 pp. 1–6, bioRxiv.
(7) (2004) 807–820. [30] Y. Li, T. Liu, H.C. Koydemir, H. Wang, K. O’Riordan, B. Bai, Y. Haga, J. Kobashi,
[10] K. De Bruyne, B. Slabbinck, W. Waegeman, P. Vauterin, B. De Baets, P. Van- H. Tanaka, T. Tamaru, et al., Rapid detection and classification of bacterial
damme, Bacterial species identification from MALDI-TOF mass spectra through colonies using a thin film transistor (TFT) image sensor and deep learning
data analysis and machine learning, Syst. Appl. Microbiol. 34 (1) (2011) 20–29. (Conference Presentation), in: Optics and Biophotonics in Low-Resource Settings
[11] T. Treebupachatsakul, S. Poomrittigul, Bacteria classification using image pro- IX, SPIE, 2023, p. PC123690A.
cessing and deep learning, in: 2019 34th International Technical Conference on [31] N.E.M. Khalifa, M.H.N. Taha, A.E. Hassanien, A.A. Hemedan, Deep bacteria:
Circuits/Systems, Computers and Communications, ITC-CSCC, IEEE, 2019, pp. robust deep learning data augmentation design for limited bacterial colony
1–3. dataset, Int. J. Reasoning-Based Intell. Syst. 11 (3) (2019) 256–264.
[12] S. Kotwal, P. Rani, T. Arif, J. Manhas, S. Sharma, Automated bacterial classifi- [32] Z. Wan-dan, S. Ru-jin, W. Cheng-wei, L. Qian-xue, X. Zhi-ping, Raman
cations using machine learning based computational techniques: Architectures, spectroscopy classification of foodborne pathogenic bacteria based on PCA-
challenges and open research issues, Arch. Comput. Methods Eng. (2022) 1–22. stacking model, in: 2019 International Conference on Intelligent Informatics and
[13] H. Sajedi, F. Mohammadipanah, H.K. Shariat Panahi, An image analysis-aided Biomedical Sciences, ICIIBMS, IEEE, 2019, pp. 304–307.
method for redundancy reduction in differentiation of identical Actinobacterial [33] R. Kang, B. Park, M. Eady, Q. Ouyang, K. Chen, Single-cell classification of
strains, Future Microbiol. 13 (3) (2018) 313–329. foodborne pathogens using hyperspectral microscope imaging coupled with deep
[14] H. Sajedi, F. Mohammadipanah, S.A.H. Rahimi, Actinobacterial strains recog- learning frameworks, Sensors Actuators B 309 (2020) 127789.
nition by Machine learning methods, Multimedia Tools Appl. 78 (14) (2019) [34] S. Patel, Bacterial colony classification using atrous convolution with transfer
20285–20307. learning, Ann. Roman. Soc. Cell Biol. (2021) 1428–1441.

10
J.P. Chaudhari et al. Engineering Science and Technology, an International Journal 46 (2023) 101500

[35] R.G. García, S.J. Rodríguez, B.B. Martínez, C.H. Gracidas, R.M. Torres, Effi- Hiren K. Mewada has obtained his M.Tech. and Ph.D. degree from Sardar Vallbhbhai
cient deep learning architectures for fast identification of bacterial strains in National Institute of Technology — Surat, Gujarat, India. Presently he is Assistant
resource-constrained devices, 2021, arXiv preprint arXiv:2106.06505. Research Professor at Prince Mohammad Bin Fahd University, Kingdom of Saudi Arabia.
[36] D.-T. Mai, K. Ishibashi, Small-scale depthwise separable convolutional neural Previously he was associate professor at Charotar University of Science and Technology,
networks for bacteria classification, Electronics 10 (23) (2021) 3005. Gujarat, India. He has more than 17 years teaching experience. His current areas of
[37] B. Mamidibathula, S. Amirneni, S.S. Sistla, N. Patnam, Texture classification interest are computer vision, signal processing, machine learning and Embedded System
using capsule networks, in: Iberian Conference on Pattern Recognition and Image design. He has published more than 60 research papers and completed several funded
Analysis, Springer, 2019, pp. 589–599. research projects. He is coauthor of one book and published five book chapters. He is
[38] Z. Zhang, Improved adam optimizer for deep neural networks, in: IEEE/ACM member of IETE and ISTE.
26th International Symposium on Quality of Service, IWQoS, IEEE, 2018, pp.
1–2.
[39] M. Talo, An automated deep learning approach for bacterial image classification, Amit Patel completed his Ph.D. from Charusat University in 2015. He is presently
2019, pp. 1–5, arXiv preprint arXiv:1912.08765. working as an Associate scientist in Charusat Space Research and Technology center
and Assistant Professor in Electronics and Communication Engineering Department,
CSPIT change. He is associated with a number of research projects sponsored by the
Board of Research in Nuclear Science and SAC-ISRO, Govt. of India. He has published
more than 40 research papers in various international journals and conferences.
Jitendra P. Chaudhari received the B.E. degree in Electronics Engineering from
Nagpur University, Nagpur, M.E. degree in Industrial Electronics engineering from M.S.
University, Baroda, Gujarat, India and Ph.D. from Charotar University of Science and Keyur Mahant is an Assistant Scientist/ Professor at CHARUSAT Space Research and
Technology in 1996, 2001 and 2016 respectively. He is working as Associate Professor Technology Center, Changa, Gujarat, India. He received B.Tech. in Electronics and
in Department of Electronics and Communication Engineering at Charotar University communication from C U Shah College of engineering and technology, India in 2008.
of Science and Technology, Changa, Gujarat. Presently he is deputed as Coordinator, He has also received M.Tech. in Communication system from CHARUSAT University,
CHARUSAT Space Research Technology Center. He is member of IE(I) and IETE. His India in 2011. He is currently involved with the microwave filters, power dividers
research interest includes, Image Processing, RF and Microwave. and circulator design using SIW technology for FMCW Radar Application, funded by
Board of Research in Nuclear Sciences (BRNS), Department of Atomic Energy (DAE),
Government of India.

11

You might also like