Professional Documents
Culture Documents
By
Awais Sultan
Reg #
Supervised by
Master of Science
In
Computer Science
at
October, 2020
A LOW COST INTERACTIVE VISUALIZATION TECHNIQUE
FOR HIGH DIMENSIONAL MULTI FACETED MEDICAL DATA
By
Awais Sultan
Reg #
Supervised by
Mr. Muhammad KashifNazir
Co-Supervised by
Mr. Syed MudassarAlam
Masters of Science
In
Computer Science
at
Riphah International University,
Faisalabad Campus, Pakistan
October, 2020
APPROVAL SHEET
“I hereby certify that the above candidate's work, including the thesis, has been completed to
my satisfaction and that the thesis is in a format and of an editorial standard recognized by the
faculty/department as appropriate for examination. The thesis has been checked through
Turnitin for plagiarism (test report attached).”
Signature (s):
Stamp: ______________________
Signature (s):
Date: ____________
SAMPLE
“DECLARATION
I certify that the research work presented in this thesis is my own and to the
best of my knowledge. All sources used and any help received in the
preparation of this dissertation havebeen acknowledged. I hereby declare that
I have not submitted this material, either inwhole or in part, for any other
degree at this or any other institution.”
Name:
Registration no: 1
Signature:
iv
ACCEPTENCE CERTIFICATE
(Department will provide you these pages)
v
“ACKNOWLEDGEMENT
First and foremost, I would like to profoundly praise the Almighty Allah SWT for
enabling me to see this great moment. I would like to thank and express my deepest
gratitude and appreciation to my supervisor Mr.MuhammadKashifNazirand co-
supervisorMr. Syed MudassarAlamwho gradually helped me in every way which I
needed to go through all difficulties. I have been extremely honored to have a
supervisor who cared so much about my work, and who responded to my questions
and queries so promptly. I really thank him and without his excellent guidance this
thesis would not have been possible.
I would also like to thank the rest of the faculty member of Riphah International
University FSD campus who have gradually offered their time, expertise, wisdom and
encouraged me to complete my thesis work in a better way.”
vi
“DEDICATION
vii
TABLE OF CONTENTS
TITLE................................................................................................i
DECLARATION…….………………………..……………....…...iv
ACCEPTANCE FORM………….……………………..….............v
ACKNOWLEDGEMENT……….…………………………....…..vi
DEDICATION…….……………………………………………...vii
TABLE OF CONTENTS…………………………......................viii
LIST OF TABLES………………………………………….……..xi
LIST OF FIGURES…………………………………......................x
LIST OF ABBREVIATIONS……………………………….........xi
ABSTRACT…………………………………………………….....xi
i
Chapter 1 INTRODUCTION...................................................................................1
1.1 “Introduction.................................................................................1
1.2 Research Background...................................................................1
1.3 Problem Statement........................................................................5
1.4 Research Questions......................................................................5
1.5 Research Objective.......................................................................5
1.6 Research Aim................................................................................5
1.7 Thesis Organization......................................................................5
1.8 Chapter Summary.........................................................................6
Chapter 2 LITERATURE REVIEW.......................................................................7
1 Introduction...................................................................................7
2.1 Background Concept....................................................................7
2.1.1 X-ray............................................................................7
2.1.2 CT-Scan.......................................................................9
2.2 Related Work................................................................................9
2.3 Chapter Summary.......................................................................27
Chapter 3 RESEARCH METHODOLOGY.....................................................28
viii
3.1 Introduction.................................................................................28
3.2 Research Framework..................................................................28
3.3 Tools and Data set......................................................................29
3.1.1 Blood smear images.........................................................29
3.4 Simulation Parameters...............................................................30
3.5 Chapter Summary.......................................................................30
Chpter 4 PROPOSED METHOD..................................................................32
4.1 Introduction.................................................................................32
4.1 Proposed Method........................................................................32
4.1.1 Parkisons Detection Using CNN..............................33
4.1.2 Brain Tumor.................................................................34
4.1.3 Cardiovascular disease using DNN.............................35
4.1.4 Pneumonia detection using CNN.................................36
4.1.5 Blood cancer detection using CNN..............................37
Chapter Summary...................................................................................38
Chapter 5 RESULTS AND DISCUSSION.....................................................39
5.1Introduction........................................................................................39
5.1 Results and Analysis........................................................................39
Chapter Summary...................................................................................46
Chapter 6 CONCLUSION AND FUTURE DIRECTIONS.................................47
6.1 Introduction.................................................................................47
6.1 Concluding Remarks........................................................................47
6.2 Limitations........................................................................................48
6.3 Future Directions and Research Opportunities.............................48
REFERENCES...................................................................................49
ix
LIST OF TABLES
x
LIST OF FIGURES
xi
LIST OF ABBREIVATIONS
DL Deep Learning
IV Interactive visualization
ML Machine learning
VA Visual analysis
xii
ABSTRACT
Medical imaging plays a vital part in diagnosing and treating illnesses since they
permit the perception of the patient’s condition in a non-invasive way. With the
advancement of diagnostic equipment, medical pictures have ended up increasingly
available. Medical images such as X-rays, research facility tests, MRI, Ultrasound, or
C.T. checks of different sorts require careful reading and exact interpretation at any
time, any put and any device that gets to be urgent require for medical experts.
Current imaging strategies for unsafe medications in computer integration and
environmental conservation as information develops and becomes more complex day
by day. On the other hand, extra-political infrastructure was required in terms of
usefulness and memory aptitudes, which may not be accessible in healing centers and
clinics with restricted assets. Hence, the easiest way to supply successful health
benefits is through a remote access system on a distributed framework.In this study, a
distributed interactive visualization framework is distributed, and order is proposed to
distinguish high-quality, sophisticated, and organized data to get superior and more
exact results for medical experts Using cardiac remodeling and mass removal
methods.
xiii
Chapter 1
INTRODUCTION
1.1 “Introduction
The topic of research thesis is An Interactive Web-Based Medical Image
Visualizationtechniquein Distributed System to Visualize High Dimensional Complex
Data. In this chapter there are some different sections in which inspiration of research
is discussed. In this chapter a brief introduction of Information Visualization (IV),
Visual Analytic System (VAS) Par Sets technique, and Association Rule Mining are
explained. This chapter provides the overview about different Information
Visualization applications and highlights the background about the understanding gap
problem of data interpretation. Research problem derived from literature, objectives
of research, research questions and significance are discussed in this chapter.”
2
Medical square rendering may be a set of procedures for visualizing two-dimensional
(2D) of three-dimensional information sets, such as attractive reverberation imaging
(MRI) and computed tomography (C.T.) gotten by a therapeutic scanner. The propels
and dispersal of the Web have started intrigued by creating tools for analyzing and
handling therapeutic pictures within the Web. It permits you to form OpenGL
situations in browser windows and programming three-dimensional intelligently
situations with the JavaScript dialect. WebSocket could be a arrange convention that
allows two-way communication between client and server. Not at all like the
Hypertext Exchange Convention (HTTP), communication between two machines
(client and server) is dynamic, which empowers the spread and improvement of
substance in real-time. Distributed frameworks speak to a structure in which the
computing hubs in an organized communicate utilizing messages to coordinate
themselves into a common objective. A curious highlight of this system is that the
failure of one center should not influence the complete framework.
Publisher-Subscription is an asynchronous informing demonstrate where publishers
and supporters trade messages without knowing each other. News is called occasions.
Within the Pups show, supporters’ flag to the broker that they are fascinated by an
event (Moraes et al. 2019). The distributor sends messages (occasions) to the broker,
which is capable of conveying the message to the enlisted supporter at the event. It is
an open-source in-memory database that can be utilized as a broker since it empowers
the PupSup demonstrate. Redis permits the broker to disseminate over numerous hubs
(Li et al., 2020).
A few attempts to create information visualization frameworks on the Web are
accessible within the literature. Recently, numerous restorative imaging applications
have utilized the GAN system. Most considers (Fried-Adar et al. 2018) to use the
GAN method from picture to picture to create a label-to-segmentation interpretation,
segment-to-image interpretation, or clinical cross-system interpretation. (Prim, Baer,
et al. 2016) prepared a full-fledged organize to memorize retinal vessel section
pictures. At that point, they learned the interpretation from the twofold character tree
to the modern retina picture. It is utilized to produce division pictures of lung areas
and heart from chest x-ray pictures (James and Dashrath 2014).
Studies referred to two GAN systems as a section and a critic and learned the
interpretation between brain MRI pictures and a brain tumor parallel segment graph.
Patch-based GAN has been prepared for interpretation between brain C.T. pictures
3
and the comparing MRI pictures. They encourage recommended an auto-environment
show for picture decontamination (Brehm, Clem, et al. 2016).
Recent studies have presented a cross-sectional picture utilizing GAN, illustrating
liver lesions from stomach C.T. picture to PET check picture. A few thinks about
have been motivated by the GAN strategy of picture location (Coates and
Stavropoulos 2014). (Stalling, Westerhoff and Hege 2005) prepared GAN with solid
tissue within the retinal locale to memorize information dispersion of healthy tissue.
They tried the GAN on the patches of undetectable sound and sporadic information
for identifying peculiarity in retinal pictures. The complexity of restricted data within
the field of therapeutic imaging has prompted research into the strategies of
counterfeit information increase to enlarge medical datasets. Within the show
consider, we center on moving forward the comes about within the assignment of
liver ulcer classification. (Bauer et al. 2013) utilized the GAN system to coordinated,
high-quality liver injury pictures (here we employ the terms rawimages, and ulcer
ROIs traded).
According to the World Health Organization, 8.2 million individuals around the world
passed on of cancer in 2012 alone, of which 745,000 were analyzed with liver cancer.
Central liver injuries are dangerous and show metastases or kind (e.g., hemangioma or
liver sores) (Kopassas et al. 2007). Computed tomography (C.T.) is one of the
foremost standard and secure imaging methods to identify, identify, and take after up
on liver injuries. Subsequently, there’s an expanding requirement and interest in
creating automated diagnostic instruments based on C.T. pictures to help radiologists
in identifying liver injuries. Previous studies have proposed strategies to consequently
classify central liver injuries on C.T. pictures (D.S. Yu, M. J. Ackermann, W. E.
Lorenzen, W. Schroeder, V. Salina, S. Aylward, D. A.).
(Holsinger and Jurisica 2014) utilized the framework highlights for liver ulcer
classification into four categories, counting the ordinary liver parenchyma course.
They used a hierarchical classifier of neural systems at each level. It got three sorts of
highlights for each tumor, relying on the structure, shape, and development bend of
the isolated tumors. In reverse, eradication was utilized to choose the leading
combination of highlights using a double calculated regression examination to classify
tumors (Zhang and Metaxas 2016).
4
1.3 Problem Statement
Medical image visualization techniques for centralized complex and extensive
distributed data require more auto computational infrastructure to improve the
meshing problem and segmentation of holes and face regular reversals.
Novelty:
In this research, an efficient and Web-based interactive framework is proposed to
visualize large and complex medical data images in a distributed system.
This application will correct meshing problems automatically to provide better
and authentic results in the visualization of medical images.
This research work can be used by medical experts, policymakers, health
professionals, insurance companies, and government agencies for funding.
The goal of this research is to provide data analytics and data visualization to explain
various aspects of the covid'19 using the currently available datasets. To gain a clear
understanding of the disease and to find ways to deal with it, further investigation is
needed. This study will include an application of machine learning methods focused
on various age groups to provide a better understanding of covid'19.
1.7 Thesis Organization
5
diminishes the doctor’s time compared to the other existing framework to check the
patient’s health. The web-based interface enables experts to pick up get to at anytime
and anyplace.
6
Chapter 2
LITERATURE REVIEW
1 Introduction
This chapter explain the existing work in interactive visualization in detail. The main
aim of this chapter explores the related work and explore all research gaps in detail
which is not discussed yet.The purpose of finding gaps in existing work help the new
researchers to find better and better solution of given problem. In the area of computer
vision, deep learning has recently played an important role. The elimination of human
judgment in the diagnosis of diseases is one of its applications. In particular, the
diagnosis of brain tumors needs high precision, where minute judgment errors can
result in a tragedy. For this reason, for medical purposes, brain tumor segmentation is
an important challenge. There are currently multiple tumor segmentation techniques,
but all lack high precision.
7
the JavaScript language to create three-dimensional virtual environments (Li, Yu,
Feng, & Zhao, 2020).
Intelligent augmented simulation conditions, for example, CAVE frameworks
have been tried in genuine symptomatic exercises, for example, 3D reverberation
investigation, and it has been indicated that they are equipped for uncovering
abandons that would sidestep 2D examination and that they have a higher expectation
to absorb information than customary radiological workstations. Just a moderately set
number of simultaneous clients, most regularly one, can deal with current vivid VR
frameworks that commonly require head checking and other equipment that doesn't
coordinate well with the working climate of the typical radiology division. Without
the standard single perspective or single client constraints of normal sound system
shows, numerous procedures have been proposed to advance genuine 3D perception
and every one of them have been applied to clinical information representation
(Preim, Klemm, et al., 2016). A snappy overview of the subject, with extraordinary
accentuation on the most firmly related methodologies. The key specialized element
of 3D presentations is heading particular light discharge, which is generally cultivated
by volumetric, holographic, or multiview procedures. Volumetric presentations blend
light fields by extending light shafts on refractive/intelligent media mounted or moved
in a vacuum.
Lenticular showcases, then again, amplify the pixel lattice of the projecting
unit, making dim zones between survey spaces. The Cambridge Multiview board is an
exemplary plan here, and a few makers produce shows dependent on variations of this
innovation. For instance, Spatial View Inc.1 has endeavored to bring to the market a
clinical workstation with an autostereoscopic show and a coordinated vision-based
interface framework (Frid-Adar et al., 2018). It's a develop innovation. A 3D sound
system impact is acquired when the left and right eyes see distinctive yet coordinating
information. In any case, the set number of perspectives subject to veils or lenticulars
in multiview frameworks produces cross-talks and discontinuities dependent on the
observer's movement.
Moreover, Pastrello et al. are of the assessment that it is valuable to join these
informational collections while utilizing pathways and protein collaboration
organizations; the subsequent organization should have the option to zero in on either
a huge scope view or more explicit limited scope sees, contingent upon the
investigation issue and test objectives. In their paper, the creators delineate a work
process that is helpful for incorporating, investigating and envisioning information
from various sources, featuring significant instrument qualities to empower these
examinations to be coordinated (Gillies, Kinahan, and Hricak, 2016).
Lenticular presentations, then again, amplify the pixel lattice of the projecting unit,
making dim zones between review spaces. The Cambridge Multiview board is an
exemplary plan around there, and a few makers produce shows dependent on
variations of this innovation. For instance, Spatial View Inc.1 has endeavored to bring
to the market a clinical workstation with an autostereoscopic show and an
incorporated vision-based interface framework (Frid-Adar et al., 2018). It's a develop
innovation. A 3D sound system impact is acquired when the left and right eyes see
distinctive yet coordinating information. Nonetheless, the predetermined number of
perspectives reliant on covers or lenticulars in multiview frameworks produces cross-
talks and discontinuities dependent on the onlooker's movement.
Moreover, Pastrello et al. are of the assessment that it is valuable to join these
informational indexes while utilizing pathways and protein association organizations;
the subsequent organization should have the option to zero in on either an enormous
scope view or more explicit limited scope sees, contingent upon the investigation
issue and test objectives. In their paper, the creators delineate a work process that is
13
helpful for coordinating, investigating and picturing information from numerous
sources, featuring significant apparatus qualities to empower these examinations to be
incorporated (Gillies, Kinahan, and Hricak, 2016).
“It is very difficult to visually depict 3D forms, partly due to the lack of detail
when the 2D (retinal) image is projected onto the 3D model. While a complex
relationship between the lighting and the geometry, orientation, and texture of the
object affects the pattern of light on the retina, different 3D shapes may have created
the same pattern of light sensations on the retina.
Therefore, visual form comprehension is generally ambiguous. The ambiguity
of diffusely shaded images, which is called bas-relief ambiguity, is not resolved by
any increase in illumination. One of the earliest depth mechanisms, despite this
complexity, is thought to be evolutionary shape-from-shading and is very successful.
The visual system relies on experience and on some assumptions to solve the
ambiguities. Surfaces tend to be regarded as convex, for instance (Stalling,
14
Westerhoff, &Hege, 2005). These assumptions are not always appropriate and can
cause the surface category and local orientation to be perceived incorrectly. In
contrast, the most used model of the human visual system indicates a single point of
light which is above and to the right. This statement has important consequences for
many perceptual phenomena besides type perception (Preim, Baer, Cunningham,
Isenberg, &Ropinski, 2016). There is, however, some indication that some (locally
independent) light sources may simply be inherently believed by the human visual
system. In contrast, the visual system is remarkably immune to light differences in
certain conditions.
For the correct interpretation of material properties, there are also indications
that more natural lighting conditions are required, such as multiple light sources. The
perception of 3D forms happens at different spatial scales. At least two levels must be
distinguished: a local scale, where the shape of individual objects is assessed, and a
global scale, where spatial relationships are assessed, including deep relationships and
proximity of objects (Bauer, Wiest, Nolte, & Reyes, 2013). In reality, there is
substantial evidence that the human visual system depicts the entire scene in a linear
scale space, with a large number of scales, where each scale is a replica of the scene
that has been convoluted by a Gaussian kernel (and subsequent scales increase the
size of the kernel; for more. Therefore, different scales may be used in studies on the
impact of depth signals.
Most surfaces are textured. This is a breach of the assumption that light is
influenced by neighboring areas of a surface in the same way and poses a challenge
for both edge-detection-based segmentation and shape-from-shading techniques.
Texture does, however, provide shape data. While there is a considerable amount of
knowledge about the large-scale image structure One of the first examinations of
texture is Gibson's (Scholl, Aach, Deserno, &Kuhlen, 2011). The most popular
texture structure model, which models texture components as Gabor patches, is Julesz
and Caelli (a sinusoid convoluted with a 2D Gaussian). Interestingly, Gabor patches
indicate a clear resemblance to the receptive field system of human vision. In
determining a surface's local curvature, texture is particularly helpful.
For example, surface textures representing principal curvature directions
(PCDs) increase the perception of shape: observers tend to interpret lines as curvature
directions on a surface. In visualization, texture has been used to represent significant
shape properties. Lines on a surface will separate the spectator into meaningful
15
substructures (Cobzas, Birkbeck, Schmidt, Jagersand, & Murtha, 2007). If shapes are
familiar, viewers look for characteristics that make such a difference possible.
Interrante and colleagues have shown that a certain style of line widely used by
illustrators assists in this distinction.”
These lines are called valley lines and represent areas of a curved surface
where the curvature along the PCD has a local minimum (i.e. the place where the
surface is flattest). The occlusion of surrounding structures is strongly affected by
these areas and are therefore drawn in dark colors. Since there are not enough features
that can be seen with valley lines, it is possible to add ridge lines, defining regions
with a local curvature limit in the PCD (i.e. the regions with the highest surface
curvature; these lines for statistical definitions and calculation algorithms) (T.S. Yoo,
M. J. Ackerman, W. E. Lorensen, W. Schroeder, V. Chalana, S. Aylward, D.
Metaxas, 20. Metaxas, 20). Such a coarse image of a surface may be helpful in
displaying an external surface in a multi-layer visualisation (e. g. to show the surface
of an organ and a deep-seated tumor as well as the surrounding risk structures).
This is a promising contrast to a semi-transparent display where ordinary
depth signals, such as occlusion and shading, can hardly be recognized by a
transparent surface. There is some debate about whether texture signals can be
correctly interpreted in orthographic projection when a 3D model is presented (a
traditional medical simulation situation). Li and Zaidi realized that the surface must
be perceived" with a noticeable amount of perspective projection, the surface must be
viewed " (Holzinger&Jurisica, 2014). Kim and colleagues, however, notice that
curvature-directed lines express form even with orthogonal projection. Using only
ridge lines could be "uninformative" if most of them are almost aligned with the
viewing direction. Therefore, a combination of ridge and valley lines offers enhanced
effectiveness (Zhang & Metaxas, 2016). Ng gives an example of a soccer player
whose MDP optimizer knows a plan and maximizes expected rewards through the use
of a mistake.
The football player was honoured by the idea that possessing time had to do
with scoring goals (Ra, Ab, & Kc, 2020). The officer stayed at the ball instead of
using ball possession to progress across the field and started "vibrating" to allow the
entire set of ball contacts (Deussen, Colditz, Stamminger, &Drettakis, 2002). This
error can apply as a specification problem, can be interpreted (no incentive for the ball
16
touch agent), convergence (the reward frequency for ball touch transfer is too high)
and optimisation (D. Wu et al., 2017).
Some MDPs are small enough with precise algorithms to solve the
consequences of eradication and erosion, but in most realistic MDPs, the Monte-Carlo
methods are necessary. In these situations, the typical solution is to run MDP
simulation software and then use the Monte Carlo optimizer for a virtual optimal
policy gradient check (Fuchs et al., 1989).A variety of ad hoc research systems for
MDP professionals may be used to iteratively investigate spatial criteria for a
"informed trial and error" (Costa et al., 2020). For the first manual transformations,
MDP practitioners typically compose an immersive client and then an ideological rule
for viewing State growth (Van Ham & Van Wijk, 2004).Even if the scientists have no
previous visual image and can automatically connect to their MDP model and MDP
optimizer, this is assumed. MDPVIS shall make three contributions.
Both MDP operations are provided by the visualisation thesis. For Markov
Decisions, we describe the MDP as the regular MDP, with the initial status distributer
from the given horizon [5,30] 4 = a,,,, SAPRμ-P0 = a. S is the final set of World
countries; A is a finite amount of potential acts in any Regime; PS S: the likelihood of
state [0, 1] depends on State Operations. Rs a (,) is compensated after State
Operations; μp [0, 1] is an aspect of discount and P0 is rewarded (Fuchs et al., 1989).
The aim is generally to increase the MDP by developing a Technique named β
to leverage the anticipated MDP benefit. Pn α| is sometimes assumed to be the state
distribution n after μ for convenience. The roll-outs are the performance of the test
device and are used for delivery by the MDP parameters. The association between
incoming parameters and the output parameters for each mixture of parameters is
between Sedlmair et altechniques. This implies that the relation is routinely changed,
the connection between the parametres, and the corresponding outputs is generated
and measures (Costa et al., 2020)
You need to check for errors to find MDP defects. In certain instances, MDPs
are known from Sedlmair et al. The test questions often claim that the roles of
compensation (R), transformation (P), regulation (T) or optimization (M) are
concerned (Fuchs et al., 1989).
MDPs have already already listed many topics covering such wide areas as RC
car monitoring, endangered species control and sports in real time. Although no big
MDP visualisation in these areas was previously provided, a range of works for
17
restricted MDP groups can be considered as visualisation (Ra et al., 2020). In one
method, a variety of works discuss discovery decisions. Broeksema et al. provide a
method for evaluating the expert system's proposals. Decisions shall be made by a
multidimensional communication analysis (MCA) in the form of Voronoi diagrams
(MDS). The Voronoi diagrams lack an intuitive double-dimensional system, but the
adjacent diagram reveals that the part of the assessment varies with various attributes.
MDP policies are also distinguished by classification of students (McGregor et al.,
2017). The main area of research is debugging classifiers, but no debugging
classification of MDPs is found in the published literature.
Groce et al. have also set goals for classified data points for review by the
customer. The end user chooses whether he or she accepts after a datapoint is
selected. Through changing the model stickers the client begs the classification (Fuchs
et al., 1989). It was an erroneous data-label that this debugging method shows that it
was the only mistake. However, there are no data marks in MDPs if the simulator, the
technique and the principal topic of analysis is primarily right. Kulesza et al. provide a
debugging method for email classification. The consumer provides input and
adaptation in an immersive bar map with a naive Bayes model. Kulesza et al.
preferred Bayes by the end users, while the interpretability of these classifiers became
too hard to grasp. Migut and Worring write a variety of knowledge graphics to
investigate the dichotomous range through a visual theoretical dashboard, as defined
by a classification computer (Van Ham & Van Wijk, 2004). The purpose of ensemble
visualisation is to provide a compact image with a specific and clear fact of several
mathematical models. The visualisation of the sample agreement eliminates ambiguity
about the impact of the prediction. In comparison, the instability of MDP visualisation
arises from shocks that the agent operates on from the world. The visualisation of the
ensemble means recognition, however MDP visualisation requires an overview of the
wide spectrum of future global policies (Prabhat & Khullar, 2016).
Vismon offers an interface to track trade in different management options for
fisheries in Alaska, a remarkable visualisation for the management of natural
resources (Shneiderman, Plaisant, & Hesse, 2013). In this scenario, the fisheries
manager filters 121 autonomous management assessments on the basis of two
separate management criteria. The manager cannot access implementation alternatives
until the strategy has been chosen. Simulation Steering is a subset of visualisation
where users can pick behaviours as they move (Y. Wu et al., 2010). The simulation
18
pathway for disease response evaluation reveals the results of persons in Afzal et al.
over time. Customers will adjust their option several times in the future to test the
response of mortality rates (Fuchs et al., 1989).
Our approach explains visualisation of the formal characteristics for a set of
subjects compared to current literary methods which aim to better visualise a given
problem area (Zockler et al., 1996). MDPVIS is accessible to wider classes of mdp by
retaining the state variables, raising the optimizer, like those with partial
observability, continuous operations and numerous agents, whereas our research
focutes on the key wording of the mdp. Partial MDPs (POMDP) observed (Deussen et
al., 2002). When tuning the Monte Carlo roll out of POMDPs, the simulator usually
eliminates unknown variables (Fuchs et al., 1989).
If the simulator is known, unknown factors are shown and checked, which
could be needed in order to better simulate the working circumstances of a policy.
This may take the form of two helicopters engaging in helicopter domain manoeuvres
(Prabhat&Khullar, 2016). MDPVIS will make multi-agent MDPs for any agent with
many variables and parameters, but the future would in particular benefit from a
multi-agent MDP view (Liu et al., 2013).. The theory is that the residual period would
remain before a computer device or generator works due to a lack of distribution and
development (Prabhat & Khullar, 2016). Driven by data diagnosis and prognoses, the
tacit probabilistic distributive methods, like Gaussian-Markov processes, are not
necessary, in comparison to models-based approaches (Shneiderman et al., 2013).
These stochastic or chaotic systems should not presuppose the like of deep schooling.
Machine training involves a range of training and high-performance computer
systems, such as cloud-based coverage. The computational workloads including Vast
Data Sets and Fog analysis are carried out at areas where enormous quantities of data
are collected and analysed rather than centralised cloud-based storage similar to cloud
computing. One of the key advantages of fog computing is to support clients by
providing large volumes of data with cutting-edge computers and cloud computing
centres in the near proximity of nearby physical devices and machines to prevent
transmission knoys. With fog calculations mounted next to the raw data source, fog
calculations greatly decrease latency.
Despite the lack of omnipresent sensors, demands of the market and scalable
high-performance computing technology, current supervision mechanisms and
prediction techniques are unable to gather vast amounts of real-time data or construct
19
broad predictive models. This is especially relevant for latency-responsive
applications (Liu et al., 2013). Wireless, cloud and deep learning sensors are used to
overcome the gap.
Publisher-Subscription is an asynchronous knowledge demonstration where
editors and backers exchange unfamiliar communications. Opportunities are named
news. Inside the display Pups the supporters are intrigued by an event for the broker
(Moraes et al. 2019). The distributor sends messages to the broker who will relay the
message to the requested supporter at the case. It can be used as a broker as it allows
the PupSup evidence. It is an open source memory storage store. Redis helps the
broker to distribute across several hubs (Li et al., 2020).
In the literature, several attempts are made to build frames to visualise details
on the site. Recently, the GAN method has been used for various restorative imaging
applications. Most recommend (Fried-Adar et al. 2018) utilising pictorial GAN to
establish an understanding of mark through segmentation, section to picture or clinical
cross system interpretation, from pictorial to pictorial. A complete organisation to
memorise the images in the retinal vessel region (Prim, Baer, et al. 2016) has been
prepared. Then, from the double-character tree to the modern retina picture, they
studied the interpreting. It is used to divide photographs from chest X-ray images of
the lung areas and hearts (James and Dashrath 2014).
Smart devices are an important part of development process control. In
addition, sensors or sensor networks that can prevent events and quantify signal are
required for the recovery of realtime data from the floor of the factory and for
monitoring the health conditions for equipment and operation. Wright et al. has
developed an accelerometer dependent surveillance method for the prediction of
cutting tool wear and surface surface finishes. For the vibrations of a high-speed steel
finishing instrument, a wireless sensor platform built on the IEEE 802.15.4
specification was used. It was shown that the wireless sensing device would calculate
cutting conditions such as the usage of the instrument on the frying machine(Kruse et
al., 1993). In order to track intelligent tool condition by utilising neural networks and
numerous sensors, Rangwala and Dornfeld suggested a computational system. In
order to quantify vibrations, an active transducers is fitted on the tool case. For
calculating cutting powers, a force dynamometer was used(Hristopulos &
Geosciences, 2015).
20
Experimental findings revealed that the frame-based monitoring device could
conduct sensor fusion and identify anomalies in the process. Li and Li also established
an acoustic pollution control device for detecting the onset of exhausted cracks. AE
transducers have been installed on the bearing housing to control the accelerated
release of concentrated stress energy(Prabhat & Khullar, 2016). For industrial motor
systems utilising wireless sensor networks, Luetal built an online and remote power
monitoring and error diagnostics framework. To test the engine shaft toque, an inline
torque transducer was used. The engine speed was calculated using an optical
encoder. In a specific industrial area, the monitoring device has been
demonstrated(Liu, Blasch, Chen, Shen, & Chen, 2013). Hou and Bergmann have
presented a system state control and defect diagnostics industrial wireless sensor
network. The computer state control device also implemented the common wireless
networking protocols (e.g. IEEE 802.15, IEEE 802.11, ZigBee, and WirelessHART).
A series of experiments on a single induction engine demonstrated the proposed
system(Cleland et al., 2018).
The aim of the process monitors is to analyse machine part health conditions
(e.g. rods and spindles), processes of manufacture (e.g. machining, joining), as well as
systems of output. Fault detection, separation and recognition are the subject of
diagnosis. The prediction is to estimate the period before a computer part or a
production device executes its expected role due to the distribution and propagation of
fault(Costa et al., 2020). The forecast period is listed as being of use in literature
(RUL). Many data-driven approaches have been established in the course of the last
two decades to diagnose and forecast manufactures. Data-driven diagnostic
approaches can explicitly be categorised into signal processing, artificial intelligence,
pattern analysis, and mathematical learning. Auto-reguressive (AR) models, two-
linear process, multivariate adaptive regression, neural networks, fluid set theory, and
deep learning provide the conventional data-driven pronostic methods(Deussen,
Colditz, Stamminger, & Drettakis, 2002).
Data-based diagnostics and prognosis need not require a detailed
comprehension of the mechanics behind processes of machining and full experience
of device conducts, unlike physics-based approaches. Contrary to model oriented
approaches, data-driven diagnostic and prognostic distributions, such as Gausian-
Markov systems, are not believed to be probabilistic. Compared with statistical
approaches, methods like machine learning are not AI dependent, since machine
21
learning involves vast quantities of training data sets and high-performance
computing systems such as cloud computing, but they do not embrace certain
stochastic or unpredictable processes like Viennese processes and gamma
processing(Carlis & Konstan, 1998).
There is a limited capacity to gather and maintain vast quantities of data in
clustered environments and a limited capacity for computer interpretation of these
data in real-time, both for current process control mechanisms and prognostic
approaches. The grid computing idea was suggested by Foster and Carl Kesselman in
1999. A network grid relates to the system of hardware and software that offers
secure, consistent, overarching and accessible access to high-end computing
capabilities. The cloud computing principle is focused on grid computation(Ra, Ab, &
Kc, 2020).
NIST notes that cloud computing is a "model for the ability to quickly provide
and distribute on-demand network access to the shared pool of configurable
computing resources (such as networks, servers, storage, applications and services)
with a minimal management or service provider interaction" Fog computing was
introduced by Cisco to expand cloud computer infrastructure to apply high-
performance computing to the edge of a company network(Zockler, Stalling, & Hege,
1996). Fog Computing, defined as Edge Computing (e.g. Wireless Router and Large
Area Network Connectivity Device) and Cloud Computing Knowledge Centres, is a
computing paradigm that delivers high-performance computational power, data
storage, and network facilities across intermediate devices.
Cloud storage requires huge data transfer to cloud data centres, with
substantial overhead efficiency. In comparison to cloud computing, analytical task
loads are taken out at places where big quantities of data are gathered and processed
in lieu of consolidated cloud storage, such as preparing with massive data sets and
visualising data processing. One of the main advantages of fog computation is that its
consumers can stop bringing the computing nodes near to local physically active
items or computers and operating software directly on large data sites between edge
devices and cloud computing data centres(Van Ham & Van Wijk, 2004). Because fog
is near to the raw data source, fog computing will minimise latency dramatically. For
latency-sensitive systems, this is especially relevant. Cisco applied fog-computing to
intelligent grids, utilising edge infrastructure, such as smart metres, to enable real-
time applications and localisation-sensitive services(Ra et al., 2020).
22
Fog computing is also a crucial aspect of the productive solution to securing
cloud-based production networks. Summary, this segment discusses how the health
conditions of machines can be tracked using sensors as well as how predictive
modelling for prognosis can be created(Shneiderman, Plaisant, & Hesse, 2013).
However, current surveillance technologies and prognostic methods cannot capture
vast amounts of real-time data or create large-scale predictive models due to a lack of
all-round sensor grids, production market requirements, and high efficiency scalable
computing systems. This paper combines wireless sensing, cloud computing and deep
learning to resolve the distance(McGregor et al., 2017).
23
networks and have low BPNN, and
accuracy CpNN are trained
and tested on the
same chest X-ray
database
24
and MNIST
dataset as source
task
independently
The chapter provides literature work in detail. The chapter explain problem in existing
literature related to interactive visualization base image detection in detail. The
25
literature show the there is still need to improve this area and techniques which
predict medical images data and give accuracy above 90%. A large numbers
researcher works on it but there are still very hot problems in this area.
26
Chapter 3
RESEARCH METHODOLOGY
3.1 Introduction
This chapter explain the overall research methodology of proposed scheme. The
chapter also explain the frame work used for this research and also give the detail of
tools used for this. For this analysis, the data set is very broad, so every data set is
described in detail. The comparison study was carried out by comparing two separate
strategies for detecting leukemia. The methods were a method of genomic
sequencing, a model of binary classification and a model of multi-class classification,
which was a method of image processing. The methods had different values for input.
A Convolutionary Neural Network (CNN) was used by both of them as a network
architecture, however. Using 3-way cross validation, they also split their datasets.
3.2 Research Framework
Leukemia is a type of cancer that can be a lethal illness, which needs an accurate and
early diagnosis to rehabilitate and cure it. Normal approaches have been turned into
automated machine tools for symptom examination, diagnosis, and prediction. A
comparative analysis was carried out in this work by comparing two separate methods
of detection of leukemia. The methods were a method of genomic sequencing, a
model of binary classification and a model of multi-class classification, which was a
method of image processing. The methods had different values for input. A
Coevolutionary Neural Network (CNN) was used by both of them as a network
architecture, however. Using 3-way cross validation, they also split their datasets. The
learning curves, confusion matrix, and classification report were the assessment tools
for assessing the outcomes. The findings showed that the genome model had better
performance and had some values that were correctly predicted with 98 percent
overall accuracy. This value was compared to the findings of the image processing
system with a value of 81 percent overall accuracy. The size of the various data sets
may be the source of the algorithms' different test results.
28
Figure 2: Blood seamer Images
29
3.5 Chapter Summary
Leukemia detection using CNN as an architecture network was interesting and
challenging because of the subject area was a complex and complicated to implement.
Still, it also incorporates intriguing aspects such as genetic. Both models used similar
hyper-parameter and neural networks, with different classification model was an
adequate ground step for comparative analysis. The models managed to score decent
results, which were presented. The data set and complete research frame work is
explained in the chapter.
30
Chapter 4
PROPOSED METHOD
1 Introduction
The chapter explain the proposed scheme in detail. It explain with the help of flow
diagrams In the area of computer vision, deep learning has recently played an
important role. The elimination of human judgment in the diagnosis of diseases is one
of its applications. In particular, the diagnosis of brain tumors needs high precision,
where minute judgment errors can result in a tragedy. For this reason, for medical
purposes, brain tumor segmentation is an important challenge. There are many
methods currently available for tumor segmentation, but they all lack high precision.
Here we present a solution by using deep learning for brain tumor segmentation. We
studied various angles of brain MR images in this work and applied various
segmentation networks. By comparing the results with a single network, the effect of
using different networks for segmentation of MR images is evaluated. Experimental
network tests indicate that a dice score of 0.73 is obtained for a single network and
0.79 is achieved for multiple networks. Magnetic Resonance Imaging (MRI) is a
medical imaging technique that is commonly used in clinical practice for the
diagnosis and treatment of brain tumors. Three distinct directions are taken from the
MR pictures. These views are called sagittal, axial and coronal. These three types of
brain MR images are shown in Figure 1. Brain Tumor segmentation techniques are a
critical component in tumor detection. Using machine learning techniques that learn
the pattern of brain tumor is useful because manual segmentation is time-consume.
4.1 Proposed Method
The section explain the proposed method of scheme in detail.
The CNN review in this study is shown in the figure. The CNN is divided into two
parts: the first part consists of two convolutionary layers, taking into account 16 filters
with dimensions of 1 × 5. Between the convolutional ones, we included an
intermediate Maxpooling layer. This section aims to derive from the inputs the key
characteristics. For classification, the second component includes three completely
32
linked layers. Dropout layers are used after convolutional and fully linked layers to
prevent over fitting. The number of weights deactivated was 20% . This architecture
was influenced by the work of Khatamino et al. where a simplification of the
AlexNetCNN was proposed by the authors. This simplification was appropriate
because the CNN parameters were trained using a smaller dataset.
33
the type of the tumor. These 3064 images belong to 233 patients. The dataset includes
708 Meningioma’s, 1426 Gliomas, and 930 Pituitary tumors, which are publicly
available in: (http://dx.doi.org/10.6084/m9.figshare.1512427). The network training
process and details are mentioned in the followings. For the single LinkNet network,
we used 2100 of images for network training that 20% of these images are considered
as validation and the rest of the data is used for the test purpose. Also for the training
of the three LinkNet networks, we separate all images into three groups. Each group
contains one type of MR image based on the image view. In each group, about 900
images are used for the training procedure and about 200 images are used as test
images. Our network uses binary cross-entropy as the loss function and the network is
tuned using this parameter.
34
01/01/2007 to allow for 10 years of follow-up for all individuals within the cohort.
We divide the EHR into: I the observation window (01/01/2000 to 12/31/2006; 7
years) for each person and (ii) the forecast window (01/01/2007 to 12/31/2016; 10
years). In the 7-year observation window (2000-2006), we obtained EHR data to train
a classifier to decide if the person will have a CVD event in the 2007-2016 10-year
prediction window.
35
standardization layer, and a ReLU layer for every last one of these squares. As seen in
Figure 2, it could have greater pooling and a dropout layer. In order to adjust the
information shape to a one-dimensional information vector, the yield of the material
extraction portion is then transferred to the smoothed layer, which is the accurately
used configuration for the thick layer grouping. The thick layer is the regular, deeply
connected layer of neural organization. It is the most well-known and commonly used
layer in which each knowledge is connected to each yield[34]. We use three thick
layers and four layers of dropouts here. The final yield is obtained from a thick layer
with sigmoid initiation work grouping the yield image to Pneumonia (addressed by
blue bolt in the figure) or ordinary (addressed by red bolt). Table 2 shows the
proposed CNN Model Engineering and Figure 3 shows the principal potential of the
code. 38,320,049 is the absolute number of model boundaries: the teachable
boundaries amount to 38,319,889 and the non-teachable boundaries only amount to
160.
36
Figure 8: Blood cancer detection using CNN
Chapter Summary
Another methodology for CNN to naturally section the most well-known types of
mind tumor, for example glioma, meningioma and pituitary, has been dispatched.
This technique does exclude ventures for preprocessing. The discoveries demonstrate
that the detachment of pictures dependent on points expands the exactness of division.
The best score acquired by Dice is 0.79. This generally high score in sagittal view
pictures was gotten from division of tumors. Photographs of Sagittal don't contain
particulars of different organs and are more noticeable than different pictures of the
tumor. In our tests, the least dice score was 0.71, which incorporates less data
contrasted with different pictures in hub see. It is foreseen that better characterization
of tumor pixels will be gotten by performing preprocessing on this gathering of
pictures and the dice score would increment. As a simple and valuable instrument for
specialists to portion mind tumors in MR pictures, our proposed strategy can be
actualized.
37
Chapter 5
1 “Introduction
A new method for CNN to automatically segment the most common forms of brain
tumor, i.e., glioma, meningioma, and pituitary, was presented in this chapter. This
method does not include steps for preprocessing. The findings indicate that the
separation of images based on angles increases the precision of segmentation. The
best score obtained by Dice is 0.79. This relatively high score in sagittal view images
was obtained from segmentation of tumors. Photos of Sagittal do not contain specifics
of other organs and are more prominent than other images of the tumor. In our tests,
the lowest dice score was 0.71, which is related to the images from the head axial
view.”
The below diagram show the results of Parkinson diseases using CCN
39
Figure 12: Lung cancer detection using CNN
40
Figure 14: Pneumonia detection using chest X-ray using CNN
41
Brain tumor detection using CNN
42
Figure 18: Brain tumor detection using CNN
43
Blood cancer / acute lymphoblastic leukemia detection using CNN
The below results show the detection of blood cancer
Figure 20: Blood cancer / acute lymphoblastic leukemia detection using CNN
Figure 21: Blood cancer / acute lymphoblastic leukemia detection using CNN
44
Figure 22: Blood cancer / acute lymphoblastic leukemia detection using CNN
Chapter Summary
“In the area of computer vision, deep learning has recently played an important role.
The elimination of human judgment in the diagnosis of diseases is one of its
applications. In particular, the diagnosis of brain tumors needs high precision, where
minute judgment errors can result in a tragedy. For this reason, for medical purposes,
brain tumor segmentation is an important challenge. Currently many methods exist for
tumor segmentation but they all lack high precision. Here we present a solution by
using deep learning for brain tumor segmentation. We studied various angles of brain
MR images in this work and applied various segmentation networks. By comparing
the results with a single network, the effect of using different networks for
segmentation of MR images is evaluated. Experimental network tests indicate that the
Dice score is 0.73 for a single network and 0.79 for multiple networks.”
45
Chapter 6
1 Introduction
This chapter explain the future direction in this work. Advanced methods can be used
to help patients detect terminal disorders such as leukemia, which is a fatal disorder
and common cancer type amongst children. Leukemia is a form of cancer that begins
in blood cells and the bone marrows, where it grows new immature blood cells when
the body does not need them. White blood count(WBC) is a routine blood test usually
done manually, to search for leukemia cells and can be automated by applying
machine learning techniques such as CNN. It is a simple and faster way to perform a
test and detect abnormality in the blood. Other practices are genomic sequencing to
detect the abnormal markers in coding and non-coding regions along with DNA
sequences. This is used to predict or detect cancer from using biomarker.
47
REFERENCES
Agus, Marco et al. 2009. “An Interactive 3D Medical Visualization System Based on
a Light Field Display.” Visual Computer 25(9): 883–93.
Bauer, Stefan, Roland Wiest, Lutz P. Nolte, and Mauricio Reyes. 2013. “A
Survey of MRI-Based Medical Image Analysis for Brain Tumor Studies.” Physics in
Medicine and Biology 58(13): 1–44.
Cobzas, Dana et al. 2007. “3D Variational Brain Tumor Segmentation Using a High
Dimensional Feature Set.” Proceedings of the IEEE International Conference
on Computer Vision: 0–7.
Frid-Adar, Maayan et al. 2018. “GAN-Based Synthetic Medical Image Augmentation
for Increased CNN Performance in Liver Lesion Classification.”
Neurocomputing 321: 321–31.
Gillies, Robert J., Paul E. Kinahan, and Hedvig Hricak. 2016. “Radiomics: Images
Are More than Pictures, They Are Data.” Radiology 278(2): 563–77.
Gotz, David, and Harry Stavropoulos. 2014. “DecisionFlow: Visual Analytics for
High-Dimensional Temporal Event Sequence Data.” IEEE Transactions on
Visualization and Computer Graphics 20(12): 1783–92.
Haskins, Grant, Uwe Kruger, and Pingkun Yan. 2020. “Deep Learning in Medical
Image Registration: A Survey.” Machine Vision and Applications 31(1): 1–
30.
Havaei, Mohammad et al. 2017. “Brain Tumor Segmentation with Deep Neural
Networks.” Medical Image Analysis 35: 18–31.
Holzinger, Andreas, and Igor Jurisica. 2014. “Knowledge Discovery and Data Mining
in Biomedical Informatics: The Future Is in Integrative, Interactive Machine
Learning Solutions.” Lecture Notes in Computer Science (including subseries Lecture
Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) 8401: 1–
18.
James, Alex Pappachen, and Belur V. Dasarathy. 2014. “Medical Image Fusion: A
Survey of the State of the Art.” Information Fusion 19(1): 4–19.
Li, Wei, Kun Yu, Chaolu Feng, and Dazhe Zhao. 2020. “SP-MIOV: A Novel
Framework of Shadow Proxy Based Medical Image Online Visualization in
Computing and Storage Resource Restrained Environments.” Future
Generation Computer Systems 105: 318–30.
48
https://doi.org/10.1016/j.future.2019.12.009.
Liu, S et al. 2015. “Eurographics Conference on Visualization (EuroVis) (2015)
Visualizing High-Dimensional Data: Advances in the Past Decade.”
www.sci.utah.edu/.
Moraes, Thiago, Paulo Amorim, Jorge Silva, and Helio Pedrini. 2019. “Web-Based
Interactive Visualization of Medical Images in a Distributed System.”
VISIGRAPP 2019 - Proceedings of the 14th International Joint Conference
on Computer Vision, Imaging and Computer Graphics Theory and
Applications 1(Visigrapp): 346–53.
reim, Bernhard, Alexandra Baer, et al. 2016. “A Survey of Perceptually Motivated 3D
Visualization of Medical Image Data.” Computer Graphics Forum 35(3): 501–25.
Preim, Bernhard, Paul Klemm, et al. 2016. “Visual Analytics of Image-
Centric Cohort Studies in Epidemiology.” Mathematics and Visualization:
221–48.
Scholl, Ingrid, Til Aach, Thomas M. Deserno, and Torsten Kuhlen. 2011. “
Challenges of Medical Image Processing.” Computer Science - Research and
Development 26(1–2): 5–13.
Stalling, Detlev, Malte Westerhoff, and Hans Christian Hege. 2005. “Amira: A
Highly Interactive System for Visual Data Analysis.” Visualization
Handbook 1: 749–67.
T.S. Yoo, M. J. Ackerman, W. E. Lorensen, W. Schroeder, V. Chalana, S. Aylward,
D. Metaxas, R. Whitaker. 2002. “Engineering and Algorithm Design for an
Image Processing API: A Technical Report on ITK - The Insight Toolkit. In Proc. of
Medicine Meets Virtual Reality,.” J. Westwood, ed., IOS Press Amsterdam
85: 586–92.
Zhang, Shaoting, and Dimitris Metaxas. 2016. “Large-Scale Medical Image
Analytics: Recent Methodologies, Applications and Future Directions.”
Medical Image Analysis 33: 98–101.
http://dx.doi.org/10.1016/j.media.2016.06.010.
49