Professional Documents
Culture Documents
Information Technology
Submitted by:
Introduction
Earlier when no major storage device was present and there were only computers
(hard disk) on which data could be stored, it was relatively easy for anyone to
analyze or abstract information from this data. But with advancements in technology,
there is an exponential increase in storage data ranging from removable sources
such as USB sticks and SD cards to dispersed source such as Cloud Services.
Therefore, big data can be described as the huge amount of structured or
unstructured data stored in various storage devices usually in the size of zettabytes
or petabytes that can be analyses to derive datasets providing patterns and
interdependency of data.
Big data has been particularly challenging to Digital forensics where it has become
extremely difficult to analyses data on three parameters of big data — Volume,
Variety and Velocity. It is of no doubt how difficult it can be to analyse data of huge
size that is being processed at a very high speed and having no idea from which
device or source. Furthermore, to find out complex interdependence and
arrangement from this data would be even more difficult leading to an increase in
cyber-crimes and a challenge for digital forensic investigators.
Data science can be incorporated in digital forensics to counter big data as it is
nearly impossible for a human to analyze such huge amount of data and realize the
complex patterns across the content. For instance, data can be collected from the
previously solved cases, evidences and databases. This data can then be used to
derive suitable algorithms set by the rules analyzed on the given data. But in the
initial test stages, the final decision should be made manually to avoid wrong
decisions.
Aim of Micro Project
Literature Review
Proposed Methodology