You are on page 1of 22

TERM PAPER

on

Title: Impact of big data on business decision-making

Submitted to

Dr. Purushottam Sharma

Name of the Student: Rishi Kumar

Enrolment No.-A2305321054

1 | Page
DECLARATION
I, Rishi Kumar, student of B.Tech (2-T-1) hereby declare that the project titled “Impact of big
data on business decision-making” Which is submitted by me to the Department of
Information Technology, Amity School of Engineering Technology, Noida, Amity University
Uttar Pradesh, in partial fulfillment of the requirement for the award of the degree of Bachelor of
Technology in Information Technology, has not been previously formed the basis for the award
of any degree, diploma or other similar title or recognition.

The author attests that permission has been obtained for the use of any copyrighted material
appearing in the Dissertation / Project report other than brief excerpts requiring only proper
acknowledgment in scholarly writing and all such use is acknowledged.

Place: - Noida

(Student Signature)
Student Name: Rishi Kumar
Enrollment No: A2305321054
Semester: 2

2 | Page
CERTIFICATE
Based on the report submitted by Rishi Kumar, student of B.Tech (Information Technology)
hereby certify that report “Impact of big data on business decision-making” which is
submitted to the department of INFORMATION TECHNOLOGY, Amity School of Engineering
and Technology, Amity University Uttar Pradesh in partial fulfillment of the requirement for the
award of the degree of Bachelor of Technology(IT) is an original contribution with existing
knowledge and faithful record of work carried out by her under my guidance and supervision.

To the best of my knowledge this work has not been submitted in part or full for any Degree or
Diploma to this university or elsewhere.

Noida

Date-

(Signature of the Guide)

Dr. Purushottam Sharma

Department of Information Technology

ASET, Noida

3 | Page
ACKNOWLEDGEMENT
The satisfaction that accompanies the successful completion of any task would be incomplete
without the mention of people whose ceaseless cooperation made it possible, whose constant
guidance and encouragement crown all efforts with success. I would like to thank Amity
University for allowing me to undertake this project. I would like to thank my faculty guide Dr.
Dr. Purushottam who is the biggest driving force behind my successful completion of the
project. He has always been there to solve any query of mine and also guided me in the right
direction regarding the project. Without his help and inspiration, I would not have been able to
complete the project.

RISHI KUMAR

A2305321054

4 | Page
Table of Contents
Acknowledgment 4

Table of Contents 5

List of Figures 6

Abstract 7

Chapter 1. INTRODUCTION 8

1.1 Big Data………………………………………………………………………………...…..........8

1.2 Types of Data………………………………………………………………………………...…...8

1.3 Characteristics of Big Data……………………………………………………………………..8

1.4 Classification of analytics of Big Data………………………………………………………....9

Chapter 2. STORAGE ……………….………………………………………………………………...11

2.1 Storage Collection…………………………………………………………………………….11

2.2 Data Transmission……………………………………………………………………………12

Chapter 3. IMPORTANCE IN BUISNESS………………………………………………..………14

Chapter 4. IMPACT OF BIG DATA……………….…………………………………………….…17

Chapter 5. ASSET MANAGEMENT……………….……………………………………………….21

5.1 Data science and aquistion……………….……………………………………………...……21

5.2 Insightful Advanced Analytics……………….…………………………………………….....21

5.3 Revenue Predictions…………………………………………………………………………..22

5.4 Future…………………………………………………………………………………………22

References ……………………………………………………………………………………………...23

5 | Page
List of Figures

1. Figure 1-3 v's of big data

2. Figure 2-Data transmission wih RMDT

3. Figure 3- Comparison between RMDT and other protocol

4. Figure 4- Representation of data utilization in terms of data

6 | Page
Abstract
Big Data Analytics (BDA) is an emerging phenomenon with the reported potential to transform
how firms manage and enhance high value businesses performance. The purpose of our study is
to investigate the impact of BDA on operations management in the manufacturing sector, which
is an acknowledged infrequently researched context. Using an interpretive qualitative approach,
this empirical study leverages a comparative case study of three manufacturing companies with
varying levels of BDA usage (experimental, moderate and heavy). The information technology
(IT) business value literature and a resource-based view informed the development of our
research propositions and the conceptual framework that illuminated the relationships between
BDA capability and organizational readiness and design. Our findings indicate that BDA
capability (in terms of data sourcing, access, integration, and delivery, analytical capabilities, and
people’s expertise) along with organizational readiness and design factors (such as BDA
strategy, top management support, financial resources, and employee engagement) facilitated
better utilization of BDA in manufacturing decision making, and thus enhanced high value
business performance. Our results also highlight important managerial implications related to the
impact of BDA on empowerment of employees, and how BDA can be integrated into
organizations to augment rather than replace management capabilities. Our research will be of
benefit to academics and practitioners in further aiding our understanding of BDA utilization in
transforming operations and production management. It adds to the body of limited empirically
based knowledge by highlighting the real business value resulting from applying BDA in
manufacturing firms and thus encouraging beneficial economic societal changes.

Chapter-I
1.1BIG DATA refers to complex and large data sets that have to be processed and analyzed to
uncover valuable information that can increases businesses profitability and for expanding
organization.

7 | Page
FUNCTIONS:

⮚ It grows exponentially with time so that it cannot be processed or analyzed using


conventional data processing techniques.

⮚ It includes data mining, data analysis and data sharing and, data visualization.

1.2 TYPES OF BIG DATA


● Structured refers to be highly organized that can readily and seamlessly stored and
accessed from a database by simply search engine algorithm. It can be processed, stored,
retrieved in a fixed format.

● Unstructured refers to the data that lacks any specific form or structure which makes
it very difficult and time consuming to analyze it. Example: - Email, etc.

● Semi-Structured is the third type of data which contains the format of both structured
and semi-structured.

1.3 CHARACTERISTICS OF BIG DATA (3 V)


a) Variety of big data refers to structured, unstructured, and semi structured data
which is gathered from multiple sources. While in the past, data could only be
collected from spreadsheets and databases but today it is now available in array forms
such as emails, PDFs, photos, videos, audios, etc. It is an important characteristic of
big data.
b) Velocity comprises the rate of changes, linking of incoming data sets at varying
speeds, and activity bursts.
c) Volume of big data indicates generation of data on a daily basis from various
sources like social media platforms, business processes, machines, networks, human
interactions. Such a large amount of data is stored in a data warehouse.

8 | Page
Figure1. 3 v of Big Data

1.4 CLASSIFICATION OF ANALYTICS IN BIG DATA


⮚ Descriptive Analytics is a statistical method to search and summarize historical data in
FiForder to identify patterns or meanings.
⮚ Data aggregations and data mining are two techniques for descripting analytics to
discover historical data. In this data is first gathered by mining and then sorted by data
aggregation to make the datasets more manageable by analysts.
Data mining describes the nest step of analysis and involves searching of data to identifying
patterns and meaning. Identified patterns are analyzed to discover the specific ways that
learners interacted within a learning environment.

Advantages: -
✔ Quickly and easily report on ROI(Return On Investment) by showing performance of
business.
✔ Identify gaps and performance issues early.

9 | Page
✔ Identify specific learners who require additional support on the numbers of student or
employee.
✔ Analyze the value and impact of resources.
⮚ Prescriptive Analytics is a statistical method for generating recommendations and
making decisions based on algorithmic models.
⮚ Predictive Analytics is a statistical method that utilizes algorithms and machine
learning to identify trends and predict future behaviors.
The software of this analytics has moved beyond the realm of statisticians and is become more
affordable and accessible for different market and industries, including field of learning and
development.
For Learning Management System, it is found to be incorporated.

Descriptive vs Predictive vs Prescriptive Analytics


⮚ Descriptive Analytics is focused solely on Historical data.
⮚ Predictive Analytics is using this historical data to develop statistical models that will
then forecast about future possibilities
⮚ Prescriptive Analytics takes Predictive Analytics a step further and takes the possible
forecasted outcomes and predicts consequences of these outcomes.

Chapter-2
2.1 STORAGE COLLECTION

10 | Page
Storage collection of huge data refers to the storage and management of large-scale datasets
through achieving reliability and availability of data accessing. Data device is employed to store,
manage, and analyze data with structured RDBMs. With the sharp growth of data, data device
is becoming increasingly more important, and far of Internet companies pursuing high capacity
of storage for being more competitive. Therefore, Compiling need for research on data storage
system for huge data. Various storage systems emerge to satisfy the strain of massive data.
Massive storage technologies has been classified as Direct Attached Storage (DAS) while
network storage has been classified into Network Attached Storage (NAS) and enclosure
Network (SAN). In DAS, various hard disks are directly connected with servers, and data
management is server centric, such storage devices are peripheral equipment, each takes a
particular amount of I/O resource and is managed by a non-public application software. For this,
DAS is barely suitable to interconnect servers with a little scale. Thus, DAS is incredibly utilized
in personal computers and small-sized servers Network storage which is to utilize network to
supply users a union interface for data accessing and sharing. Network storage equipment
includes special data exchange equipment, disk array, tap library, and other storage media,
likewise as special storage software. it's characterized with strong expandability. NAS which is
an device equipment of a network. it's directly connected to a network through a hub or switch
through TCP/IP protocols. In NAS, data is transmitted within the range of files. As Compared to
DAS, the I/O at a NAS server is reduced since the server accesses a tool indirectly through a
network. NAS is network-oriented, While SAN is incredibly designed for data storage with a
scalable and bandwidth intensive network, In SAN, data storage management is comparatively
independent within a storage local area network, where multipath based data switching among
any internal nodes which is employed to understand a maximum degree of knowledge sharing
and data management. From the organization of an information storage system, DAS, NAS, and
SAN can all be divided into three parts:
(i) Disc array is that the inspiration of a storage system and also the essential guarantee for data
storage;
(ii) Connection and network subsystems, which give connection among one or more-disc arrays
and servers
(iii) Storage management software, which handles data sharing, disaster recovery, and other
storage management tasks of multiple servers.

11 | Page
Figure2. Data transmission with RMDT

2.2 TRANMISSION OF BIG DATA


In 1969, when the net has been developed the primary data transmission between the 2
computers with a distance of 600 km was held. this could be often the primary step to the
information transmission. The Wide Area Network was developed at that point and also the
distances between communication nodes has been increases correspondingly.
The two well-known data transmission types: TCP and UDP. The difference between them is
barely of reliability.  
TCP could even be a reliable transport protocol that's widely used for data transport in IP
networks and sub-networks. it's suitable for the transmission of information streams with
guaranteed integrity, reliability, and checking of delivery. Since it absolutely was developed 40
years ago, this protocol has significant drawbacks.  Data at that point were transmitted by means
of TCP via local networks where:
* RTT was up to twenty MS;
* capacity was ones and 10s of megabits per second.
The new ways for data transition to consume existing network resources more effectively is
being developed by FILA group whether or not the standard of legacy infrastructure is poor.
there's a thought to insert the algorithms for available bandwidth analysis within the
communicating during reliable data transmission. Also, the transport protocol with this algorithm
will have quick adaptation to changing of bandwidth in an exceedingly connection. This group
has been developed another protocol, Reliable Multi-Destination Transport (RMDT) is
additionally a point-to-multipoint data transport protocol, which provides reliable multi-gigabit
Big Data distribution across the planet. It works in any IP-based network and can handle high
packet losses further as high RTTs and jitter. 
12 | Page
RMDT protocol features include:
* Multiple recipients within one session
* Reliability on the complete rate over WAN
* End-to-end 10G communication (1×10)
* Efficient system resources consumption
* Legacy IP infrastructure
* Suitable for streaming
RMDT protocol are often employed in WAN networks so on assure maximum bandwidth
utilization. it's resistant against high losses, RTT, and delay jitter. 

Figure3. Comparison between RMDT and other protocol

Chapter-3

13 | Page
3.1IMPORTANCE OF BIG DATA IN BUISNESS
FUNDAMENTALS
Big knowledge has nice vital in company yet as businesses. The importance of massive
knowledge is depended upon however company utilizes its collected knowledge that provides the
potential to its growth. Mainly firms yet as organization uses knowledge to enable:

✔ Price Saving: The tools of massive knowledge like Hadoop and Cloud based mostly
Analytics bring price advantage to business once great amount if knowledge to be keep.
✔ Time Reductions: The high-speed tools like Hadoop and in-memory analytics will
simply establish new sources of knowledge of information that helps businesses in
analyzing data directly and creating fast choices supported the training.
✔ Perceive the market conditions. the massive knowledge has nice importance in
determinative current market state of affairs. for instance, by analyzing customers’
buying behaviors, an organization will conclude the trending and therefore the most
marketing product that helps a good in going sooner than its competitors.
✔ Management on-line reputation: Its tools will do sentiment analysis as feedback
regarding United Nations agency is voice communication regarding your company.
✔ Boost client Acquisition and Retention: The client is that the most vital quality
of any business. today to beat peers and going success, it's vital to urge public review and
care of its sentiments. there's no single business which will claim success while not 1st
having to determine a solid client base. the utilization of massive knowledge permits
businesses to look at varied client connected patterns and trends that is very important to
trigger loyalty.
✔ Solve Advertisers drawback and supply promoting Insights: Its analytics
will facilitate in dynamical all business operations which has the flexibility to match
client expectation and dynamical company’s merchandise.
✔ Driver of Innovations and merchandise Development: Another large
advantage of massive knowledge is that the ability to assist firms initiate and redevelop
their products.

14 | Page
Figure 4. Representation of data utilization in terms of data

Big data and Business Intelligence measure two technologies helps plenty in firms in decision-
making method, each have nice variations. They take issue in an exceedingly approach that they
work the maximum amount as within the sort of knowledge they analyze. ancient bismuth
methodology is predicated on the principle of grouping all business knowledge into a central
server. Typically, this knowledge is analyzed in offline mode, when storing the data in
Surroundings referred to as knowledge Warehouse. The data is structured in an exceedingly
standard electronic information service with an extra set of indexes and sorts of access to the
tables (multidimensional cubes). an enormous knowledge answer differs in several aspects to
bismuth to use:
* In an exceedingly huge knowledge surroundings, data is kept on a distributed classification
system, instead of on a central server that is far safer and a lot of versatile houses.
*Huge knowledge solutions carry the process functions to the information, instead of the
information to the functions that helps in handling larger amounts of knowledge in an
exceedingly a lot of agile approach.
* Huge knowledge will analyze knowledge in numerous formats, each structured and
unstructured. the quantity of unstructured knowledge (those not keep in an exceedingly
ancient knowledgebase) is growing at levels abundant beyond the structured data. huge
knowledge solutions it to resolve them by permitting a worldwide analysis of assorted
sources of knowledge.

15 | Page
* Knowledge processed by huge knowledge solutions are often historical or come back from
period sources. Thus, firms will create choices And improve their business in an agile and
economical approach.
* Huge knowledge technology uses parallel mass process (MPP) ideas, that improves the
speed of study. By this, several directions square measure dead at the same time. Since, the
varied jobs square measure divided into many parallel execution elements, at the top the
results square measure reunited and conferred. Big knowledge has become the fact of doing
business yet as organization these days. If this knowledge is managed well, it will result in
powerful insights and quality deciding. huge data analytics is that the method of examining
giant knowledge sets containing a spread of information sorts * to find some information in
databases, * to spot attention-grabbing patterns and establish relationships to resolve issues,
market trends, client preferences, and different helpful data. Companies and businesses that
implement huge knowledge Analytics have many business advantages. firms implement huge
knowledge Analytics as a result of they require to form a lot of sophisticated business
choices. {a knowledge an}. Data Warehouse (DW) could be a assortment of company
information and knowledge derived from operational systems and external data sources. data
warehouse is meant to support business choices by permitting data consolidation, analysis
and coverage at totally different mixture levels. knowledge is kept into the information
Warehouse through the processes of extraction, transformation and loading.

Chapter-4
4.1 IMPACT OF BIG DATA ON ORGANIZATION AND
INDIVIDUALS

16 | Page
The use of massive information will offer businesses attention-grabbing, and generally stunning,
insights into our lives. Data mining play roles in several branches like banking, exchange,
promoting management, retail sales, signal process, insurance, telecommunication, electronic
commerce, health, medicine, biology, genetics, industry, construction, education, intelligence,
science, and engineering. numerous businesses apply data processing to essential business
processes to realize competitive advantage and facilitate businesses grow. In this, data processing
applications within the main sectors covering the mentioned sectors has been represented as:
Education Sector: it's been discovered that data processing has been employed in several studies
within the education sector including:

● Deciding the standing of students’ pass and fail,


● factors poignant the success of the scholars enrolling at the university,
● making the preference of university department,
● deciding the factors that influence the preference order of latest registered students,
● selecting a profession in step with the demographic and private characteristics,
● preventing students from failing and deciding the factors that have an effect on success,
● verify the relationships between the kind of college during which students graduate and
their university departments evaluating the study activities of distance education
students,
● deciding the profiles and preferences of scholars getting into the university entrance
examination,
● deciding the connection between educational success and participation in extracurricular
activities of university students,
● deciding the relationships between the socio-economic level of scholars and also the
level of educational learning, and · deciding whether or not there's a relationship
between student entry scores and faculty accomplishment. These usage areas in
academic sector facilitate academics to manage their categories, to know their students'
learning, and to produce active feedback to learners.

Telecommunication Sector: data processing are often employed in the


telecommunication sector

● to predict mobile user movements within the communications sector,


● to work out the long run movements of mobile users, · to observe frauds,
● to scale back a lot of human-based analysis,
● to work out the factors that influence customers to decision a lot of at bound times,
● to work out user templates for social network usage,
● to spot new prospects victimization demographic information,
● to spot the characteristics of shoppers WHO want special action as suspension or
deactivation, and
● to stop client loss. so as to stop client loss, telecommunication organizations will
eliminate this downside by resorting to ways for developing ways, low price and
effective campaigns.

Health Sector: Data processing techniques and application tools are a lot of valuable
for health sector. Data processing is employed within the health sector:

17 | Page
● to diagnose the sickness,
● to work out the treatment technique to be applied to the sickness,
● to estimate the resource use and patient numbers in hospitals,
● to line the success of treatment ways applied within the hospital,
● to classify the patient information in step with factors like age, gender, race and
treatment,
● to work out the bad factors in surgeries,
● to stop corruption in hospital expenditures. Public Sector: data processing is
commonly accustomed predict public safety and security issues within the public
sector. data processing techniques supply open opportunities for the general
public sector to optimize selections. These selections are supported general
trends extracted from past expertise and historical information. with the
exception of that, it's necessary: · to work out the tax connected corruption,
● to predict the impact of changes within the legal system on the budget,
● to work out waste and stop injury caused by waste, estimate population,
● to forecast the weather,
● verify new job opportunities,
● to live performance of staff, manage the business processes,
● to classify public expenditures, arrange the proper use of resources,
● to forecast the long run of public investment, analyze the information in defense
trade, and
● to work out that offenders are doubtless to commit crimes in terms of safety. One
of the foremost vital examples during this space is Government application in
Turkey. quicker feedback is being received as data processing ways are
employed in conjunction with the net page re-arrangement in step with the
behavior of the user within the past by deciding the concurrent access of the
knowledge and also the order during which the net pages on E-Government are
visited

Construction Sector: data processing is employed within the construction


sector in construction, project management, hydraulics, activity health and safety
`applications, analysis of earthquake information, groundwork studies and plenty of
alternative areas. in sight of the studies meted out during this context, it's been found
that studies are done:

● to make the knowledge classification theme in project documents


● to work out the tax connected corruption, · to predict the impact of changes
within the legal system on the budget,
● to work out waste and stop injury caused by waste,
● to estimate the value of main road construction
● to estimate population, forecast the weather, verify new job opportunities
● to estimate the compressive strength of the cement product · to live
performance of staff, manage the business processes,
● to outline the characteristics of activity accidents within the housing industry
● to classify public expenditures, arrange the proper use of resources,
● to live employee productivity

18 | Page
● to work out the concrete compressive strength
● to forecast the long run of public investment, analyze the information in
defense trade,
● to work out the connection of leadership-motivation between the chief and
also the employee
● to work out that offenders ar doubtless to commit crimes in terms of safety,
● to work out the situation of information mining technique in construction
management.

Engineering and Science Sector: giant quantities of information are


collected from scientific fields like natural philosophy, bioinformatics,
computing, criminal science, engineering, geosciences, arithmetic, computer
code etc. data processing provides several advantages in engineering and science
sector including:

✔ managing the method of the computer code employed in corporations,


✔ reducing the quantity of tasks,
✔ increasing the speed of the computer code Development Life Cycle,
✔ saving time and energy
✔ providing competitive advantage to the organizations with the anticipated
analysis,
✔ up the producing method
✔ biological literature analysis,
✔ remote sensing,
✔ soil quality analysis
✔ detective work crime pattern,
✔ period feature extraction for flow analysis,
✔ getting smart quality seed
✔ evolving new crop breeds
✔ classifying the astronomical objects,
✔ system modeling,
✔ discovering the relationships for best utilization of the cold storages and
use of canal water
✔ classifying the sequences in bioinformatics.

These usage areas in engineering and science sector facilitate users to boost
system performance of computer code used, to supply insight into several
elements of used engineering computer code development processes, and to
set up the longer-term decision-making method.

19 | Page
Chapter-5
5 IMPACTS OF BIG DATA IN ASSET MANAGEMENT
Alternative data in asset management goes beyond traditional data and refers to styles of data
that haven't been primarily utilized in the past to tell investment strategies. Big Data makes these
forms of data sets more accessible to investors.

In the past, institutional investors and investment managers will examine fundamental market
data like company earnings to tell their investment strategies. Investors are now gazing
alternative data sets like Mastercard transactions, survey data, weather forecasts, social media
data, and more to form their decisions. Big data makes easier for institutional investors for
creating strategies.

20 | Page
Investors are using alternative data, like web scraping or crowdsourcing as they become more
common practice in asset management.

With the massive number of information occupied by asset manager, they perform various
processes:

5.1 Data Science and Acquisition

Data science skills are getting used by financial services organizations to aggregate data from
various sources and make predictive models to boost decision-making.

Due to Big Data’s complexity and blend of structured and unstructured data models, data
scientists are needed to extract meaningful insights from data sets. More and more investors are
using data science includes data intelligence like AI and machine learning, to know data

5.2 Insightful Advanced Analytics

Asset managers are using advanced analytics to enable portfolio managers to access and
understand previously unreadable data. Data visualization and customized dashboards help asset
managers uncover investment patterns supported consumer behavior, transaction data, payroll
data from companies, and more.

Asset management firms can leverage pre-configured and pre-selected key performance
indicators created by industry-leading data science teams. The visualized models don’t require
in-house data science teams to make and are easy to integrate with existing data analysis
workflows.

5.3 Daily Revenue Signals for Revenue Prediction

The quantity of knowledge available also makes it possible for asset managers to receive daily
revenue signals they'll use to drive corporate or sector revenue predictions.

With the number of knowledge available, there's also the necessity for proven accuracy and
effectiveness. As an increasing number of companies leverage Big Data, whether or not it can
add Alpha to investment decisions could be a major consideration for asset managers.

By having access to those signals and alternative data sources, firms make more effectively make
portfolio allocation decisions, manage risk and reduce the likelihood of falling behind
competitors with seasoned data analytics solutions in situ.

5.4 The Future of knowledge Analytics in Asset Management

Data Analytics in Asset Management is undoubtedly an enormous topic moving forward. Asset
managers who are unable to use big data technology and leverage data to reinforce their research,
21 | Page
using big data for investment higher cognitive process, assist with risk analysis, and other
capabilities are likely to struggle to remain competitive in coming years.

The leading kinds of alternative data most companies are already using or attempt to start using
include web scraping, crowdsourcing, credit cards and POS systems, social media sentiment
analysis, search trends, and web traffic.

It is up to asset managers to find differing kinds of information and the way to best leverage it
for his or her clients. Asset management companies are hiring more data specialists for in-house
expertise that some years from now could make all the difference in their ability to stay getting
the foremost out of huge Data.

Moving forward, data intelligence software like computer science (AI) and machine learning will
still streamline investment processes across the board for asset managers, and can likely leave
them with longer to assist a greater number of clients.

It is also important to recollect that the longer data science technologies are refined and matured,
additionally to the growing amounts of information and data types being generated worldwide,
the higher they’ll become at extracting actionable insights and deciphering the foremost complex
of knowledge sets.

References
1. https://fila-lab.de/index.php/our-work/big-data-transmission/

2. https://www.coforge.com/salesforce/blog/data-analytics/understanding-the-3-vs-of-big-data-
volume-velocity-and-variety/

3. https://www.mckinsey.com/business-functions/quantumblack/our-insights/how-companies-are-
using-big-data-and-analytics

4. https://www.analyticssteps.com/blogs/companies-uses-big-data

22 | Page

You might also like