Professional Documents
Culture Documents
Highlighted CC104.Rev
Highlighted CC104.Rev
CHAPTER 1
One of the associated models in information
management is the DIKAR model. This model shows
Information Management - Choo (1995) defined the transition of raw facts into information; then,
information management (IM) as the management of from information, knowledge can be derived.
processes and systems responsible for acquiring,
creating, organizing, distributing, and using - Knowing this model helps organizations identify the
information. technology needs to assist in the efficient
transformation of data for their business needs.
The five keys to IM:
Leadership Data
Operation Information
Process Knowledge
Data Action
Technology Result
- Data is a raw form of knowledge and, on its own, The manual file system is the traditional pen paper-
doesn’t carry any significance or purpose. In other folder file management. It is composed of a
words, you have to interpret data for it to have collection of file folders, categorized and stored in
meaning. Data can be simple—and may even seem cabinets
useless until it is analyzed, organized, and interpreted. The computerized file system uses computers and
storage devices to acquire, organize, and store data.
There are two main types of data: It also facilitates faster data tracking and report
Quantitative data is provided in numerical form, generation as compared to the manual method.
like the weight, volume, or cost of an item.
Qualitative data is descriptive, but non-numerical, Data processing - occurs when data is collected and
like the name, sex, or eye color of a person. translated into usable information. Usually
performed by a data scientist or team of data
Information is defined as knowledge gained through scientists, it is important for data processing to be
study, communication, research, or instruction. done correctly as not to negatively affect the end
Essentially, information is the result of analyzing and product, or data output.
interpreting pieces of data.
Data processing starts with data in its raw form and
- Whereas data is the individual figures, numbers, or converts it into a more readable format (graphs,
graphs, information is the perception of those pieces documents, etc.),
of knowledge.
Six stages of Data processing
INFORMATION AND ITS IMPORTANCE
1.Data collection - Collecting data is the first step in
INFORMATION MANAGEMENT data processing. Data is pulled from available
sources, including data lakes and data warehouses.
Choo (1995) defined information management (IM) as
the management of processes and systems 1 2.Data preparation - Once the data is collected, it
responsible for acquiring, creating, organizing, then enters the data preparation stage.
distributing, and using information.
Data preparation, often referred to as “pre processing” Data inconsistency
is the stage at which raw data is cleaned up and This happens when multiple copies of the data have
organized for the following stage of data processing. conflicting versions. This is usually a result of a
particular update done in one document but failed
3.Data input - The clean data is then entered into its to update other existing copies.
destination (data warehouse). and translated into a
language that it can understand. Lack of data integrity
The existence of data redundancies might
-Data input is the first stage in which raw data begins to compromise data integrity. Integrity is defined as
take the form of usable information. the consistency and accuracy of data.
4.Processing - During this stage, the data inputted to 3. Lack of design structure and models
the computer in the previous stage is actually
processed for interpretation File systems usually lack design and data models as
compared to database systems.
-Processing is done using machine learning algorithms,
though the process itself may vary slightly depending - Data models contribute to the overall structure
on the source of data being processed. and logic flow that enable databases to be more
organized in managing data.
5.Data output/interpretation- The
output/interpretation stage is the stage at which data is 4. Exposure to various data anomalies
finally usable to non data scientists.It is translated,
readable, and often in the form of graphs, videos, Anomaly means inconsistency in the pattern from
images, plain text, etc.). the normal form.
6. Data storage - The final stage of data processing is In Database Management System (DBMS), anomaly
storage. After all of the data is processed, it is then means the inconsistency occurred in the relational
stored for future use. While some information may be table during the operations performed on the
put to use immediately, much of it will serve a purpose relational table.
later on.
The other reason for the database anomalies is that
File System Problems all the data is stored in a single table
1.Dependence in structural and data changes So, to remove the anomalies of the database,
normalization is the process which is done where
Data dependence happens when data structure the splitting of the table and joining of the table
changes affect data access. (different types of join) occurs
Data independence is when these changes do not affect Types of data anomalies are as follow:
the data nor the related programs.
Update anomaly - happens when an update
Data structure changes include: change in keys, change done results in data inconsistency or
in data types. discrepancy.
1. Data dictionary management A data dictionary Centralized database is located and maintained
holds the metadata about the database structures. at a single site.
DBMS offers features that help manage and generate Distributed database is distributed across
data dictionaries for reference and documentation several different sites and servers.
purposes.
2. Based on data stored:
2. Data storage management Database performance
can be fine-tuned to be more efficient in the storage General-purpose database usually includes a wide
and retrieval of information. range of topics and/or types of documents.
3. Data presentation DBMS can aid in the acquisition, Discipline-specific database focuses on a single
storage, and transformation of data presented in a academic discipline or a group of related fields.
structured way to satisfy logical expectations.
4.Based on the how it will be used and the time
4. Security management and multi-access control sensitivity of information:
Access roles and levels can be configured through an
Access Control List (ACL), which allows the limited Operational database stores and supports real-
rights and privileges per user type. time business data.
Analytical database contains historical data that
5. Back-up and recovery management Back-up and can be used for business intelligence.
recovery options are available to avoid database
outages and data loss. 5.Based on the degree which the data are
structured:
6. Data integrity management Integrity is the
consistency and accuracy of data. Through the Relational database is used to handle structured
enforcement of the integrity rules, DBMS is capable data.
of ensuring data integrity. Structured data – are easily searchable by
algorithms.
7. Database application and communication Non-relational database (NoSQL) is for handling
interfaces DBMS can handle the Structured Query unstructured data.
Language (SQL) and its connection with other Unstructured data – have no predefined structure.
programming languages and applications. Also, it
supports communication with other databases. CHAPTER 2
Types of Databases Data Model and data modeling
3. Based on the number of users: Database - design focuses on how the database
structure will be used to store and manage end user
Single user database supports one user at a given data.
time.
- Data Modeling and Data Models Database design
Workgroup database is intended for a group of users. primarily focuses on the processing and handling of
end-user data using the database structure.
Enterprise database supports multiple groups of
users working together and accessing files - Data modeling, the first step in designing a
simultaneously, such as organizations and database, refers to the process of creating a specific
corporations. data model for a determined problem domain.
4
In developing a database, the first phase is data 2. Data models are a vital data architecture
modeling, which refers to creating a data model for a component, along with flow diagrams, architectural
given problem domain. blueprints, a unified vocabulary, and other artifacts.
- Data modeling reduces the complexities of database The network model is known for its flexible way of
design. It is an iterative and progressive process that representing data. The structure follows a graph
begins with a simple understanding of the problem format where objects are nodes and relationships are
domain, and as your understanding increases, so does arcs.
the level of detail in the data model.
4. Object-Oriented Model
Ideally…
The object-oriented model envisions entities as
Ideally, data models are living documents that evolve objects. Each object has attributes (state) and methods
along with changing business needs. (behavior). An object belongs to a single class only and
as its instance.
- They play an important role in supporting business
processes and planning IT architecture and strategy. The “objects” involved are abstractions of real-world
Data models can be shared with vendors, partners, entities. Objects are grouped in class hierarchies, and
and/or industry peers. have associated features.
A Logical View of Data Relational database model’s Click super - Super key is the set of all the keys which help
to edit Master title style structural and data to identify rows in a table uniquely.
independence
candidate key - Candidate keys are those attributes
-allow us to ignore the structure’s physical complexity. that uniquely identify rows of a table.
-enables us to view data logically rather than physically. alternate key - All the keys which did not become the
primary Key are called alternate keys
• The logical view allows a simpler file concept of data
storage. foreign key - Foreign Key is used to establish
• The use of logically independent tables is easier to relationships between two tables.
understand.
composite key - a composite key is a set of two or
Relational Database Model (RDM) – is presented as data more attributes that help identify each tuple in a
tables that collect groups of elements into relations. table uniquely.
These models work based on the idea that each table unique key - Unique Key is a column or set of
setup will include a primary key or identifier. Relations columns that uniquely identify each record in a table.
are represented by tables. 7
1. Primary Key DATA DICTIONARY
• It is the attribute that is used to differentiate each row
in a table. A data dictionary is a centralized repository of
• Responsible for entity integrity. metadata about the data in a database. It contains
• Requirements: information about the structure of the database,
• the value must be different from all other values for such as the names and descriptions of tables,
that attribute; unique columns, data types, constraints, relationships, and
• the value null cannot be used for an attribute to be other relevant attributes.
used as a key
SYSTEM CATALOG
2. Foreign Key
• It forms a relationship by placing a primary from one A system catalog, also known as a system catalog
table to another table. table or data dictionary table, is a special set of
• It enforces referential integrity. tables within a database that stores metadata about
• It is safeguards the tables from actions that would the database itself.
discredit links between tables.
• It ensures that the values being entered in the foreign - It contains information about the database objects
key column is valid and exists in the table it points to. such as tables, indexes, views, users, privileges, and
other system-related details.
Relational Schema
DATA REDUNDANCY REVISITED
A schema is a logical grouping of database objects
related to one another and becomes a guide on how the Update anomalies – inconsistencies that might
database is constructed. happen when we do an update in a record.
REFERENTIAL INTEGRITY -
• requires that a foreign key must have a matching
primary key or it must be null. This constraint is specified
between two tables (parent and child); it maintains the
correspondence between rows in these tables.