Professional Documents
Culture Documents
School Of Computer
Engineering
A Special
Thanks to
J. Han and M. Kamber.
&
Tan, Steinbach, Kumar
for their slides and books, which I have
used for preparation of these slides.
Chapter Contents
3
q DataVisualization
q Summary
What is Data?
4
q Collection of data objects and Attributes
their attributes
q An attribute is a property or
characteristic of an object Tid Refund Marital Taxable
Status Income Cheat
Ø Examples: eye color of a
1 Yes Single 125K No
person, temperature, etc.
2 No Married 100K No
Ø Attribute is also known as
3 No Single 70K No
variable, field, characteristic,
4 Yes Married 120K No
or feature
5 No Divorced 95K Yes
q A collection of attributes describe Objects
6 No Married 60K No
an object
7 Yes Divorced 220K No
Ø Object is also known as 8 No Single 85K Yes
record, point, case, sample, 9 No Married 75K No
entity, or instance 10 No Single 90K Yes
10
Data Objects & Attribute
5
q Data sets are made up of data objects.
q A data object represents an entity.
q Examples:
ü sales database: customers, store items, sales
ü medical database: patients, treatments
ü university database: students, professors, courses
q Also called samples , examples, instances, data points, objects, tuples.
q Data objects are described by attributes.
q Database rows -> data objects; columns ->attributes.
q Attribute (or dimensions, features, variables): a data field, representing a
characteristic or feature of a data object.
E.g., customer _ID, name, address
q Types:
Ø Nominal
Ø Binary
Ø Numeric: quantitative
ü Interval-scaled, Ratio-scaled
Attribute Values
6
q Ordinal
ü Values have a meaningful order (ranking) but magnitude between
successive values is not known.
ü Size = {small, medium, large}, grades, army rankings
ü Other examples of ordinal attributes include Grade (e.g., A+, A, A−, B+,
and so on) and Professional rank. Professional ranks can be enumerated
in a sequential order, such as assistant, associate, and full for professors,
q The central tendency of an ordinal attribute can be represented by its mode and
its median (the middle value in an ordered sequence), but the mean cannot be
defined.
q Qualitative attributes are describes a feature of an object, without giving an
actual size or quantity. The values of such qualitative attributes are typically
words representing categories.
Types of Attributes...
10
Distinctness: =
Order: < >
Addition: + -
Multiplication: */
12
Properties of Attribute Values...
13
Attribute Transformation Comments
Level
Nominal Any permutation of values If all employee ID numbers were
reassigned, would it make any
difference?
Interval new_value =a * old_value + b where a and b Thus, the Fahrenheit and Celsius
are constants temperature scales differ in terms
of where their zero value is and the
size of a unit (degree).
13
Discrete and Continuous Attributes
14
q Discrete Attribute
– Has only a finite or countably infinite set of values
– Examples: zip codes, counts, or the set of words in a collection of
documents
– Often represented as integer variables.
– Note: binary attributes are a special case of discrete attributes
q Continuous Attribute
– Has real numbers as attribute values
– Examples: temperature, height, or weight.
– Practically, real values can only be measured and represented using a finite
number of digits.
– Continuous attributes are typically represented as floating-point variables.
14
Types of data sets
15
q Record
Ø Relational records
Ø Data matrix, e.g., numerical matrix, crosstabs
Ø Document data: text documents: term-frequency vector
Ø Transaction data
q Graph
Ø World Wide Web
Ø Social and infomation network
Ø Molecular Structures
q Ordered
Ø Video data: sequence of images
Ø Temporal data: time-series
Ø Sequential Data: transaction sequences
Ø Genetic sequence data
q Spatial, image and multimedia
Ø Spatial data: maps
Ø Image data
Ø Video data 15
Important Characteristics of Structured Data
16
q Dimensionality
Ø Curse of Dimensionality
q Sparsity
q Resolution
16
Record Data
17
q Data that consists of a collection of records, each of which consists of a fixed set of
attributes
17
Data Matrix
18
q If data objects have the same fixed set of numeric attributes, then the data objects
can be thought of as points in a multi-dimensional space, where each dimension
represents a distinct attribute
q Such data set can be represented by an m by n matrix, where there are m rows,
one for each object, and n columns, one for each attribute
18
Document Data
19
q Each document becomes a `term' vector,
Ø each term is a component (attribute) of the vector,
Ø the value of each component is the number of times the corresponding term
occurs in the document.
19
Transaction Data
20
TID Items
1 Bread, Coke, Milk
2 Beer, Bread
3 Beer, Coke, Diaper, Milk
4 Beer, Bread, Diaper, Milk
5 Coke, Diaper, Milk
20
Graph Data
21
2
5 1
2
5
21
Chemical Data
22
22
Ordered Data
23
Sequences of transactions
Items/Events
An element of the
sequence
23
Ordered Data
24
Sequences of transactions
Items/Events
An element of the
sequence
24
Ordered Data
25
q Genomic sequence data
GGTTCCGCCTTCAGCCCCGCGCC
CGCAGGGCCCGCCCCGCGCCGTC
GAGAAGGGCCCGCCTGGCGGGCG
GGGGGAGGCGGGGCCGCCCGAGC
CCAACCGAGTCCGACCAGGTGCC
CCCTCTGCTCGGCCTAGACCTGA
GCTCATTAGGCGGCAGCGGACAG
GCCAAGTAGAACACGCGAAGCGC
TGGGCTGCCTGCTGCGACCAGGG
25
Ordered Data
26
q Spatio-Temporal Data
Average Monthly
Temperature of land
and ocean
26
Data Quality
27
q What kinds of data quality problems?
q How can we detect problems with the data?
q What can we do about these problems?
Data Preparation = Cleaning the Data
q Data Preparation can take 40-80% (or more) of the effort in a data mining project
ü Dealing with NULL (missing) values
ü Dealing with errors
ü Dealing with noise
ü Dealing with outliers (unless that is your science!)
ü Transformations: units, scale, projections
ü Data normalization
ü Relevance analysis: Feature Selection
ü Remove redundant attributes
ü Dimensionality Reduction
27
Noise
28
q For objects, noise is an extraneous object
q For attibutes, noise refers to modification of original values
Ø Examples: distortion of a person’s voice when talking on a poor phone and
“snow” on television screen
28
Outliers
29
q Outliers are data objects with characteristics that are considerably different than
most of the other data objects in the data set
29
Missing Values
30
q Reasons for missing values
Ø Information is not collected
(e.g., people decline to give their age and weight)
Ø Attributes may not be applicable to all cases
(e.g., annual income is not applicable to children)
30
Types of Missing Values
31
q Some definitions are based on representation: Missing data is the lack of a
recorded answer for a particular field.
Ø Missing completely at random (MCAR)
Ø Missing at Random (MAR)
Ø Missing Not at Random (MNAR)
31
Missing Completely at Random (MCAR)
32
q Missingness of a value is independent of attributes
ü Fill in values based on the attribute
ü Analysis may be unbiased overall
q The missingness on the variable is completely unsystematic.
q Example when we take a random sample of a population, where each member has
the same chance of being included in the sample.
32
Missing at Random (MAR)
33
q Missingness is related to other variables
q Fill in values based other values
q Almost always produces a bias in the analysis
33
Missing not at Random (MNAR) - Nonignorable
34
q Missingness is related to unobserved measurements
q When the missing values on a variable are related to the values of that variable
itself, even after controlling for other variables.
34
Formalize the definitions
35
q Let's X represent a matrix of the data we “expect” to have; X = {Xo, Xm} where
Xo is the observed data and Xm the missing data.
1. MCAR: P(R| Xo , Xm ) = P(R)
2. MAR: P(R| Xo , Xm ) = P(R| Xo)
q Let’s define R as a matrix with the same dimensions as X where Ri,j = 1 if the
datum is missing, and 0 otherwise.
3. MNAR: No simplification.
35
Duplicate Data
36
q Data set may include data objects that are duplicates, or almost duplicates of one
another
Ø Major issue when merging data from heterogeous sources
q Examples:
ü Same person with multiple email addresses
q Data cleaning
Process of dealing with duplicate data issues
q When should duplicate data not be removed?
36
Data Preprocessing
37
q Aggregation
q Sampling
q Dimensionality Reduction
q Feature subset selection
q Feature creation
q Discretization and Binarization
q Attribute Transformation
37
Aggregation
38
q Combining two or more attributes (or objects) into a single attribute (or object)
q Purpose
Ø Data reduction
ü Reduce the number of attributes or objects
Ø Change of scale
ü Cities aggregated into regions, states, countries, etc
Ø More “stable” data
ü Aggregated data tends to have less variability
38
Aggregation...
39
39
Sampling
40
q Sampling is the main technique employed for data selection.
Ø It is often used for both the preliminary investigation of the data and the final
data analysis.
q Statisticians sample because obtaining the entire set of data of interest is too
expensive or time consuming.
q Sampling is used in data mining because processing the entire set of data of
interest is too expensive or time consuming.
q The key principle for effective sampling is the following:
Ø using a sample will work almost as well as using the entire data sets, if the
sample is representative
Ø A sample is representative if it has approximately the same property (of
interest) as the original set of data
40
Types of Sampling
41
q Simple Random Sampling
ü There is an equal probability of selecting any particular item
q Sampling without replacement
ü As each item is selected, it is removed from the population
q Sampling with replacement
ü Objects are not removed from the population as they are selected for the
sample.
ü In sampling with replacement, the same object can be picked up more than
once
q Stratified sampling
ü Split the data into several partitions; then draw random samples from each
partition
41
Sample Size
42
42
Sample Size
43
q What sample size is necessary to get at least one object from each of 10
groups.
43
Curse of Dimensionality
44
44
Dimensionality Reduction
45
q Purpose:
Ø Avoid curse of dimensionality
Ø Reduce amount of time and memory required by data mining algorithms
Ø Allow data to be more easily visualized
Ø May help to eliminate irrelevant features or reduce noise
q Techniques
Ø Principle Component Analysis
Ø Singular Value Decomposition
Ø Others: supervised and non-linear techniques
45
Dimensionality Reduction: PCA
46
q Goal is to find a projection that captures the largest amount of variation in data
x2
x1
q Find the eigenvectors of the covariance matrix
q The eigenvectors define the new space
46
Dimensionality Reduction: ISOMAP
47
47
Feature Subset Selection
48
q Another way to reduce dimensionality of data
q Redundant features
Ø duplicate much or all of the information contained in one or more other
attributes
ü Example: purchase price of a product and the amount of sales tax paid
q Irrelevant features
Ø contain no information that is useful for the data mining task at hand
ü Example: students' ID is often irrelevant to the task of predicting
students' GPA
q Techniques:
Ø Brute-force approch: Try all possible feature subsets as input to data mining
algorithm
Ø Embedded approaches: Feature selection occurs naturally as part of the data
mining algorithm
Ø Filter approaches: Features are selected before data mining algorithm is run
Ø Wrapper approaches: Use the data mining algorithm as a black box to find
best subset of attributes
48
Feature Creation
49
q Create new attributes that can capture the important information in a data set
much more efficiently than the original attributes
49
Mapping Data to a New Space
50
q Fourier transform
q Wavelet transform
50
Discretization Using Class Labels
51
q Entropy based approach
51
Discretization Without Using Class Labels
52
52
Attribute Transformation
53
q A function that maps the entire set of values of a given attribute to a new set of
replacement values such that each old value can be identified with one of the new
values
Ø Simple functions: xk, log(x), ex, |x|
Ø Standardization and Normalization
53
Similarity and Dissimilarity
54
q Similarity
Ø Numerical measure of how alike two data objects are.
Ø Is higher when objects are more alike.
Ø Often falls in the range [0,1]
q Dissimilarity
Ø Numerical measure of how different are two data objects
Ø Lower when objects are more alike
Ø Minimum dissimilarity is often 0
Ø Upper limit varies
q Proximity refers to a similarity or dissimilarity
54
Similarity/Dissimilarity for Simple Attributes
55
q p and q are the attribute values for two data objects.
55
Euclidean Distance
56
q Euclidean Distance
n 2
dist ( pk qk )
k 1
56
Euclidean Distance
57
p1
point x y
2
p1 0 2
p3 p4
p2 2 0
1
p2 p3 3 1
0 p4 5 1
0 1 2 3 4 5 6
p1 p2 p3 p4
p1 0 2.828 3.162 5.099
p2 2.828 0 1.414 3.162
p3 3.162 1.414 0 2
p4 5.099 3.162 2 0
Distance Matrix
57
Minkowski Distance
58
q Minkowski Distance is a generalization of Euclidean Distance
1
n r r
dist ( | pk qk |)
k 1
58
Minkowski Distance:Example
59
q r = 1. City block (Manhattan, taxicab, L1 norm) distance.
A common example of this is the Hamming distance, which is just the number
of bits that are different between two binary vectors
q r = 2. Euclidean distance
q Do not confuse r with n, i.e., all these distances are defined for all numbers of
dimensions.
59
Minkowski Distance
60
L1 p1 p2 p3 p4
p1 0 4 4 6
p2 4 0 2 4
p3 4 2 0 2
p4 6 4 2 0
point x y
p1 0 2 L2 p1 p2 p3 p4
p2 2 0 p1 0 2.828 3.162 5.099
p3 3 1 p2 2.828 0 1.414 3.162
p4 5 1 p3 3.162 1.414 0 2
p4 5.099 3.162 2 0
L p1 p2 p3 p4
p1 0 2 3 5
p2 2 0 1 3
p3 3 1 0 2
p4 5 3 2 0
Distance Matrix
60
Mahalanobis Distance
61
1 T
mahalanobi s ( p , q ) ( p q ) ( p q )
1 n
j ,k
n 1 i 1
( X ij X j )( X ik X k )
Covariance Matrix:
0 .3 0 .2
0 . 2 0 . 3
A: (0.5, 0.5)
B: (0, 1)
C: (1.5, 1.5)
Mahal(A,B) = 5
Mahal(A,C) = 4
62
Common Properties of a Distance
63
q Distances, such as the Euclidean distance, have some well known properties.
63
Similarity Between Binary Vectors
64
q Common situation is that objects, p and q, have only binary attributes
64
SMC versus Jaccard: Example
65
p= 1000000000
q= 0000001001
65
Cosine Similarity
66
q If d1 and d2 are two document vectors, then
cos( d1, d2 ) = (d1 d2) / ||d1|| ||d2|| ,
where indicates vector dot product and || d || is the length of vector d.
q Example:
d1 = 3 2 0 5 0 0 0 2 0 0
d2 = 1 0 0 0 0 0 0 1 0 2
d1 d2= 3*1 + 2*0 + 0*0 + 5*0 + 0*0 + 0*0 + 0*0 + 2*1 + 0*0 + 0*2 = 5
||d1|| = (3*3+2*2+0*0+5*5+0*0+0*0+0*0+2*2+0*0+0*0)0.5 = (42) 0.5 = 6.481
||d2|| = (1*1+0*0+0*0+0*0+0*0+0*0+0*0+1*1+0*0+2*2) 0.5 = (6) 0.5 = 2.245
66
Recommended Text and Reference Books
67
q Text Book:
Ø J. Han and M. Kamber. Data Mining: Concepts and Techniques. Morgan
Kaufmann, 3rd ed., 2011
q Reference Books:
Ø H. Dunham. Data Mining: Introductory and Advanced Topics. Pearson
Education. 2006.
Ø I. H. Witten and E. Frank. Data Mining: Practical Machine Learning Tools
and Techniques. Morgan Kaufmann. 2000.
Ø D. Hand, H. Mannila and P. Smyth. Principles of Data Mining.Prentice-Hall.
2001.
67
68