You are on page 1of 3

Categorical: => If the data is Nominal/Ordinal then consider Bar/ Pie Charts

Quantitative: => If the data is interval/ratio then consider box plot for small data and Histogram chart for large data

Besides summarizing data by means of tables and/or graphs, it can also be useful to describe the center of a distribution. We can
do that by means of so-called measures of central tendency: the mode, median, and mean.
Yet to adequately describe a distribution we need more information. We also need information about the variability or dispersion
of the data. We need, in other words, measures of dispersion. Well-known measures of dispersion are the range,
the interquartile range, the variance, and the standard deviation. 
A graph that nicely presents the variability of a distribution is the box plot.

The mode will be used for Categorical variables


Median/Mean will be used for Quantitative variables => If the data is skewed/has outliners use Median else use Mean 

Sometimes Mode/Median/Mean can’t describe the data distribution hence we need more info to draw conclusions about the data.
The measure of variability is an important factor to measure that. 
1. Range: - Highest-Lowest value in the data variables tells how widely the data is distributed
2. Interquartile Range: - Divide the data distribution into 4 parts. To achieve that mark the Median has half which divides
the data into 2 parts. The upper part is again divided by Median to get Q1 and Q2 Interquartile, and the lower part is
again divided by Median to get Q3 and Q4 Interquartile
3. Variance
4. Standard Deviation

Although Range and Interquartile range do the job it doesn’t consider all the values in the variables so to get the accurate
variability of a variable, we need Variance and Standard Deviation.

Sometimes researchers want to know if a specific observation is common or exceptional. To answer that question, they express a
score in terms of the number of standard deviations it is removed from the mean. This number is what we call a z-score. If we
recode original scores into z-scores, we say that we standardize a variable.

Sample space: Collection of all possible outcomes for a random phenomenon


Event: A subset of the sample space.
Each event has a probability and to find the probability we can use tree diagram.

Calculate probabilities of combines events:


 For a particular outcome, all probabilities along that path should be multiplied.
 For an event that includes multiple outcomes, the probabilities for all the outcomes
should be summed.

The probabilities that values of different variables co-occur are called joint probabilities. Such
joint probabilities can be summed over one variable to form so-called marginal probabilities (the
probabilities for a single variable). In the first video-lecture these concepts of joint and marginal
probability are explained by using an example of beach-visitors which are characterized by
gender and activity.

In the subsequent video the conditional probability is defined and clarified. Conditional
probability is the probability of an event, given that another event occurs. In this lecture the
example with beach-visitors is used again, but now to illustrate how probabilities of the activity
distribution would f.i. change conditional on gender. Mathematically, the conditional probability
of A given B equals the joint probability of A and B, divided by the (marginal)probability of
event B.

The third video explains that independence of random events is closely related to the conditional
probability between these events. It appears that random events are independent if the joint
probability of these events is equal to the product of the marginal probabilities or, equivalent, if
the conditional probability of random variable equals its marginal probability.

The fourth and final video of this module brings together joint and conditional probabilities as
well as tree-diagrams to derive and intuitively explain Bayes' law. It shows how the observation
that joint probabilities can be calculated based on the conditional probability of A given B as
well as B given A, leads to Bayes’ law. The last video lecture ends by explaining that while
Bayes' follows mechanically by rewriting two probability-relations, it is often applied much
more generally - by updating a prior belief in or knowledge about a hypothesis A with additional
evidence B towards a posterior belief (A conditional on B).

You might also like