You are on page 1of 3

STATISTICS

COMPRE
1. Solve for the mean, median and mode of the set of data below: 24 28 25 26 23
25 27 28 30
MEAN-26.22
solution:24 +28+ 25+ 26+ 23+ 25+ 27+ 28+ 30 divided by 9
236 divided by 9 = 26.222
MEDIAN-26
solution Md= 23 24 25 25 26 27 28 28 30
MODE- 25 and 28 the distribution is bimodal
solution
23, 24, (25, 25) ,26, 27, (28, 28) ,30
2. Compute for the Standard Deviation and Variance of the set of data given: 6 10
8 13 5 11 10
mean-9
VARIANCE- 6.85714285714
STANDARD DEVIATION- 2.618614

3. What is Chi-Square? When do we use Chi-Square? Make a hypothetical


problem and data where chi-square is used. 

A chi-square statistic is a test that compares expected data to


actual observed data (or model results). It requires data that is random,
raw, mutually exclusive, collected from independent variables, and drawn
from a large enough sample. This is a common method of hypothesis
testing.

For example, since the Philippine national election is next year, a


random survey was conducted among 2,000 voters, both male and
female. The respondents were divided into three groups according on their
gender and whether they were in favor of the BLUE Team, ORANGE
Team, PINK Team, or RED Team. Then, considering a grid with the
columns, BLUE, ORANGE, PINK, or RED as well as male and female
rows. Assume the following data from the 2,000 respondents: Finding the
predicted frequencies is the first step in calculating the chi squared
statistic. For each "cell" in the grid, they are calculated. There are six total
anticipated frequencies because there are two gender categories and
three political perspective categories. The predicted frequency is
calculated using the following formula: Row Total * (Column Total)/N.
The sum of these values, or 32.41, is the chi-squared statistic. We
may next look at a chi-squared statistic table to evaluate if the result is
statistically significant or not, given the degrees of freedom in our setup.
4. What is ANOVA. Give an example of a problem where ANOVA is used.
The analysis of variance (ANOVA) is a statistical method that divides a data set's
observed aggregate variability into two parts: systematic components and
random factors. Random factors have no statistical impact on the supplied data
set, whereas systematic influences do. In a regression research, analysts utilize
the ANOVA test to examine the impact of independent factors on the dependent
variable.

Exercising ANOVA in Real Life #1


A large-scale farm is trying to figure out which of three fertilizers produces the
maximum crop output. They apply each fertilizer to ten different fields and then
calculate the total yield at the end of the season.
Researchers can use a one-way ANOVA with "kind of fertilizer" as the factor and
"crop yield" as the response to see if there is a statistically significant difference
in the mean yield that results from these three fertilizers.

We can conclude that there is a statistically significant difference if the overall p-


value of the ANOVA is less than our significance level (usually 0.10, 0.05, 0.01).

Degrees of freedom
The amount of independent pieces of information used to calculate an estimate is
referred to as degrees of freedom. It's not nearly the same as the sample's total
number of items. To calculate the df for the estimate, subtract 1 from the total
number of elements. Assume you're looking for the average weight loss on a low-
carb diet. You may use four people, giving you three degrees of freedom (4 – 1 =
three), or one hundred persons, giving you df = 99.

mean, median, and mode, in mathematics, the three principal ways of


designating the average value of a list of numbers. The arithmetic mean is found
by adding the numbers and dividing the sum by the number of numbers in the
list. This is what is most often meant by an average. The median is the middle
value in a list ordered from smallest to largest. The mode is the most frequently
occurring value on the list.

Standard deviation and variance are both determined by using the mean of


a group of numbers in question. The mean is the average of a group of
numbers, and the variance measures the average degree to which each
number is different from the mean. The extent of the variance correlates to
the size of the overall range of numbers—meaning the variance is greater
when there is a wider range of numbers in the group, and the variance is
less when there is a narrower range of numbers.

Variance
The variance is the average of the squared differences from the mean. To figure
out the variance, first calculate the difference between each point and the mean;
then, square and average the results.

For example, if a group of numbers ranges from 1 to 10, it will have a mean of
5.5. If you square the differences between each number and the mean, and then
find their sum, the result is 82.5. To figure out the variance, divide the sum, 82.5,
by N-1, which is the sample size (in this case 10) minus 1. The result is a
variance of 82.5/9 = 9.17. Standard deviation is the square root of the variance
so that the standard deviation would be about 3.03.

The Correlation Coefficient (r)


The sample correlation coefficient (r) is a measure of the closeness of association of the points
in a scatter plot to a linear regression line based on those points, as in the example above for
accumulated saving over time. Possible values of the correlation coefficient range from -1 to +1,
with -1 indicating a perfectly linear negative, i.e., inverse, correlation (sloping downward) and +1
indicating a perfectly linear positive correlation (sloping upward).

You might also like