Professional Documents
Culture Documents
Francis
Additional Mathematics Project Work
2020
Introduction to Statistics
Introduction of Obesity
Problem Solving
Histogram
Frequency Polygon
Ogive
Further Exploration
Student Role
Conclusion
Introduction of Statistcs
Definition
Statistical Methods
a) Descriptive statistic
A descriptive statistic (in the count noun sense) is a summary statistic that
quantitatively describes or summarizes features of a collection
of information, while descriptive statistics in the mass noun sense is the process of
using and analyzing those statistics. Descriptive statistics is distinguished
from inferential statistics (or inductive statistics), in that descriptive statistics aims to
summarize a sample, rather than use the data to learn about the population that the
sample of data is thought to represent.
b) Inferential statistic
Statistical inference is the process of using data analysis to deduce properties of an
underlying probability distribution. Inferential statistical analysis infers properties of
a population, for example by testing hypotheses and deriving estimates. It is assumed
that the observed data set is sampled from a larger population. Inferential statistics can
be contrasted with descriptive statistics. Descriptive statistics is solely concerned with
properties of the observed data, and it does not rest on the assumption that the data
come from a larger population.
History of Statiscal Method
The earliest European writing on statistics dates back to 1663, with the publication
of Natural and Political Observations upon the Bills of Mortality by John
Graunt. Early applications of statistical thinking revolved around the needs of states to
base policy on demographic and economic data, hence its stat- etymology. The scope
of the discipline of statistics broadened in the early 19th century to include the
collection and analysis of data in general. Today, statistics is widely employed in
government, business, and natural and social sciences.
The mathematical foundations of modern statistics were laid in the 17th century with
the development of the probability theory by Gerolamo Cardano, Blaise
Pascal and Pierre de Fermat. Mathematical probability theory arose from the study
of games of chance, although the concept of probability was already examined
in medieval law and by philosophers such as Juan Caramuel. The method of least
squares was first described by Adrien-Marie Legendre in 1805.
Application of Statistics
Statistical consultants can help organizations and companies that don't have in-house
expertise relevant to their particular questions.
Statistical methods
Descriptive statistics
Main article: Descriptive statistics
Inferential statistics
Main article: Statistical inference
A random variable that is a function of the random sample and of the unknown
parameter, but whose probability distribution does not depend on the unknown
parameter is called a pivotal quantity or pivot. Widely used pivots include the z-
score, the chi square statistic and Student's t-value.
Between two estimators of a given parameter, the one with lower mean squared
error is said to be more efficient. Furthermore, an estimator is said to be unbiased if
its expected value is equal to the true value of the unknown parameter being
estimated, and asymptotically unbiased if its expected value converges at the limit to
the true value of such parameter.
This still leaves the question of how to obtain estimators in a given situation and carry
the computation, several methods have been proposed: the method of moments,
the maximum likelihood method, the least squares method and the more recent
method of estimating equations.
The best illustration for a novice is the predicament encountered by a criminal trial.
The null hypothesis, H0, asserts that the defendant is innocent, whereas the alternative
hypothesis, H1, asserts that the defendant is guilty. The indictment comes because of
suspicion of the guilt. The H0 (status quo) stands in opposition to H1 and is maintained
unless H1 is supported by evidence "beyond a reasonable doubt". However, "failure to
reject H0" in this case does not imply innocence, but merely that the evidence was
insufficient to convict. So the jury does not necessarily accept H0 but fails to
reject H0. While one can not "prove" a null hypothesis, one can test how close it is to
being true with a power test, which tests for type II errors.
Error
Working from a null hypothesis, two basic forms of error are recognized:
Mean squared error is used for obtaining efficient estimators, a widely used class of
estimators. Root mean square error is simply the square root of mean squared error.
A least squares fit: in red the points to be fitted, in blue the fitted line.
Many statistical methods seek to minimize the residual sum of squares, and these are
called "methods of least squares" in contrast to Least absolute deviations. The latter
gives equal weight to small and big errors, while the former gives more weight to
large errors. Residual sum of squares is also differentiable, which provides a handy
property for doing regression. Least squares applied to linear regression is
called ordinary least squares method and least squares applied to nonlinear
regression is called non-linear least squares. Also in a linear regression model the non
deterministic part of the model is called error term, disturbance or more simply noise.
Both linear regression and non-linear regression are addressed in polynomial least
squares, which also describes the variance in a prediction of the dependent variable (y
axis) as a function of the independent variable (x axis) and the deviations (errors,
noise, disturbances) from the estimated (fitted) curve.
Measurement processes that generate statistical data are also subject to error. Many of
these errors are classified as random (noise) or systematic (bias), but other types of
errors (e.g., blunder, such as when an analyst reports incorrect units) can also be
important. The presence of missing data or censoring may result in biased
estimates and specific techniques have been developed to address these problems.
Interval estimation
Main article: Interval estimation
Confidence intervals: the red line is true value for the mean in this example, the
blue lines are random confidence intervals for 100 realizations.
Most studies only sample part of a population, so results don't fully represent the
whole population. Any estimates obtained from the sample only approximate the
population value. Confidence intervals allow statisticians to express how closely the
sample estimate matches the true value in the whole population. Often they are
expressed as 95% confidence intervals. Formally, a 95% confidence interval for a
value is a range where, if the sampling and analysis were repeated under the same
conditions (yielding a different dataset), the interval would include the true
(population) value in 95% of all possible cases. This does not imply that the
probability that the true value is in the confidence interval is 95%. From
the frequentist perspective, such a claim does not even make sense, as the true value is
not a random variable. Either the true value is or is not within the given interval.
However, it is true that, before any data are sampled and given a plan for how to
construct the confidence interval, the probability is 95% that the yet-to-be-calculated
interval will cover the true value: at this point, the limits of the interval are yet-to-be-
observed random variables. One approach that does yield an interval that can be
interpreted as having a given probability of containing the true value is to use
a credible interval from Bayesian statistics: this approach depends on a different way
of interpreting what is meant by "probability", that is as a Bayesian probability.
Significance
Main article: Statistical significance
Statistics rarely give a simple Yes/No type answer to the question under analysis.
Interpretation often comes down to the level of statistical significance applied to the
numbers and often refers to the probability of a value accurately rejecting the null
hypothesis (sometimes referred to as the p-value).
In this graph the black line is probability distribution for the test statistic,
the critical region is the set of values to the right of the observed data point
(observed value of the test statistic) and the p-value is represented by the green
area.
Referring to statistical significance does not necessarily mean that the overall result is
significant in real world terms. For example, in a large study of a drug it may be
shown that the drug has a statistically significant but very small beneficial effect, such
that the drug is unlikely to help the patient noticeably.
Statistical consultants can help organizations and companies that don't have in-house
expertise relevant to their particular questions.
Machine learning models are statistical and probabilistic models that capture patterns
in the data through use of computational algorithms.
Statistics in academy
The rapid and sustained increases in computing power starting from the second half of
the 20th century have had a substantial impact on the practice of statistical science.
Early statistical models were almost always from the class of linear models, but
powerful computers, coupled with suitable numerical algorithms, caused an increased
interest in nonlinear models (such as neural networks) as well as the creation of new
types, such as generalized linear models and multilevel models.
Increased computing power has also led to the growing popularity of computationally
intensive methods based on resampling, such as permutation tests and the bootstrap,
while techniques such as Gibbs sampling have made use of Bayesian models more
feasible. The computer revolution has implications for the future of statistics with a
new emphasis on "experimental" and "empirical" statistics. A large number of both
general and special purpose statistical software are now available. Examples of
available software capable of complex statistical computation include programs such
as Mathematica, SAS, SPSS, and R.
Statistics applied to mathematics or the arts
Definition of obesity
Overweight and obesity are defined as abnormal or excessive fat accumulation that
presents a risk to health. A body mass index (BMI) over 25 is considered overweight,
and over 30 is obese. The issue has grown to epidemic proportions, with over 4
million people dying each year as a result of being overweight or obese in 2017
according to the global burden of disease.
Rates of overweight and obesity continue to grow in adults and children. From 1975
to 2016, the prevalence of overweight or obese children and adolescents aged 5–19
years increased more than four-fold from 4% to 18% globally.
Obesity is one side of the double burden of malnutrition, and today more people are
obese than underweight in every region except sub-Saharan Africa and Asia. Once
considered a problem only in high-income countries, overweight and obesity are now
dramatically on the rise in low- and middle-income countries, particularly in urban
settings. The vast majority of overweight or obese children live in developing
countries, where the rate of increase has been more than 30% higher than that of
developed countries.
Classification of Obesity
Body Mass Index (BMI) is a person’s weight in kilograms divided by the square of
height in meters. A high BMI can be an indicator of high body fatness.
To calculate BMI, see the Adult BMI Calculator or determine BMI by finding your
height and weight in this BMI Index Chartexternal icon.
If your BMI is less than 18.5, it falls within the underweight range.
If your BMI is 18.5 to <25, it falls within the normal.
If your BMI is 25.0 to <30, it falls within the overweight range.
If your BMI is 30.0 or higher, it falls within the obese range.
At an individual level, BMI can be used as a screening tool but is not diagnostic of the
body fatness or the health of an individual. A trained healthcare provider should perform
appropriate health assessments in order to evaluate an individual’s health status and risks.
If you have questions about your BMI, talk with your health care provider.
BMI does not measure body fat directly, but research has shown that BMI is moderately
correlated with more direct measures of body fat obtained from skinfold thickness
measurements, bioelectrical impedance, underwater weighing, dual energy x-ray
absorptiometry (DXA) and other methods 1,2,3. Furthermore, BMI appears to be
strongly correlated with various adverse health outcomes consistent with these more
direct measures of body fatness
Effects on health
People who have obesity, compared to those with a normal or healthy weight, are at
increased risk for many serious diseases and health conditions, including the
following:
Causes
The balance between calorie intake and energy expenditure determines a person's
weight. If a person eats more calories than he or she burns (metabolizes), the person
gains weight (the body will store the excess energy as fat). If a person eats fewer
calories than he or she metabolizes, he or she will lose weight. Therefore, the most
common causes of obesity are overeating and physical inactivity. Ultimately, body
weight is the result of genetics, metabolism, environment, behavior, and culture.
History of obesity
In the evolutionary history of humankind, bodily fat seems to have served nature’s
purpose by outfitting the species with a built-in mechanism for storing its own food
reserves. During prehistoric times, when the burden of disease was that of pestilence
and famine, natural selection rewarded the “thrifty” genotypes of those who could
store the greatest amount of fat from the least amount of the then erratically available
foods and to release it as frugally as possible over the long run. This ability to store
surplus fat from the least possible amount of food intake may have made the
difference between life and death, not only for the individual but also more
importantly for the species. Those who could store fat easily had an evolutionary
advantage in the harsh environment of early hunters and gatherers.
The esthetic value and cultural significance attached to obesity is reflected in the
mysterious nude female figurines of Stone Age Europe, dating back to more than
20,000 years ago, considered to be matriarchal icons of fertility or the mother
goddess. The best known of these earliest representations of the human form is the
one discovered in Willendorf, Australia in 1908. Commonly known as the Venus of
Willendorf, its squat body, bulbous contours, pendulous breasts, and prominent belly
are as esthetically a factual rendering of gross obesity as can be.
Problem Solving
Obesity can lead to various health problems such as diabetes, heart disease and stroke.
Your school has decided to carry out an “Obesity Awareness Campaign” with the
aim to create awareness among students about obesity related health problems.
In this campaign, you are required to obtain the height and weight of students in your
class.
A.
C.
x x2 f x - x̄ fx2 (x - x̄ )2 f (x - x̄ )2
44.5 1980.25 6 -20 11881.50 400 2400
54.5 2970.25 11 -10 32672.75 100 1100
64.5 4160.25 3 0 12480.75 0 0
74.5 5550.25 5 10 27751.25 100 500
84.5 7140.25 1 20 7140.25 400 400
94.5 8930.25 1 30 8930.25 900 900
104.5 10920.25 2 40 21840.50 1600 3200
114.5 13110.25 1 50 13110.25 2500 2500
E. Calculation
1. Mean, x̄
267 599.5 193.5 372.5 84.5 94.5 209 114 .5 1935
30 30
= 64.5
2. Median
N 1 C. f .
M L i
f
15 6
M 50 9
11
=57.36
3. Mode
=54.5
4. Best Measure of Tendency
The mean is the best measure of tendency because it uses all values in the data set to
give an average values.
5. Standard Deviation, σ
1st method
f ( x x)
2
f
11000
30
19.14
2nd method
fx
2
(x)2
f
4526.92 4160.25
19.14
F. Conclusion
From my result, the mean is 64.5 and the standard deviation is 11.55. So, it
indicated that the dispersion is not too wide from the value of the mean obtained.
From the mean and standard deviation, the heaviest is 114 kg and the lightest is
44 kg. So, it indicated that the weight of the student not extremely heavy nor
light.
Further Exploration
Further Exploration:
The body mass index (BMI) gives and indication of the physical state of a person as
being underweight, normal, overweight or orbese. BMI can be calculated by using the
following formula:
weight ( kg )
BMI
height (m) height (m)
The table below shows the BMI and the corresponding physical state of a person:
BMI Category
Below 18.5 Underweight
18.5 - 24.9 Normal
25 - 29.9 Overweight
30 and above Obese
From my survey, I can conclude that 11 students are underweight, 13 are normal,
4 are overweight and 2 students are obese. So, 36.67% of the students are
underweight, another 43.33% are normal, 13.33% are overweight and 6.67% are
obese. The highest BMI score is 32.67 while the lowest BMI score is 15.57.
V) Student Role
Obesity has been proven to be crucial factor of many chronic diseases which
affect human lives every single day. Such as, heart diseases, fatty liver and diabetes.
So, we need to make take extra precautionary steps in order to prevent ourselves from
becoming obese. A student’s role in avert obesity is through the students themselves.
A student would first need to implement good eating habits and prevent themselves
from oily and snack foods which will leads to obesity. This implementation would
also attract other students and peers to practice good eating habits to avoid this
problems. Students would also need to get fit by doing exercise and workout to
release sweat and body acid. Dieting is a factor of good health and makes our fitter
and has better body mass.. Lastly, students would also need to go for regular check-
ups with doctors to know our initial body weight and body problems. If a student can
carry out these steps, obesity can prevented and healthy lifestyles will be the new
normal for human beings.
VI) Conclusion
Based on this project, I have concluded that this project has made me understand
statistics much more and I have understood the importance of statistics in making
choices in daily life.
From my survey, 6.67% of the students are obese. 13.33% of students are
overweight, 43.33% of students are normal and lastly underweight are 36.67% of the
students. This shows that most of the students in my survey are either underweight or
normal category. This shows a good result and this project is a complete success.
Most of the students in the normal category. The charts shows a low result of
obesity. Underweight students are quite high and these students will face many
negative effects. This stastistics may help the students on how to make their decision
by looking at this BMI graph.
Students who are underweight most probably are cause by stress and skipping
their meal that makes them thinner and malnutrition. Their body will start to not
function properly. Students in this category should try to increase their weight by
meeting a nutritionist. These people are much more experienced in this field and will
be able to give a good suggestion and advice on dieting and the right food intake in
daily basis. These students should also try not to take empty calories or junk food as it
would not help in the process. These students should also consider high protein meats
which can help build stronger muscles.
As for the overweight and obese students, they should prevent from taking junk
food and oily food as it will contribute to weight gain which will end up in obesity.
Whereas some students may think that skipping their meal will help them in reducing
their weight.This is not a proper way of dieting. Make sure that every meal taken is
low calories and has good protein. This will help in dieting. Obesity will increase the
student’s risk of chronic diseases such as heart disease and diabetes. They should
exercise more frequently to burn off the extra fat. Meeting a dietition will help those
obese students in how to eat a healthy meal instead of skipping meal. Not just that,
this profession will help them to give a suggestion on how to diet by exercising and
doing workouts.