You are on page 1of 299

LESSON - 1

INTRODUCTION TO BUSINESS RESEARCH METHODS

Research - Meaning - Types - Nature and scope of research - Problem formulation statement of research objective - value and cost of information - Decision theory Organizational structure of research - Research process - research designs - exploratory
-Descriptive - Experimental research.
OBJECTIVE
To equip the students with the basic understanding of the research methodology and
provide an insight into the application of modern analytical tools and techniques for the
purpose of management decision making.
STRUCTURE

Value of Business Research


Scope of Research
Types of Research
Structure of Research

LEARNING OBJECTIVES

To understand the importance of business research as a management decision


making tool
To define business research
To understand the difference between basic and applied research
To understand when business research in needed and when it should be
conducted
To identify various topics for business research.

INTRODUCTION
The task of business research is to generate accurate information for use in decision
making. The emphasis of business research is shifting the decision makers from
intuitive information that is based on own judgment and gathering information into
systematic and objective investigation.
DEFINITION
The business research is defined as the systematic and objective process of gathering,
recording and analyzing data for aid in making business decisions. Literally, research
means to "search again". It connotes patient study and scientific investigation where in
the research takes more careful look to discover to know about the subject of study. The

data collected and analyzed are to be accurate, the research need to be very objective.
Thus, the role of researcher is to be detached and impersonal rather than engaging in
biased attempt. This means without objectivity the research is useless. The definition is
restricted to take decision in the aspects of business alone. This generates and provides
the necessary qualitative and quantitative information upon which, as base, the
decisions are taken. This information reducing the uncertainly of decisions and reduces
the risk of making wrong decisions. However research should be an "aid" to managerial
judgment, not a substitute for it. There is more to management than research. Applying
research remains a managerial art.
The study of research methods provides with the knowledge and skills that need to solve
the problems and meet the challenges of the fast-paced decision making environment.
There are two important factors stimulate an interest in scientific approach to decision
making:
1. The is an increased need for more and better information, and
2. The availability of technical tools to meet this need.
During the last decade, we have witnessed dramatic changes in the business
environment. These changes have created new knowledge needs for the manager to
consider when evaluating any decision. The trend toward complexity has increased the
risk associated with business decision, making it more important to have sound
information base. The following are the few reasons which makes the researcher to
lookout for newer and better information based on which the decisions are taken:

There are more variables to consider on every decision


More knowledge exists in every field of management
The quality and theory and models to explain the tactics and strategic results are
improving.
Better arrangement of information
Advancement in computer allowed to create better database.
The power and ease of use of today's computer have given to capability to analyze
the data to solve managerial problems

The development of scientific method in business research lags behind the similar
developments in physical science research which is more rigorous and much more
advanced. But business research is of recent origin and moreover the finding cannot be
patented that of physical science research. Business research normally deals with topics
such as human attitudes, behavior and performance. Even with these hindrances,
business research is making strides in the scientific arena. Hence, the managers who are
not proposed ior this scientific application in business research will be at severe
disadvantage.
Value of Business Research

The prime value of business research is that it reduces uncertainty by providing


information that improves decision making process. The decision making process is
associated with the development and implementation of a strategy, which involves the
following
1. Identifying problems or opportunities
Before any strategy can be developed, an organization must determine where it wants to
go and how it will get there. Business research can help managers to plan strategies by
determining the nature of situations or by identifying the existence of problems of
opportunities that are present in the organization. Business research may be used as a
scanning activity to provide information about what business happening or its
environment. Once it defines and indicates problems and opportunities, managers may
evaluate alternatives very easily and clear enough to make a decision.
2. Diagnosing and Assessing problems and opportunities
The important aspect of business research is the provision of diagnostic information
that clarifies the situation. It there is a problem, they need to specify what happened and
why. If an opportunity exists, they need to explore, clarify and refine the nature of
opportunity. This will help in developing alternative courses of action that are practical.
3. Selecting and implementing a course of action
After the alternative course of action has been clearly identified, business research is
conducted to obtain scientific information which will aid in evaluating the alternatives
and selecting the best course of action.
Need for Research
When a manager faced with two or more possible course if action, the researcher
carefully need to take decision whether or not to conduct the research. The following are
the determinants.
1. Time Constraints
In most of the business environment, the decisions most must be made immediately,
but conducting research systematically takes time. There will not be much time to relay
on research. As a consequence, the decisions are sometimes made without adequate
information and through understanding of the situation.
2. Availability of Data
Often managers possess enough information with no research. When they lack adequate
information, research must be considered. The managers should think whether the
research will be able to generate information needed to answer the basic question about
which the decision is to be taken.

3. Nature of Information
The value of research will depend upon the nature of decisions to be made. A routine
decision does not require substantial information or warrants. However, for important
and strategic decision, more likely research needs to be conducted.
4. Benefits vs. Costs
The decision to conduct research boils down to these important questions.
1. Will the rate of return be worth the investment?
2. Will the information improve the quality of the decision?
3. Is the research expenditure the best use of available funds?
Thus, the cost of information should not exceed the benefits i.e. value of information.
What is Good Research?
Good research generates dependable data can be used reliable for making managerial
decisions. The following are the tips of good research.

Purpose clearly defined, i.e., understanding problems clearly


The process described in sufficient details
The design carefully planned to yield results
Careful consideration must be given and maintain high ethical standards
Limitations properly revealed
Adequate analysis of the data and appropriate tools used
Presentation of data should be comprehensive, early understood and presented
unambiguously
Conclusion should base on the data obtained and justified.

Scope of Research
The scope of research on management is limited to business. A researcher conducting
research within an organization may be referred as a "marketing researcher" or
"organizational researcher", although business research is specialized and the term
encompasses all the functional areas Production, Finance, Marketing, HR etc. The
different functional areas may investigate different phenomenon, but they are
comparable to one another because they use similar research methods. There are many
kinds of areas are resembled in the business environment like forecasting, trends
environment, capital formation, portfolio analysis, cost analysis, risk analysis, TQM, job
satisfaction, organizational effectiveness, climate, culture, market potential,
segmentation, sales analysis, distribution channel, computer information needs
analysis, social values and establish and etc.

Types of Research
Research is to develop and evaluate concepts and theories. In broader sense research
can be classified as.
1) Basic Research or Pure Research
It does not directly involve the solution to a particular problem. Although basic research
generally cannot be implemented, this is conducted to verify the acceptability of a given
theory or to discuss more about a certain concept.
2) Applied Research
It is conducted when a decision must be made about a specific real-life problem. It
encompasses those studies undertaken to answer question to specific problems or to
make decision about particular course of action.
However, the procedures and techniques utilized by both researchers do not differ
substantially. Both employ scientific method to answer questions. Broadly, the scientific
method refers to techniques and procedures that help researcher to know and
understand business phenomenon. The scientific method requires systematic analysis
and logical interpretation of empirical evidence (facts from observation or
experimentation) to confirm or dispose prior conceptions. In basic research, it first tests
the prior conceptions or assumptions or hypothesis and then makes inferences and
conclusions. In the applied research the use of scientific method assures objectivity in
gathering facts and taking decision.
At the outset, it may be noted that there are several ways of studying and tackling a
problem. There is no single perfect design. Research designs have been classified by
authors in different ways. Different types of research designs have emerged on the
account of the different perspectives from which the problem or opportunity is viewed.
However, the research designs broadly classified into three categories - exploratory,
descriptive, and causal research. The research can be classified on the basis of either
technique or function. Experiment, surveys and observation are few common
techniques. The technique may be qualitative of quantitative. Based on the nature of the
problems or purpose of study the above three are used invariably used in management
parlance.
3) Exploratory Research
The focus is mainly on discovering of ideas. An exploratory research is generally based
on secondary data that are already available. It is to be understood that this type of
study is conducted to classify ambiguous problems. These studies provide information
to use in analyzing situations. This will helps to crystallize a problem and identify
information needs for further research. The purpose of exploration is usually to develop
hypotheses or question for further research. The exploration may be accomplished with
different techniques. Both qualitative and quantitative techniques are applicable

although exploration studies relies on more heavily a qualitative technique like


experience survey, focus group.
4) Descriptive Research
The major purpose is to describe characteristics of a population or phenomenon.
Descriptive research seeks to determine to answers to who, what, when, where and how
questions. Unlike explorative, these studies are based on some previous understanding
of the nature of the research problem. Descriptive studies can be divided into two broad
categories-Cross sectional and Longitudinal. The former type is more frequently used. A
cross section study is concerned with the sample of elements from a given population. It
is carried out once and represents one point at time. The longitudinal studies are based
on panel data or panel methods. A panel is sample of representatives who are
interviewed and then re-interviewed from time to time, that is, longitudinal studies are
repeated over an extended period.
5) Causal Research
The main goal of causal research is the identify cause and effect relationship among
variables. It attempts to establish that when we do one thing what another thing will
happen. Normally explorative and descriptive studies precede causal research.
However, based on the breadth and depth of study, another method is frequently used is
management called case study research. This places more emphasis on a full contextual
analysis of fever events or conditions and their interrelations. An emphasis on details
provides valuable insight for problem solving, evaluation and strategy. The detail is
gathered from multiple sources of information.
Value of Information and Cost
Over the part decade many cultural, technological are competitive factors have created a
variety of new challenges, problems and opportunities for today's decision makers in
business. First, the rapid advances in interactive marketing communication
technologies have increased the need for database management skills. Moreover,
advancements associated with the so called information super high ways have created
greater emphasis on secondary data collection, analysis and interpretation. Second,
there is a growing movement emphasizes quality improvements. This placed more
importance on cross sectional information then over before. Third is the expansion of
global markets which introduces a new set of multicultural problem and question.
These three factors that influence the research process and it take steps into seeking new
information in management perspective. There may be situations where management is
sufficiently clear that no additional information is likely to change its decision. In such
cases, it is obvious that the value of information is negligible. In contrast, there are
situations where the decisions look out for information which is not available easily.
Unless the information collected does not led to change or modify a decision, the
information has no value. Generally information is most valuable in cases i) where there

is unsure of what is to be done and ii) where extreme profits or losses involved. A
pertinent question is-how much information should be collected in a given situation?
Since the collected information involves a cost, it is necessary to ensure that the benefit
from the information is more than the cost involved in its collection.
Decision Theory
With reference to the above discussion, an attempt is needed to see how information can
be evaluated for setting up a limit. The concept of probability is the basis of decision
maker under conditions of uncertainly. These are three basic sources of assigning
probabilities
1) Based on a logic / deduction: For e.g., when a coin is tossed, the probability of
getting a head or tail is 0.5.
2) Past experience / Empirical evidence: The experience gained in the process
resolving these problems in the past. On the basis of its past experience, it may be in a
better positive to estimate the probability of new decisions.
3) Subjective Estimate: The most frequently used method, it is based on the
knowledge and information with respect to researcher for the probability estimates.
The above discussion was confined to single stage problem wherein the researcher is
required to select the best course of action on the basis of information available at a
point at time. However, there are problems with multiple stages wherein a sequence of
decisions involved. Each decision leads to a chance event which in turn influences the
next decision. In those cases, a Decision Tree Analysis i.e. graphical derives depicting
the sequence of action-event combination, will be useful in making a choice between two
alternatives. If the decision tree is not helpful, more sophisticated technological known
as Bayesian Analysis can be used. Here, the probabilities can be revised on account of
the availability of new information using prior, posterior and pre-posterior analysis.
There is a great deal between budgeting and value assessment in management decision
to conduct research. An appropriate research study should help managers avoid losses
and increase sales or profits; otherwise, research can be wasteful. The decision maker
wants a cost-estimate for a research project and equally precise assurance that useful
information will result from the research. Even if the researcher can give good cost and
information estimates, the decision maker or manager still must judge whether the
benefits out-weigh the costs.
Conceptually, the value of research information is not difficult to determine. In business
situation the research should provide added revenues or reduce expenses. The value of
research information may be judge in terms of the difference between the results of
decisions made with the information and the result that would be made without it. It is
simple to state, in actual application, it presents difficult measurement problems.
Guideline for approximately the cost-to-value of Research

1. Focus on the most important issues of the project: Identify certain issues as
important and others as peripheral to the problems. Unimportant issues are only to
drain resources.
2. Never try to do much: There is limit to the amount of information that can be
collected. The researcher must take a trade-off between the number of issues that can be
dealt with the depth of each issue. Therefore it is necessary to focus on those issues of
greatest potential value.
3. Determine whether secondary, primary information or combination is
needed: The most appropriate must be selected that should address the stated
problem.
4. Analyze all potential methods of collecting information: Alternative data
sources and research designs are available that will allow detailed investigation of issues
at a relatively low cost.
5. Subjectively asses the value of information: The researcher need to ask some
fundamental questions relating to objections. For example, a) Can the information be
collected at all? b) Can the information tell something more that already what we have?
c) Will the information provide significant insights? d) What benefits will be delivered
from this information?
Structure of Research
Business research can take many forms, but systematic inquiry is a common thread.
Systematic inquiry requires an orderly investigation. Business research is a sequence of
highly interrelated activity. The steps research process overlap continuously.
Nevertheless, research on management often follows a general pattern. The styles are:
1. Defining the problem
2. Planning a research design
3. Planning a sample
4. Collecting data
5. Analyzing the data
6. Formulating the conclusions and preparing the report
SUMMARY
This paper outlined the importance of business research. Difference between basic and
applied research have been dealt in detail. This chapter has given the meaning, scope,
types, and structure of the research.

KEY TERMS

Research
Value of research
Need for research
Good Research
Scope of research
Types of research
Basic research
Applied research
Scientific method
Exploratory research
Descriptive research
Cross - sectional and longitudinal
Causal research
Decision theory
Structure

QUESTIONS
1. What are some examples of business research in your particular field of interest?
2. What do you mean by research? Explain its significance in modem times.
3. What is the difference between applied and basic research?
4. What is good research?
5. Discuss: Explorative, descriptive and causal research.
6. Discuss the value of information and cost using decision theory.
7. Discuss the structure of business research.

- End of Chapter LESSON - 2


PROBLEM DEFINITION

OBJECTIVES

To discuss the nature of decision makers objectives and the role they play in
defining the research
To understand that proper problem definition is essential for effective business
research
To discuss the influence of the statement of the business problem on the specific
research objectives
To state research problem in terms of clear and precise research objectives.

STRUCTURE

Problem definition
Situation analysis
Measurable Symptoms
Unit of analysis
Hypothesis and Research objectives

PROBLEM DEFINITION
Before choosing a research design, manager and researcher need a sense of direction for
the investigation. It is extremely important to define the business problem carefully
because the definition determines the purposes of the research and ultimately the
research design. Well defined problem is half solved problems - hence, researcher must
understand how to define problem. The formal quantitative research process should not
begin until the problem has been clearly defined. Determination of research problem
consists of three important tasks namely,
1. Classifying in argument information needs.
2. Redefining research problem, and
3. Establishing hypothesis & research objectives.
Step 1: To ensure that appropriate information created through this process, researcher
must assist decision maker in making sure that the problem or opportunity has been
clearly defined and the decision maker is aware of the information requirements. This
include the following activities namely,
i) Purpose: Here, the decision maker holds the responsibility of addressing a
recognized decision problem or opportunity. The researcher begins the process by
asking the decision maker to express his or her reasons for thinking there is a need to
undertake research. By this questioning process, the researcher can develop insights as
what they believe to be the problems. One method that might be employed to familiarize
the decision maker is the iceberg principle. The dangerous part of many problems,
like submersed portion of the iceberg, is neither visible nor understood by managers. If
the submerged position of the problem is omitted from the problem definition, then
result may be less than optimal.
ii) Situation Analysis or understanding the situation: To gain the complete
understanding, both should perform a basic situation analysis of the circumstances

surrounding the problem area. A situational analysis is a popular tool that focuses on
the informal gathering of background information to familiarize the overall complexity
of the decision. A situation analysis attempts to identify the event and factors that have
led to the current decision problem situation. To objectively understand the client's
domain (i.e., industry, competition, product line, markets etc) the researchers not rely
only on the information provided by client but also others. In short the researcher must
develop expertise in the client's business.
iii) Identifying and separating measurable symptoms: Once the researcher
understands the overall problem situation, they must work with decision maker to
separate the problems from the observable and measurable symptoms that may have
been initially perceived as the being the decision problem.
iv) Determining unit of analysis: The researcher must be able to specify whether
data should be collected about individual, households, organizations, departments,
geographical areas, specific object or some contributions of these. The unit of analysis
will provide direction in later activities such as scale measurement development and
drawing appropriate sample of respondents.
v) Determining relevant variables: Here, the focus is on the identifying the
different independent or dependent variables. It is determination of type of information
(i.e., facts, estimates, predictions, relationships) and specific constructs (i.e. concepts or
ideas about an object, attributes, or phenomenon that are worth measurement)
Step 2:
Once the problem is understood and specific information requirements are identified,
then the researcher must redefine the problem in more specific terms. In reframing the
problems and questions as information research questions, they must use their scientific
knowledge expertise. Establishing research questions specific to problems will force the
decision maker to provide additional information that is relevant to the actual problems.
In other situations, redefining problems as research problems can lead to the
establishment of research hypothesis rather than questions.
Step 3: (Hypothesis & Research objective)
A hypothesis is basically an unproven started of a research question in a testable
format. Hypothetical statement can be formulated about any variable and can express a
possible relationship between two or more variables. While research questions and
hypotheses are similar in their intent to express relationship, the hypotheses tend to be
more specific and declarative, whereas research questions are more interrogative. In
other words hypotheses are statement that can be empirically tested.
Research objectives are precise statements of what a research project will attempt to
achieve. It indirectly represents a blueprint of research activities. Research objectives
allow the researcher to document concise, measurable and realistic events that either
increase or decrease the magnitude of management problems. More importantly it

allows for the specification of information required to assist management decision


making capabilities.
SUMMARY
Nature of decision makers objectives, and the role they play in defining the research
have been dealt in detail. This chapter has given the steps involved in defining the
problem.
KEY TERMS

Problem definition
Iceberg principle
Situation analysis
Unit of analysis
Variables
Hypotheses
Research objectives

QUESTIONS
1. What is the task of problem definition?
2. What is the iceberg principle?
3. State a problem in your field of interest, and list some variables that might be
investigated to solve this problem.
4. What do you mean by hypothesis?
5. What is a research objective?

- End of Chapter LESSON 3


RESEARCH PROCESS AND DESIGN

OBJECTIVES

To list the stages in the business research process


To identify and briefly discuss the various decision alternatives available to the
researcher during each stage of the research process.

To classify business research as exploratory, descriptive, and causal research


To discuss categories of research under exploratory, descriptive and causal
research.

STRUCTURE

Research process
Research design
Types of research designs
Explorative research
Descriptive research
Casual research

Research Process
Before discussing the phases and specific steps of research process, it is important to
emphasize the need for information and when the research is conducted or not. In this
context, the research process may be called as information research process that would
be more appropriate in business parlance. The information research is used to reflect
the evolving changes occurring within the management and the rapid changes facing
many decision makers regarding how firms conduct both internal and external
activities. Hence, understanding the process of transforming raw data into usable
information from broader information and expanding the applicability of the research
process in solving business problems and opportunities is very important.
Overview: The research process has been described anywhere from 6 to 11
standardized stages. Here, the process consist of four distinct inter related phases that
have logical, hierarchical ordering depicted below.
Diagram: Four phases of Research Process

However, each phase should be viewed as a separate process that consists of


combination of integrated steps and specific procedures. The four phases and
corresponding step guided by the principles of scientific method, which involves
formalized research procedures that can be characterized as logical, objective,
systematic, reliable, valid and ongoing.
Integrative steps within research process

The following exhibit represents the interrelated steps of the 4 phases of the research
process. Generally researchers should follow the steps in order. However, the
complexity of the problem, the level of risk involved and management needs will
determine the order of the process.
Exhibit
Phase 1: Determination of Research Problem
Step 1: Determining management information needs
Step 2: Redefining the decision problem as research problem.
Step 3: Establishing research objectives.
Phase 2: Development of Research Design
Step 4: Determining to evaluate research design.
Step 5: Determining the data source.
Step 6: Determining the sample plan and sample size.
Step 7: Determining the measurement scales.
Phase 3: Execution of the Research Design
Step 8: Data collection and processing data.
Step 9: Analysing the data.
Phase 4: Communication of the Results
Step 10: Preparing and presenting the final report to management.
Step 1: Determining management information needs
Before the researcher becomes involved usually, the decision maker has to make a
formal statement of what they believe is the issue. At this point, the researcher's
responsibility is to make sure management has clearly and correctly specified the
opportunity or question. It is important for the decision maker and the researcher to
agree on the definition of the problem so that the result of the research process will
produce useful information. Actually the researcher should assist the decision maker in
determining whether the referred problem is really a problem or just a symptom or a yet
unidentified problem. Finally the researchers list the factors that could have a direct or
indirect impact on the defined problem or opportunity.

Step 2: Redefining the decision problem as research problem.


Once the researcher and decision makers have identified the specific information needs,
the researcher must redefine the problem in scientific terms since the researcher feel
more comfortable using a scientific framework. This is very critical because, it influences
many other steps. It is the researcher's responsibility to state the initial variables
associated with the problem in the form of one or more question formats (how, what,
where, when or why). In addition, the researcher need to focus on determining what
specific information is required (i.e., facts, estimates, predictions, relationships or some
combination) and also of quality of information which includes the value of information.
Step 3: Establishing research objectives.
The research objective should follow from the definition of research problem established
in Step 2. Formally stated objective provides the guidelines for determining other steps
to be undertaken. The undertaking assumption is that, if the objectives are achieved, the
decision maker will have information to solve the problem.
Step 4: Determining to evaluate research design.
The research design serves as a master plan of methods and procedures that should be
used to collect and analyze the data needed by the decision maker. In this master plan,
the researcher must consider the design technique (survey, observation, and
experiment), the sampling methodology and procedures, the schedule and the budget.
Although every problem is unique, but most of the objectives can be met using one of
three types of research designs: exploratory, descriptive, and casual. Exploratory focuses
on collecting either secondary or primary data and using unstructured formal or
informal procedures to interpret them. It is often used simply classify problems or
opportunity and it is not intended to provide conclusive information. Some examples of
exploratory studies are focus group interview, expensive surveys and pilot studies.
Descriptive studies that describe the existing characteristic which generally allows to
draw inference and can lead to a course of action. Causal studies are designed to collect
information about cause and effect relationship between two or more variables.
Step 5: Determining the data source.
This can be classified as being either secondary or primary. Secondary data can usually
be gathered faster and at less cost than primary data. Secondary data are historical data
previously collected and assembled for some research problem other than the current
situation. In contrast primary data represent firsthand data, and yet to have meaningful
interpretation and it employs either surveyor observation.
Step 6: Determining the sample plan and sample size.
To make inference or prediction about any phenomenon we need to understand where
or who is supplying the raw data and how representative those data are. Therefore,
researchers need to identify the relevant defined target population. The researcher can

choose between a sample (small population) and census (entire population). To achieve
this research objective, the researcher needs to develop explicit sampling plan which will
serve as a blueprint for defining the target population. Sampling plans can be classified
into two general types: probability (equal chance) and non probability. Since sampling
size affects quality and general ability, researchers, must think carefully about how
many people to include or how many objects to investigate.
Step 7: Determining the measurement scales.
This step focuses on determining the dimensions of the factors being investigated and
measuring the variables that underlie the defined problem. This determines how much
raw data can be collected and the amount of data to be collected. The level of
information (nominal, ordinal, interval, and ratios), the reliability, the validity and
dimension (uni vs. multi) will determine the measurement process.
Step 8: Data collection and processing data.
There are two fundamental approaches to gather raw data. One is to ask questions about
variables and phenomena using trained interviewers or questionnaires. The other is to
observe variables or phenomena using professional observers or high tech mechanical
devices. Self - administered surveys, personal interviews, computer simulations,
telephone interviews are some of the tools to collect data. The questioning allows a
wider variety of collective of data about not only past, present but also the state of mind
or intentions. Observation can be characterized as natural, contrived, disguised or
undisguised, structured or unstructured, direct or indirect, human or mechanical, and
uses the devices like video camera, tape recorders, audiometer, eye camera, psychogalvanometer or pupil meter. After the raw data collected, a coding scheme is needed so
that the raw data can be entered into computers. It is assigning logical numerical
description to all response categories. The researcher must then clean the raw data of
either coding or data entry error.
Step 9: Analysing the data.
Using a variety of data analysis technique, the researcher can create new, complex data
structure by continuing two or more variables into indexes, ratios, constructs and so on.
Analysis can vary from simple frequency distribution (percentage) to sample statistic
measures (mode, median, mean, standard deviation, and standard error) to multivariate
data analysis.
Step 10: Preparing and presenting the final report to management.
This step is to prepare and present the final research report to management. The report
should contain executive summary, introduction, problem definition and objectives,
methodology, analysis, results and finding, finally suggestions and recommendation. It
also includes appendix. Any researcher is expected not only submit well produced
written report but also oral presentation.

Research Design
Kerlinger, in his "Foundations of Behavioral Research" book defines, "Research design
is the plan structure, and strategy of investigation conceived so as to obtain answers to
research questions and to control variance". The plan is overall scheme or program of
the research. It includes an outline of what the investigator will do from writing
hypotheses and their operational implication to the final analysis of data. A structure is
the framework, and the relations among variables of a study. According to Green & Tull,
a research design is the specification of methods and procedures for acquiring the
information needed. It is the overall operational pattern or framework of the project
that stipulates what information is to be collected from which sources by what
procedures.
From the above definitions it can be understood that the research design is more or less
a blueprint of research, which lays down the methods and procedure for requisite
collection of information and measurement and analysis with a view to arrive at
meaningful conclusions of the proposed study.
Types of Research Design
The different types of design are explained in the previous section (refer types of
research). There are three frequently used classification is give below.
I. Explorative
II. Descriptive
III. Casual
Here the focus will be how these studies are conducted and methods are explained:
I. Categories of Explorative Research
There are four general categories of explorative research methods. Each category
provides various alternative ways of getting information.
1. Experience Surveys
It is an attempt to discuss issues and ideas with top executive and knowledge
people who have experience in the field. This research in the form of experience
survey may be quite informal. This activity intends only to get ideas about the'
problems. Often an experience survey may consist of interviews with a small
number of people who have been carefully selected. The respondents will
generally be allowed to discuss the questions with few constraints. Hence, the
purpose of such experts is to help formulate the problem and classify concepts
rather than develop conclusive evidence.

2. Secondary Data Analysis


Another quick source of background information is trade literature. Using
secondary data may equally important to applied research. Investigating date
that has been completed for some purpose other than the present one is one of
the frequent forms of exploratory research. Also, it is to remember that this
method often used in descriptive analysis.
3. Case Studies
It is to obtain information one or few situation that are similar to the present one.
The primary advantage of case study is that entire entity can be investigated in
depth and with meticulous attention in details. The results from this type should
be seen as tentative. Generalizing from a few cases can be dangerous, because
more situations are not typical in same sense. But even if situations are not
directly comparable, a number of insights can be gained and hypothesis
suggested for future research.
4. Pilot Studies
In the context of exploratory research, a pilot study implies that some aspect of
the research will be on a small scale. This generates primary data usually for
qualitative analysis. The major categories are discussed below:
a. Focus Group Interview
The popular method in the qualitative research is an unstructured free-flowing
interview with a small group of people. It is not rigid, but flexible promote that
encourages discussions. The primary advantages are that they are relatively brief,
easy to execute, quickly analyzed and inexpensive. However, a small group will
rarely be a representative sample, no matter how carefully it is selected.
b. Projective Techniques
It is an indirect means of questioning that enable the respondent to project
beliefs and feeling onto a third party, an inanimate object or a situation.
Respondents are not required to provide answer to a structural format. They are
encouraged to describe a situation in their own words, with little prompting by
the researcher, within the context of their own experiences, attitudes, personality
and to express opinions and emotions. The most common techniques are used
associations, sentence completion, Thematic Apperception Test (TAT) and role
playing.
c. Depth interview
It is similar to focus group but in the interviewing session the researcher asks
many questions and probes for elaboration. Here the role of researcher

(interviewers) is more important. He must be highly skillful who can influence


respondents to talk freely without disturbing the direction. It may last more
hours, hence it is expensive.
II. Categories of Descriptive Research
In contrast to exploratory, descriptive studies are more formalized and typically
structured with clearly stated hypothesis. When the researcher interested in knowing
the characteristics of certain groups such as age, sex, educational level, occupation or
income, a descriptive study may be necessary. Descriptive studies can be divided into
two broad categories - cross sectional and longitudinal. The following are the methods
used in descriptive studies:
1. Secondary Data Analysis
2. Primary Data Analysis
3. Case studies
Several methods are available to collecting the information (i.e., observation,
questionnaire, and examination of records with the merits and limitations, the
researcher may use one or more of these methods which have been discussed in details
in later chapters. Thus the descriptive studies methods must be selected keeping in view
the objectives of the study and the resources available. The said design can be
appropriately referred to as a survey design using observing or questioning process, lit.
III. Categories of Causal Research
As the name implies, a causal design investigates the cause and effect relationship
between two or more variables. The causal studies may be classified as informal and
formal or quasi / true and complex designs. The methods used in experimental research
are discussed hereunder:
1. The one-shot case study (after-only design)
2. Before-after without control group.
3. After-only with control group.
4. Before-after with one control group
5. Four-group, Six- study design (Solomon four group design)
6. Time series design
7. Completely randomized design

8. Randomized block design.


9. Factorial design
10. Latin square design
The first two methods are called as quasi experimental designs, next 3 methods are
called as true experimental designs and last 4 methods are called complex designs. In
this experimental design the following symbols are used in describing the various
experimental designs:
X = Exposure of a group to an experimental treatment.
O = Observation or measurement of the dependent variable.
1. The one-shot case study design (after only design): The one-shot
design, there should be a measure of what would happen when test units were not
exposed to X to compare with the measure when subjects were exposed to X. This
is diagrammed as follows:
X

O1

2. Before - after without control group design: In this, the researcher is


likely to conclude that the difference between O2 and O1 (O2 O1) is the measure of
the influence of the experimental treatment. The design is as follows:
O1

O2

3. After only with control group design: The diagram is as follows:


Experimental Group:
Control Group:

X
X

O1
O2

The design is to randomly selected subjects and randomly assign to experimental


or the control group. The treatment is then measured in both groups at same time.
The effect is calculated as follows:
O2 - O1
4. Before-after with one control group design: This is explained as follows:
Experimental Group:
Control Group:

O1
O3

X
X

O2
O4

As the diagram above indicated, the subjects of the experimental group are tested
before and after being exposed to the treatment. The control group is tested twice,
at the same time of experimental group, but these subjects are not exposed to the
treatment. The effect is calculated as follows:
(O2 - O1) - (O4 O3)
5. Four-group, six- study design: (Solomon four groups design) Combining,
the before - after with control group design and the after-only with control group
design provides a means for controlling testing affects, as well as other sources of
extraneous variations. The diagram as follows:
Experimental Group 1
Control Group 1

O1

O3

Experimental Group 2:

O4

Control Group 2:

O2

O5
O6

6. Time series design: When experiments are conducted over long periods of
time, they are more vulnerable to history effects due to changes in population,
attitudes, economic patterns and the like. Hence, this is also called quasiexperimental design. This design can be diagrammed as follows:
O1

O2

O3

O4

O5

O6

Several observations have been taken before and after the treatment to determine
the patterns after the treatment are similar to the pattern of before the treatments.
7. Completely randomized design (CRD): CRD is an experimental design
that uses a random process to assign experimental units to treatments. Here,
randomization of experimental units to control extraneous variables while
manipulating a single independent variable, the treatment variable.
8. Randomized block design (RBD): The RBD is an extension of the CRD. A
form of randomization is utilized to control for most extraneous variation. Here an
attempt is made to isolate the effects of the single variable by blocking its effects.
9. Factorial design: A FD allows for testing the effects of two or more
treatment (factors) at various levels. It allows for the simultaneous manipulation
of 2 or more variable at various levels. This design will measure main effect (i.e.,
the influence on the dependent variable by each independent variable and also
interaction effect.

10. Latin Square Design (LSD): The LSD attempts to control or block out the
effect of two or more confounding extraneous factors. This design is so named
because of the layout of the table that represents the design.
SUMMARY
This chapter has outlined the stages in business research prawn. Various types of
research design have been dealt in detail.
KEY TERMS

Exploratory research
Descriptive research
Causal research
Focus group
Projective techniques
Case studies
Depth Interview
Experimental designs
Quasi experimental design
True experimental designs
Complex designs

QUESTIONS
1. Explain the different phases of research process.
2. Briefly describe the different steps involved in a research process.
3. What are major types researches in business?
4. Discuss the categories of exploratory and descriptive research.
5. Explain different experimental designs.
REFERENCES
1. Bellenger and et al, Marketing Research, Home Wood Illinois, Inc. 1978.
2. Boot, John C.G. and Cox., Edwin B., Statistical Analysis for Managerial Decisions, 2nd
ed. New Delhi: McGraw Hill Publishing Co. Ltd.
3. Edwards, Allen, Statistical Methods, 2nd ed., New York. 1967.

- End of Chapter -

LESSON 4
INTRODUCTION TO STATISTICS AND POPULATION PARAMETERS

OBJECTIVES

Meaning and definition of Statistics


Nature of Statistical study
Importance of Statistics in business and also its limitations

STRUCTURE

Nature of statistical study


Importance of statistics in business
Statistical quality control method
Limitation of statistics

INTRODUCTION
At the outset, it may be noted that the word 'Statistics' is used rather curiously in two
senses-plural and singular. In the plural sense, it refers to a set of figures. Thus, we
speak of production and sale of textiles, television sets, and so on. In the singular sense,
Statistics refers to the whole body of analytical tools that are used to collect the figures,
organize and interpret them and, finally, to draw conclusions from them.
It should be noted that both the aspects of Statistics are important if the quantitative
data are to serve their purpose. If Statistics, as a subject, is inadequate and consists of
poor methodology, we would not know the right procedure to extract from the data the
information they contain. On the other hand, if our figures are defective in the sense
that they are inadequate or inaccurate, we would not reach the right conclusions even
though our subject is well developed. With this brief introduction, let us first see how
Statistics has been defined.
Statistics has been defined by various authors differently. In the initial period the role of
Statistics was confined to a few activities. As such, most of the experts gave a narrow
definition of it. However, over a long period of time as its role gradually expanded,
Statistics came to be considered as much wider in its scope and, accordingly, the experts
gave a wider definition of it.
Spiegal, for instance, defines Statistics, highlighting its role in decision-making
particularly under uncertainty, as follows:

"Statistics is concerned with scientific method for collecting, organising, summarising,


presenting and analysing data as well as drawing valid conclusions and making
reasonable decisions on the basis of such analysis."
This definition covers all the aspects and then tries to link them up with decisionmaking. After all, Statistics as a subject must help one to reach a reasonable and
appropriate decision on the basis of the analysis of numerical data collected earlier.
Using the term 'Statistics' in the plural sense, Secrist defines Statistics as
"Aggregate of facts, affected to a marked extent by multiplicity of causes, numerically
expressed, enumerated or estimated according to reasonable standards of accuracy,
collected in a systematic manner for a predetermined purpose, and placed in relation to
each other".
This definition of Secrist highlights a few major characteristics of statistics as given
below:
1. Statistics are aggregates of facts. This means that a single figure is not statistics.
2. Statistics are affected by a number of factors. For example, sale of a product depends
on a number of factors such as its price, quality, competition, the income of the
consumers, and so on.
3. Statistics must be reasonably accurate. Wrong figures, if analysed, will lead to
erroneous conclusions. Hence, it is necessary that conclusions must be based on
accurate figures.
4. Statistics must be collected in a systematic manner. If data are collected in a
haphazard manner, they will not be reliable and will lead to misleading conclusions.
5. Finally, statistics should be placed in relation to each other. If one collects data
unrelated to each other, then such data will be confusing and will not lead to any logical
conclusions. Data should be comparable over time and over space.
THE NATURE OF A STATISTICAL STUDY
Having briefly looked info the definition of Statistics, we should know at this stage as to
what the nature of a Statistical study is. Whether a given problem pertains to business or
to some other field, there are some well defined steps that need to be followed in order
to reach meaningful conclusions.
1. Formulation of the Problem: To begin with, we have to formulate a problem on
which a study is to be done. We should understand the problem as clearly as possible.
We should know its scope so that we do not go beyond it or exclude some relevant
aspect.

2. Objectives of the Study: We should know what the objectives of the proposed
study are. We should ensure that the objectives are not extremely ambitious or else the
study may fail to achieve them because of limitations of time, finance or even
competence of those conducting the study.
3. Determining Sources of Data: The problem and the objectives, thus properly
understood, will enable us to know as to what data are required to conduct the study.
We have to decide whether we should collect primary data or depend exclusively on
secondary data. Sometimes the study is based on both the secondary and the primary
data. When study is to be based on secondary data, whether partly or fully, it is
necessary to ensure that the data are quite suitable and adequate for the objectives of
the study.
4. Designing Data Collection Forms: Once the decision in favour of collection of
primary data is taken, one has to decide the mode of their collection. The two methods
available are: (i) observational method, and (ii) survey method. Suitable questionnaire is
to be designed to collect data from respondents in a field survey.
5. Conducting the Field Survey: Side by side when the data collection forms are
being designed, one has to decide whether a census surveyor a sample survey is to be
conducted. For the latter, a suitable sample design and the sample size are to be chosen.
The field survey is then conducted by interviewing sample respondents. Sometimes, the
survey is done by mailing questionnaires to the respondents instead of contacting them
personally.
6. Organising the Data: The field survey provides raw data from the respondents. It
is now necessary to organise these data in the form of suitable tables and charts so that
we may be aware of their salient features.
7. Analysing the Data: On the basis of the preliminary examination of the data
collected as well as the nature and scope of our problem, we have to analyse data. As
several statistical techniques are available, we should take special care to ensure that the
most appropriate technique is selected for this purpose.
8. Reaching Statistical bindings: The analysis in the preceding step will bring out
some statistical findings of the study. Now we have to interpret these findings in terms
of the concrete problem with which we started our investigation.
9. Presentation of Findings: Finally, we have to present the findings of the study,
properly interpreted, in a suitable form. Here, the choice is between an oral presentation
and a written one. In the case of an oral presentation, one has to be extremely selective
in choosing the material, as in a limited time one has to provide a broad idea of the
study as well as its major findings to be understood by the audience in proper
perspective. In case of a written presentation, a report has to be prepared. It should be
reasonably comprehensive and should have graphs and diagrams to facilitate the reader
in understanding it in all its ramifications.

IMPORTANCE OF STATISTICS IN BUSINESS


There is an increasing realisation of the importance of Statistics in various quarters.
This is reflected in the increasing use of Statistics in the government, industry, business,
agriculture, mining, transport, education, medicine, and so on. As we are concerned
with the use of Statistics in business and industry here, the description given below is
confined to these areas only.
Three major functions where statistics can be found useful in a business enterprise:
A. The planning of operations - This may relate to either special projects or to the
recurring activities of a firm over a specified period.
B. The setting up of standards - This may relate to the size of employment, volume
of sales, fixation of quality norms for the manufactured product, norms for the daily
output, and so forth.
C. The function of control - This involves comparison of actual production achieved
against the norm or target set earlier. In case the production has fallen short of the
target, it gives remedial measures so that such a deficiency does not occur again.
A point worth noting here is that although these three functions - planning of
operations, setting standards, and control-are separate, but in practice they are very
much interrelated.
Various authors have highlighted the importance of Statistics in business. For instance,
Croxton and Cowden give numerous uses of Statistics in business such as project
planning, budgetary planning and control, inventory planning and control, quality
control, marketing, production and personnel administration. Within these also they
have specified certain areas where Statistics is very relevant. Irwing W. Burr, dealing
with the place of Statistics in an industrial organisation, specifies a number of areas
where Statistics is extremely useful. These are: customer wants and market research,
development design and specification, purchasing, production, inspection, packaging
and shipping, sales and complaints, inventory and maintenance, costs, management
control, industrial engineering and research.
It can be seen that both the lists are extremely comprehensive. This clearly points out
that specific statistical problems arising in the course of business operations are
multitudinous. As such, one may do no more than highlight some of the more important
ones to emphasise the relevance of Statistics to the business world.
Personnel Management
This is another sphere in business where statistical methods can be used. Here, one is
concerned with the fixation of wage rates, incentive norms and performance appraisal of
individual employee. The concept of productivity is very relevant here. On the basis of
measurement of productivity, the productivity bonus is awarded to the workers.

Comparisons of wages and productivity are undertaken in order to ensure increases in


industrial productivity.
Seasonal Behaviour
A business firm engaged in the sale of certain product has to decide how much stock of
that product should be kept. If the product is subject to seasonal fluctuations then it
must know the nature of seasonal fluctuations in demand. For this purpose, seasonal
index of consumption may be required. If the firm can obtain such data or construct a
seasonal index on its own then it can keep a limited stock of the product in lean months
and large stocks in the remaining months. In this way, it will avoid the blocking of funds
in maintaining large stocks in the lean months. It will also not miss any opportunity to
sell the product in the busy season by maintaining adequate stock of the product during
such a period.
Export Marketing
Developing countries have started giving considerable importance to their exports.
Here, too, quality is an important factor on which exports depend. This apart, the
concerned firm must know the probable countries where its product can be exported.
Before that, it must select the right product, which has considerable demand in the
overseas markets. This is possible by carefully analysing the statistics of imports and
exports. It may also be necessary to undertake a detailed survey of overseas markets to
know more precisely the export potential of a given product.
Maintenance of Cost Records
Cost is an important consideration for a business enterprise. It has to ensure that cost of
production, which includes cost of raw materials, wages, and so forth, does not mount
up or else this would jeopardize its competitiveness in the market. This implies that it
has to maintain proper cost records and undertake an analysis of cost data from time to
time.
Management of Inventory
Closely related to the cost factor is the problem of inventory management. In order to
ensure that the production process continues uninterrupted, the business firm has to
maintain an adequate inventory. At the same time, excessive inventory means blocking
of funds that could have been utilized elsewhere. Thus, the firm has to determine a
magnitude of inventory that is neither excessive nor inadequate.
While doing so, it has to bear in mind the probable demand for its product. All these
aspects can be well looked after if proper statistics are maintained and analyzed.
Expenditure on Advertising and Sales

A number of times business firms are interested to know whether there is an association
between two or more variables such as advertising expenditure and sales. In view of
increasing competitiveness, business and industry spend a large amount on advertising.
It is in their interest to find out whether such advertising expenditure promotes the
sales. Here, by using correlation and regression techniques it can be ascertained that the
advertising expenditure is worthwhile of not.
Mutual Funds
Mutual funds which have come into existence in recent years, provide an avenue to a
person to invest his savings so that he may get a reasonably good return. Different
mutual funds have different objectives as they have varying degrees of risk involved in
the companies they invest in. Here, Statistics provides certain tools or techniques to a
consultant or financial adviser through which he can provide sound advice to a
prospective investor.
Relevance in Banking and Insurance Institutions
Banks and insurance companies frequently use varying statistical techniques in their
respective areas of operation. They have to maintain their accounts and analyze these to
examine their performance over a specified period.
The above discussion is only illustrative and there are numerous other areas where the
use of Statistics is so common that without its use they may have to close down their
operations.
STATISTICAL QUALITY CONTROL METHODS
In the sphere of production, for example, statistics can be useful in various ways to
ensure the production of quality goods. This is achieved by identifying and rejecting
defective or substandard goods. The sale targets can be fixed on the basis of sale
forecasts, which are done by using varying methods of forecasting. Analysis of sales
done against the targets set earlier would indicate the deficiency in achievement, which
may be on account of several causes: (i) targets were too high and unrealistic (ii)
salesmen's performance has been poor (iii) emergence of increase in competition, and
(iv) poor quality of company's product, and so on. These factors can be further
investigated.
LIMITATIONS OF STATISTICS
The preceding discussion highlighting the importance of Statistics in business should
not lead anyone to conclude that Statistics is free from any limitation. As we shall see
here, Statistics has a number of limitations.
There are certain phenomena or concepts where Statistics cannot be used. This is
because these phenomena or concepts are not amenable to measurement. For example,

beauty, intelligence, courage cannot be quantified. Statistics has no place in all such
cases where quantification is not possible.
1. Statistics reveal the average behaviour, the normal or the general trend. An
application of the 'average' concept if applied to an individual or a particular situation
may lead to a wrong conclusion and sometimes may be disastrous. For example, one
may be misguided when told that the average depth of a river from one bank to the other
is four feet, when there may be some points in between where its depth is far more than
four feet. On this understanding, one may enter those points having greater depth,
which may be hazardous.
2. Since Statistics are collected for a particular purpose, such data may not be relevant
or useful in other situations or cases. For example, secondary data (i.e., data originally
collected by someone else) may not be useful for the other person.
3. Statistics is not 100 percent precise as is Mathematics or Accountancy. Those who use
Statistics should be aware of this limitation.
4. In Statistical surveys, sampling is generally used as it is not physically possible to
cover all the units or elements comprising the universe. The results may not be
appropriate as far as the universe is concerned. Moreover, different surveys based on
the same size of sample but different sample units may yield different results.
5. At times, association or relationship between two or more variables is studied in
Statistics, but such a relationship does not indicate 'cause and effect' relationship. It
simply shows the similarity or dissimilarity in the movement of the two variables. In
such cases, it is the user who has to interpret the results carefully, pointing out the type
of relationship obtained.
6. A major limitation of Statistics is that it does not reveal all pertaining to a certain
phenomenon.
7. There is some background information that Statistics does not cover. Similarly, there
are some other aspects related to the problem on hand, which are also not covered. The
user of Statistics has to be well informed and should interpret Statistics keeping in mind
all other aspects having relevance on the given problem.
SUMMARY
This chapter outlined the importance and growth of statistics. Various applications of
statistics in the domain of management have been dealt in detail.
KEY TERMS

Statistics
Statistical quality control methods
Seasonal behaviour

IMPORTANT QUESTIONS
1. Define statistics
2. What do you mean by statistical quality methods?
3. Explain the application of statistics in various business domains.

- End of Chapter LESSON 5


ESTIMATION OF POPULATION PARAMETERS

OBJECTIVE

To acquire knowledge of estimation of parameter

STRUCTURE

Estimation of population parameters


Measure of central tendency
Mean, median, mode
Geometric mean
Harmonic mean

A population is any entire collection of people, animals, plants or things from which we
may collect data. It is the entire group we are interested in, which we wish to describe or
draw conclusions about.
In order to make any generalizations about a population, a sample, that is meant to be
representative of the population, is often studied. For each population there are many
possible samples. A sample statistic gives information about a corresponding population
parameter. For example, the sample mean for a set of data would give information about
the overall population mean.
It is important that the investigator carefully and completely defines the population
before collecting the sample, including a description of the members to be included.
Example
The population for a study of infant health might be all children born in the India, in the
1980's. The sample might be all babies born on 7th May in any of the years.

Population Parameters
Parameter is a value, usually unknown (and which therefore has to be estimated), used
to represent a certain population characteristic. For example, the population mean is a
parameter that is often used to indicate the average value of a quantity.
Within a population, a parameter is a fixed value which does not vary. Each sample
drawn from the population has its own value of any statistic that is used to estimate this
parameter. For example, the mean of the data in a sample is used to give information
about the overall mean in the population from which that sample was drawn.
MEASURES OF CENTRAL LOCATION (OR CENTRAL TENDENCY)
The most important objective of statistical analysis is to determine a single value for the
entire mass of data, which describes the overall level of the group of observations and
can be called a representative set of data. It tells us where the centre of the distribution
of data is located on the scale that we are using. There are several such measures, but we
shall discuss only those that are most commonly used. These are: Arithmetic Mean,
Mode and Median. These values are very useful in not only presenting the overall
picture of the entire data, but also for the purpose of making comparisons among two or
more sets of data.
As an example, questions like, "How hot is the month of June in Mumbai?" can be
answered, generally, by a single figure of the average temperature for that month. For
the purpose of comparison, suppose that we want to find out if boys and girls at the age
10 differ in height. By taking the average height of boys of that age and the average
height of the girls of the same age, we can compare and note the difference.
While, arithmetic mean is the most commonly used measure of central location, mode
and median arc more suitable measures under certain set of conditions and for certain
types of data. However, all measures of central tendency should meet the following
requisites:

It should be easy to calculate and understand.


It should be rigidly defined. It should have one and only one interpretation so
that the personal prejudice or bias of the investigator does not affect the value or
its usefulness.
It should be representative of the data. If it is calculated from a sample, then the
sample should be random enough to be accurately representing the population.
It should have sampling stability. It should not be affected by sampling
fluctuations. This means that if we pick 10 different groups of college students at
random and we compute the average of each group, then we should expect to get
approximately the same value from these groups.
It should not be affected much by extreme values. If a few very small or very large
items are presented in the data, they will unduly influence the value of the
average by shifting it to one side or the other and hence the average would not be

really typical of the entire series. Hence, the average chosen should be such that it
is not unduly influenced by extreme values.
Let us consider these three measures of the central tendency:
MODE
In statistics, mode means the most frequent value assumed by a random variable, or
occurring in a sampling of a random variable. The term is applied both to probability
distributions and to collections of experimental data.
Like the statistical mean and the median, die mode is a way of capturing important
information about a random variable or a population in a single quantity. The mode is in
general different from mean and median, and may be very different for strongly skewed
distributions.
The mode is not necessarily unique, since the same maximum frequency may be
attained at different values. The worst case is given by so-called uniform distributions,
in which all values are equally likely.
Mode of a probability distribution
The mode of a probability distribution is the value at which its probability density
function attains its maximum value, so, informally speaking; the mode is at the peak.
Mode of a sample
The mode of a data sample is the element that occurs most often in the collection. For
example, the mode of the sample [1, 3, 6, 6, 6, 6, 7, 7, 12, 12, 17] is 6. Given the list of
data [1, 1, 2, 4, 4] the mode is not unique.
For a sample from a continuous distribution, such as [0.935..., 1.211..., 2.430..., 3.668...,
3.874...], the concept is unusable in its raw form, since each value will occur precisely
once. The usual practice is to discreteise the data by assigning the values to equidistant
intervals, as for making a histogram, effectively replacing the values by the midpoints of
the intervals they are assigned to. The mode is then the value where the histogram
reaches its peak. For small or middle-sized samples the outcome of this procedure is
sensitive to the choice of interval width if chosen too narrow or too wide; typically one
should have a sizable fraction of the data concentrated in a relatively small number of
intervals (5 to 10), while the fraction of the data falling outside these intervals is also
sizable.
Comparison of mean, median and mode
For a probability distribution, the mean is also called the expected value of the random
variable. For a data sample, the mean is also called the average.

When do these measures make sense?


Unlike mean and median, the concept of mode also makes sense for "nominal data" (i.e.,
not consisting of numerical values). For example, taking a sample of Korean family
names, one might find that "Kim" occurs more often than any other name. Then "Kim"
might be called the mode of the sample. However, this use is not common.
Unlike median, the concept of mean makes sense for any random variable assuming
values from a vector space, including the real numbers (a one-dimensional vector space)
and the integers (which can be considered embedded in the real numbers). For example,
a distribution of points in the plane will typically have a mean and a mode, but the
concept of median does not apply. The median makes sense when there is a linear order
on the possible values.
Uniqueness and Definedness
For the remainder, the assumption is that we have (a sample of) a real-valued random
variable.
For some probability distributions, the expected value may be infinite or undefined, but
if defined, it is unique. The average of a (finite) sample is always defined. The median is
the value such that the fractions not exceeding it and not falling below it are both at least
1/2. It is not necessarily unique, but never infinite or totally undefined. For a data
sample it is the "halfway" value when the list of values is ordered in increasing value,
where usually for a list of even length the numerical average is taken of the two values
closest to "halfway". Finally, as said before, the mode is not necessarily unique.
Furthermore, like the mean, the mode of a probability distribution can be (plus or
minus)-infinity, but unlike the mean it cannot be just undefined. For a finite data
sample, the mode is one (or more) of the values in the sample and is itself then finite.
Properties
Assuming definedness, and for simplicity uniqueness, the following are some of the
most interesting properties.
All three measures have the following property: If the random variable (or each value
from the sample) is subjected to the linear or affine transformation which replaces Xby
ax+b, so are the mean, median and mode.
However, if there is an arbitrary monotonic transformation, only the median follows; for
example, if X is replaced by exp(X), the median changes from m to exp(m) but the mean
and mode won't.
Except for extremely small samples, the median is totally insensitive to "outliers" (such
as occasional, rare, false experimental readings). The mode is also very robust in the
presence of outliers, while the mean is rather sensitive.

In continuous uni-modal distributions the median lies, as a rule of thumb, between the
mean and the mode, about one third of the way going from mean to mode. In a formula,
median = (2 x mean + mode)/3. This rule, due to Karl Pearson, is however not a hard
and fast rule. It applies to distributions that resemble a normal distribution.
Example for a skewed distribution
A well-known example of a skewed distribution is personal wealth: Few people are very
rich, but among those some are excessively rich. However, many are rather poor.
A well-known class of distributions that can be arbitrarily skewed is given by the lognormal distribution. It is obtained by transforming a random variable X having a
normal distribution into random variable Y = exp(X). Then the logarithm of random
variable Y is normally distributed, whence the name.
Taking the mean of X to be 0, the median of Y will be 1, independent of the standard
deviation of X. This is so because X has a symmetric distribution, so its median is also
0. The transformation from X to Y is monotonic, and so we find the median exp(0) = 1
for Y.
When X has standard deviation = 0.2, the distribution of Y is not very skewed. We find
(see under Log-normal distribution), with values rounded to four digits:
Mean = 1.0202
Mode = 0.9608
Indeed, the median is about one third on the way from mean to mode.
When X has a much larger standard deviation, = 5, the distribution of Y is strongly
skewed. Now
Mean = 7.3891
Mode = 0.0183
Here, Pearson's rule of thumb fails miserably.
MEDIAN
In probability theory and statistics, a median is a number dividing the higher half of a
sample, a population, or a probability distribution from the lower half. The median of a
finite list of numbers can be found by arranging all the observations from lowest value to
highest value and picking the middle one. If there is an even number of observations,
one often takes the mean of the two middle values.

At most, half the population has values less than the median and at most half have
values greater than the median. If both groups contain less than half the population,
then some of the population is exactly equal to the median.
Popular explanation
The difference between the median and mean is illustrated in a simple example.
Suppose 19 paupers and one billionaire are in a room. Everyone removes all money from
their pockets and puts it on a table. Each pauper puts $5 on the table; the billionaire
puts $1 billion (that is, $109) there. The total is then $1,000,000,095. If that money is
divided equally among the 20 persons, each gets $50,000,004.75. That amount is the
mean (or "average") amount of money that the 20 persons brought into the room. But
the median amount is $5, since one may divide the group into two groups of 10 persons
each, and say that everyone in the first group brought in no more than $5, and each
person in the second group brought in no less than $5. In a sense, the median is the
amount that the typical person brought in. By contrast, the mean (or "average") is not at
all typical, since no one present - pauper or billionaire - brought in an amount
approximating $50,000,004.75.
Non-uniqueness
There may be more than one median. For example if there are an even number of cases,
and the two middle values are different, then there is no unique middle value. Notice,
however, that at least half the numbers in the list are less than or equal to either of the
two middle values, and at least half are greater than or equal to either of the two values,
and the same is true of any number between the two middle values. Thus either of the
two middle values and all numbers between them are medians in that case.
Measures of statistical dispersion
When the median is used as a location parameter in descriptive statistics, there are
several choices for a measure of variability: the range, the inter-quartile range, and the
absolute deviation. Since the median is the same as the second quartile, its calculation is
illustrated in the article on quartiles. To obtain the median of an even number of
numbers, find the average of the two middle terms.
Medians of particular distributions
The median of a normal distribution with mean and variance 2 is . In fact, for a
normal distribution, mean = median = mode.
The median of a uniform distribution in the interval [a, b] is (a + b) / 2, which is also the
mean.
Medians in descriptive statistics

The median is primarily used for skewed distributions, which it represents differently
than the arithmetic mean. Consider the multiset {1, 2, 2, 2, 3, 9}. The median is 2 in this
case, as is the mode, and it might be seen as a better indication of central tendency than
the arithmetic mean of 3.166....
Calculation of medians is a popular technique in summary statistics and summarizing
statistical data, since it is simple to understand and easy to calculate, while also giving a
measure that is more robust in the presence of outlier values than is the mean.
MEAN
In statistics, mean has two related meanings:
- The average in ordinary English, which is also called the arithmetic mean (and is
distinguished from the geometric mean or harmonic mean). The average is also called
sample mean. The expected value of a random variable, which is also called the
population mean.
- In statistics, means are often used in geometry and analysis. A wide range of means
have been developed for these purposes, which are not much used in statistics. See the
other means section below for a list of means.
Sample mean is often used as an estimator of the central tendency such as the
population mean. However, other estimators are also used.
For a real-valued random variable X, the mean is the expectation of X. If the expectation
does not exist, then the random variable has no mean.
For a data set, the mean is just the sum of all the observations divided by the number of
observations. Once we have chosen this method of describing the communality of a data
set, we usually use the standard deviation to describe how the observations differ. The
standard deviation is the square root of the average of squared deviations from the
mean.
The mean is the unique value about which the sum of squared deviations is a minimum.
If you calculate the sum of squared deviations from any other measure of central
tendency, it will be larger than for the mean. This explains why the standard deviation
and the mean are usually cited together in statistical reports.
An alternative measure of dispersion is the mean deviation, equivalent to the average
absolute deviation from the mean. It is less sensitive to outliers, but less tractable when
combining data sets
Arithmetic Mean
The arithmetic mean is the "standard" average, often simply called the "mean".

The mean may often be confused with the median or mode. The mean is the arithmetic
average of a set of values, or distribution; however, for skewed distributions, the mean is
not necessarily the same as the middle value (median), or most likely (mode). For
example, mean income is skewed upwards by a small number of people with very large
incomes, so that the majority has an income lower than the mean. By contrast, the
median income is the level at which half the population is below and half is above. The
mode income is the most likely income, and favors the larger number of people with
lower incomes. The median or mode is often more intuitive measures of such data.
That said, many skewed distributions are best described by their mean - such as the
Exponential and Poisson distributions.
An amusing example
Most people have an above average number of legs. The mean number of legs is going to
be less than 2, because there are people with one leg, people with no legs and no people
with more than two legs. So since most people have two legs, they have an above average
number.
Geometric Mean
The geometric mean is an average that is useful for sets of numbers that are interpreted
according to their product and not their sum (as is the case with the arithmetic mean).
For example rates of growth.

For example, the geometric mean of 34, 27, 45, 55, 22, 34 (six values) is (34 x 27 x 45 x
55 x 22 x 34)1/6 = (1699493400)1/6 = 34.545
Harmonic Mean
The harmonic mean is an average which is useful for sets of numbers which are defined
in relation to some unit, for example speed (distance per unit of time).

An example
An experiment yields the following data: 34, 27, 45, 55, 22, 34. We need to find the
harmonic mean. No. of items is 6, therefore n = 6. Value of the denominator in the
formula is 0.181719152307. Reciprocal of this value is 5.50299727522. Now, we multiply
this by n to get the harmonic mean as 33.0179836513.
Weighted Arithmetic Mean
The weighted arithmetic mean is used, if one wants to combine average values .rom
samples of the same population with different sample sizes:

The weights i represent the bounds of the partial sample. In other applications they
represent a measure for the reliability of the influence upon the mean by respective
values.
SUMMARY
This chapter has given the meaning of population parameters. The procedures of
measuring the above population parameters are dealt with in detail in the chapter.
KEYTERMS

Population parameters
Mean, Mode and Median
Arithmetic mean
Geometric mean
Harmonic mean
Skewed distribution

IMPORTANT QUESTIONS
1. Explain the methods to measure the Median, Mode and Mean

2. What are the different types of Means?

- End of Chapter -

LESSON 6
HYPOTHESIS TESTING

OBJECTIVES

To learn the process of Hypothesis Testing


To find out the details of Null hypothesis and Alternate hypothesis
To learn the precise meaning of probability value
To understand the Type I error and Type II errors
To gain knowledge on various methods to test hypothesis

STRUCTURE

Hypothesis test
Statistical and practical significance
One and two failed tests
Type I and Type II Errors

Statistical hypothesis is an assumption about a population parameter. This


assumption may or may not be true.
The best way to determine whether a statistical hypothesis is true is to examine the
entire population. Since this is often impractical, researchers typically examine a
random sample from the population. If the sample data are consistent with the
statistical hypothesis, the hypothesis is accepted if not, the hypothesis is rejected.
There are two types of statistical hypotheses.

Null hypothesis: usually the hypothesis that sample observations result purely
from chance effects.
Alternative hypothesis: the hypothesis that sample observations are influenced by
some non-random cause.

For example, suppose we wanted to determine whether a coin was fair and balanced. A
null hypothesis might be that half the flips would result in Heads and half in Tails. The
alternative hypothesis might be that the number of Heads and Tails would be very

different. Suppose we flipped the coin 50 times resulting in 40 Heads and 10 Tails Given
this result, we would be inclined to reject the null hypothesis and accept the alternative
hypothesis.
HYPOTHESIS TESTS
Statisticians follow a formal process to determine whether to accept or reject a null
hypothesis, based on sample data. This process, called hypothesis testing, consists of
four steps.
1. Formulate the hypotheses. This involves stating the null and alternative
hypotheses. The hypotheses are stated in such a way that they are mutually exclusive.
That is, if one is true, the other must be false, and vice versa.
2. Identify the test statistic. This involves specifying the statistic (e.g., a mean score,
proportion) that will be used to assess the validity of the null hypothesis.
3. Formulate a decision rule. A decision rule is a procedure that the researcher uses
to decide whether to accept or reject the null hypothesis.
4. Accept or reject the null hypothesis. Use the decision rule to evaluate the test
statistic. If the statistic is consistent with the null hypothesis, accept the null hypothesis;
otherwise, reject the null hypothesis.
This section provides an introduction to hypothesis testing. Basic analysis involves some
hypothesis testing. Examples of hypotheses generated in marketing research abound

The department store is being patronized by more than 10 percent of the


households
The heavy and light users of a brand differ in terms of psychographic
characteristics
One hotel has a more upscale image than its close competitor
Familiarity with a restaurant results in greater preference for that restaurant.

The null hypothesis is a hypothesis about a population parameter. The purpose of


hypothesis testing is to test the viability of the null hypothesis in the light of
experimental data. Depending on the data, the null hypothesis either will or will not be
rejected as a viable possibility.
Consider a researcher interested in whether the time to respond to a tone is affected by
the consumption of alcohol. The null hypothesis is that 1 - 2 = 0, where 1 is the mean
time to respond after consuming alcohol and 2 is the mean time to respond otherwise.
Thus, the null hypothesis concerns the parameter 1 - 2 and the null hypothesis is that
the parameter equals zero.
The null hypothesis is often the reverse of what the experimenter actually believes; it is
put forward to allow the data to contradict it. In the experiment on the effect of alcohol,

the experimenter probably expects alcohol to have a harmful effect. If the experimental
data show a sufficiently large effect of alcohol, then the null hypothesis that alcohol has
no effect can be rejected.
It should be stressed that researchers very frequently put forward a null hypothesis in
the hope that they can discredit it. For a second example, consider an educational
researcher who designed a new way to teach a particular concept in science, and wanted
to test experimentally whether this new method worked better than the existing method.
The researcher would design an experiment comparing the two methods. Since the null
hypothesis would be that there is no difference between the two methods, the researcher
would be hoping to reject the null hypothesis and conclude that the method he or she
developed is the better of the two.
The symbol H0 is used to indicate the null hypothesis. For the example just given, the
null hypothesis would be designated by the following symbols: H0: 1 - 2 = 0, or by
H0: 1 = 2
The null hypothesis is typically a hypothesis of no difference as in this example where it
is the hypothesis of no difference between population means. That is why the word
"null" in "null hypothesis" is used - it is the hypothesis of no difference.
Despite the "null" in "null hypothesis", there are occasions when the parameter is not
hypothesized to be 0. For instance, it is possible for the null hypothesis to be that the
difference between population means is a particular value. Or, the null hypothesis could
be that the mean SAT score in some population is 600. The null hypothesis would then
be stated as: H0: = 600. Although the null hypotheses discussed so far have all
involved the testing of hypotheses about one or more population means, null hypotheses
can involve any parameter.
An experiment investigating the correlation between job satisfaction and performance
on the job would test the null hypothesis that the population correlation () is 0.
Symbolically, H0: = 0. Some possible null hypotheses are given below:
H0: = 0
H0: = 10
H0: 1 - 2 = 0
H0: = 0.5
H0: 1- 2 = 0
H0: 1 - 2 - 3
H0: 1 2 = 0

Steps in hypothesis testing


1. The first step in hypothesis testing is to specify the null hypothesis (H0) and the
alternative hypothesis (H1). If the research concerns whether one method of presenting
pictorial stimuli leads to better recognition than another, the null hypothesis would
most likely be that there is no difference between methods (H0: 1 - 2 = 0). The
alternative hypothesis would be H1: 1 2. If the research concerned the correlation
between grades and SAT scores, the null hypothesis would most likely be that there is no
correlation (H0: = 0). The alternative hypothesis would be H0: 0.
2. The next step is to select a significance level. Typically the 0.05 or the 0.01 level is
used.
3. The third step is to calculate a statistic analogous to the parameter specified by the
null hypothesis. If the null hypothesis were defined by the parameter 1 - 2, then the
statistic M1 - M2 would be computed.
4. The fourth step is to calculate the probability value (often called the p-value). The p
value is the probability of obtaining a statistic as different or more different from the
parameter specified in the null hypothesis as the statistic computed from the data. The
calculations are made assuming that the null hypothesis is true.
5. The probability value computed in Step 4 is compared with the significance level
chosen in Step 2. If the probability is less than or equal to the significance level, then the
null hypothesis is rejected; if the probability is greater than the significance level then
the null hypothesis is not rejected. When the null hypothesis is rejected, the outcome is
said to be statistically significant; when the null hypothesis is not rejected then the
outcome is said be not statistically significant.
6. If the outcome is statistically significant, then the null hypothesis is rejected in favor
of the alternative hypothesis. If the rejected null hypothesis were that 1 - 2 = 0, then
the alternative hypothesis would be that 1 2. If M1 were greater than M2 then the
researcher would naturally conclude that 1 2.
7. The final step is to describe the result and the statistical conclusion in an
understandable way. Be sure to present the descriptive statistics as well as whether the
effect was significant or not. For example, a significant difference between a group that
received a drug and a control group might be described as follows:
Subjects in the drug group scored significantly higher (M = 23) than did subjects in the
control group (M = 17), t(18) = 2.4, p = 0.027. The statement that "t(18) = 2.4" has to do
with how the probability value (p) was calculated. A small minority of researchers might
object to two aspects of this wording. First, some believe that the significance level
rather than the probability level should be reported. The argument for reporting the
probability value is presented in another section. Second, since the alternative
hypothesis was stated as 1 2, some might argue that it can only be concluded that the
population means differ and not that the population mean.

This argument is misguided. Intuitively, there are strong reasons for inferring that the
direction of the difference in the population is the same as the difference in the sample.
There is also a more formal argument. A non significant effect might be described as
follows:
Although subjects in the drug group scored higher (M = 23) than did subjects in the
control group (M = 20), the difference between means was not significant, t(18) = 1.4. p
= 0.179. It would not have been correct to say that there was no difference between the
performances of the two groups. There was a difference. It is just that the difference was
not large enough to rule out chance as an explanation of the difference. It would also
have been incorrect to imply that there is no difference in the population. Be sure not to
accept the null hypothesis.
The Precise Meaning of the Probability Value
There is often confusion about the precise meaning of the probability computed in a
significance test. As stated in Step 4 of the steps in hypothesis testing, the null
hypothesis (H0) is assumed to be true. The difference between the statistic computed in
the sample and the parameter specified by H0 is computed and the probability of
obtaining a difference this large or large is calculated. This probability value is the
probability of obtaining data as extreme or more extreme than the current data
(assuming H0 is true). It is not the probability of the null hypothesis itself. Thus, if the
probability value is 0.005, this does not mean that the probability that the null
hypothesis is true is .005. It means that the probability of obtaining data as different or
more different from the null hypothesis as those obtained in the experiment is 0.005.
The inferential step to conclude that the null hypothesis is false goes, as follows:
The data (or data more extreme) are very unlikely given that the null hypothesis is true.
This means that: (1) a very unlikely event occurred or (2) the null hypothesis is false.
The inference usually made is that the null hypothesis is false.
To illustrate that the probability is not the probability of the hypothesis, consider a test
of a person who claims to be able to predict whether a coin will come up heads or tails.
One should take a rather skeptical attitude toward this claim and require strong
evidence to believe in its validity. The null hypothesis is that the person can predict
correctly half the time (H0: = 0.5). In the test, a coin is flipped 20 times and the person
is correct 11 times. If the person has no special ability (H0 is true), then the probability of
being correct 11 or more times out of 20 is 0.41. Would someone who was originally
skeptical now believe that there is only a 0.41 chance that the null hypothesis is true?
They almost certainly would not since they probably originally thought H0 had a very
high probability of being true (perhaps as high as 0.9999). There is no logical reason for
them to decrease their belief in the validity of the null hypothesis since the outcome was
perfectly consistent with the null hypothesis.
The proper interpretation of the test is as follows:

A person made a rather extraordinary claim and should be able to provide strong
evidence in support of the claim if the claim is to be believed. The test provided data
consistent with the null hypothesis that the person has no special ability since a person
with no special ability would be able to predict as well or better more than 40% of the
time. Therefore, there is no compelling reason to believe the extraordinary claim.
However, the test does not prove the person cannot predict better than chance; it simply
fails to provide evidence that he or she can. The probability that the null hypothesis is
true is not determined by the statistical analysis conducted as part of hypothesis testing.
Rather, the probability computed is the probability of obtaining data as different or
more different from the null hypothesis (given that the null hypothesis is true) as the
data actually obtained.
According to one view of hypothesis testing, the significance level should be specified
before any statistical calculations are performed. Then, when the probability (p) is
computed from a significance test, it is compared with the significance level. The null
hypothesis is rejected if p is at or below the significance level; it is not rejected if p is
above the significance level. The degree to which p ends up being above or below the
significance level does not matter. The null hypothesis either is or is not rejected at the
previously stated significance level. Thus, if an experimenter originally stated that he or
she was using the 0.05 significance level and p was subsequently calculated to be 0.042,
then the person would reject the null hypothesis at the 0.05 level. If p had been 0.0001
instead of 0.042 then the null hypothesis would still be rejected at the 0.05 level. The
experimenter would not have any basis to be more confident that the null hypothesis
was false with a p of 0.0001 than with a p of 0.041. Similarly, if the p had been 0.051
then the experimenter would fail to reject the null hypothesis.
He or she would have no more basis to doubt the validity of the null hypothesis than if p
had been 0.482. The conclusion would be that the null hypothesis could not be rejected
at the 0.05 level. In short, this approach is to specify the significance level in advance
and use p only to determine whether or not the null hypothesis can be rejected at the
stated significance level.
Many statisticians and researchers find this approach to hypothesis testing not only too
rigid, but basically illogical. Who in their right mind would not have more confidence
that the null hypothesis is false with a p of 0.0001 than with a p of 0.042? The less likely
the obtained results (or more extreme results) under the null hypothesis, the more
confident one should be that the null hypothesis is false. The null hypothesis should not
be rejected once and for all. The possibility that it was falsely rejected is always present,
and, all else being equal, the lower the p value, the lower this possibility.
Statistical and Practical Significance
It is important not to confuse the confidence with which the null hypothesis can be
rejected with size of the effect. To make this point concrete, consider a researcher
assigned the task of determining whether the video display used by travel agents for
booking airline reservations should be in color or in black and white. Market research
had shown that travel agencies were primarily concerned with the speed with which

reservations can be made. Therefore, the question was whether color displays allow
travel agents to book reservations faster. Market research had also shown that in order
to justify the higher price of color displays, they must be faster by an average of at least
10 seconds per transaction. Fifty subjects were tested with color displays and 50 subjects
were tested with black and white displays.
Subjects were slightly faster at making reservations on a color display (M = 504.7
seconds) than on a black and white display (M = 508.2) seconds, although the difference
is small, it was statistically significant at the .05 significance level. Box plots of the data
are shown below.

The 95% confidence interval on the difference between means is:


-7.0 < colour black & white -0.1
which means that the experimenter can be confident that the color display is between
7.0 seconds and 0.1 seconds faster. Clearly, the difference is not big enough to justify the
more expensive color displays. Even the upper limit of the 95% confidence interval
(seven seconds) is below the minimum needed to justify the cost (10 seconds).
Therefore, the experimenter could feel confident in his or her recommendation that the
black and white displays should be used. The fact that the color displays were
significantly faster does not mean that they were much fasten It just means that the
experimenter can reject the null hypothesis that there is no difference between the
displays.
The experimenter presented this conclusion to management but management did not
accept it. The color displays were so dazzling that despite the statistical analysis, they
could not believe that color did not improve performance by at least 10 seconds. The
experimenter decided to do the experiment again, this time using 100 subjects for each
type of display. The results of the second experiment were very similar to the first.
Subjects were slightly faster at making reservations on a color display (M = 504.7
seconds) than on a black and white display (M = 508.1 seconds). This time the
difference was significant at the 0.01 level rather than the 0.05 level found in the first
experiment. Despite the fact that the size of the difference between means was no larger,

the difference was "more significant" due to the larger sample size used. If the
population difference is zero, then a sample difference of 3.4 or larger with a sample size
of 100 is less likely than a sample difference of 3.5 or larger with a sample size of 50.
The 95% confidence interval on the difference between means is:
-5.8 < colour black & white -0.9
and the 99% interval is:
-6.6 < colour black & white -0.1
Therefore, despite the finding of a "more significant" difference between means the
experimenter can be even more certain that the color displays are only slightly better
than the black and white displays. The second experiment shows conclusively that the
difference is less than 10 seconds.
This example was used to illustrate the following points:
(1) an effect that is statistically significant is not necessarily large enough to be of
practical significance and
(2) the smaller of two effects can be "more significant" than the larger.
Be careful how you interpret findings reported in the media. If you read that a particular
diet lowered cholesterol significantly, this does not necessarily mean that the diet
lowered cholesterol enough to be of any health value. It means that the effect on
cholesterol in the population is greater than zero.
TYPE I AND II ERRORS
There are two kinds of errors that can be made in significance testing:
(1) a true null hypothesis can be incorrectly rejected and
(2) a false null hypothesis can fail to be rejected.
The former error is called a Type I error and the latter error is called a Type II error.
These two types of errors are defined in the table.

The probability of a Type I error is designated by the Greek letter alpha () and is called
the Type I error rate; the probability of a Type II error is designated by the Greek
letter beta (), and is called the Type II error rate.
A Type II error is only an error in the sense that an opportunity to reject the null
hypothesis correctly was lost. It is not an error in the sense that an incorrect conclusion
was drawn since no conclusion is drawn when the null hypothesis is not rejected. A Type
I error, on the other hand, is an error in every sense of the word. A conclusion is drawn
that the null hypothesis is false when, in fact, it is true. Therefore, Type I errors are
generally considered more serious than Type II errors. The probability of a Type I error
() is called the significance level and is set by the experimenter. There is a tradeoff
between Type I and Type II errors. The more an experimenter protects himself or
herself against Type I errors by choosing a low level, the greater the chance of a Type II
error. Requiring very strong evidence to reject the null hypothesis makes it very unlikely
that a true null hypothesis will be rejected. However, it increases the chance that a false
null hypothesis will not be rejected, thus lowering power. The Type I error rate is almost
always set at .05 or at .01, the latter being more conservative since it requires stronger
evidence to reject the null hypothesis at the .01 level then at the .05 level.
One and Two Tailed Tests
In the section on "Steps in hypothesis testing", the fourth step involves calculating the
probability that a statistic would differ as much or more from parameter specified in the
null hypothesis as does the statistic obtained in the experiment. This statement implies
that a difference in either direction would be counted. That is, if the null hypothesis
were H0: 1 - 2 = 0, and the value of the statistic M1 - M2 were +5, then the probability
of M1 - M2 differing from zero by five or more (in either direction) would be computed.
In other words, probability value would be the probability that either M1 - M2 5 or M1 M2 -5.Assume that the figure shown below is the sampling distribution of M1 - M2.

The figure shows that the probability of a value of +5 or more is 0.036 and that the
probability of a value of -5 or less is .036. Therefore the probability of a value either
greater than or equal to +5 or less than or equal to -5 is 0.036 + 0.036 = 0.072.
A probability computed considering differences in both directions is called a "twotailed" probability. The name makes sense since both tails of the sampling distribution
are considered. There are situations in which an experimenter is concerned only with
differences in one direction. For example, an experimenter may be concerned with
whether or not 1 - 2 is greater than zero. However, if 1 - 2 is not greater than zero,
the experimenter may not care whether it equals zero or is less than zero. For instance, if
a new drug treatment is developed, the main issue is whether or not it is better than a
placebo. If the treatment is not better than a placebo, then it will not be used. It does not
really matter whether or not it is worse than the placebo.
When only one direction is of concern to an experimenter, then a "one-tailed" test can
be performed. If an experimenter were only to be concerned with whether or not 1 - 2
is greater than zero, then the one-tailed test would involve calculating the probability of
obtaining a statistic as greater than the one obtained in the experiment.
In the example, the one-tailed probability would be the probability of obtaining a value
of M1 - M2 greater than or equal to five given that the difference between population
means is zero.

The shaded area in the figure is greater than five. The figure shows that the one-tailed
probability is 0.036.
It is easier to reject the null hypothesis with a one-tailed than with a two-tailed test as
long as the effect is in the specified direction. Therefore, one-tailed tests have lower
Type II error rates and more power than do two-tailed tests. In this example, the onetailed probability (0.036) is below the conventional significance level of 0.05 whereas
the two-tailed probability (0.072) is not. Probability values for one-tailed tests are one
half the value for two-tailed tests as long as the effect is in the specified direction.
One-tailed and two-tailed tests have the same Type I error rate. One-tailed tests are
sometimes used when the experimenter predicts the direction of the effect in advance.
This use of one-tailed tests is questionable because the experimenter can only reject the

null hypothesis if the effect is in the predicted direction. If the effect is in the other
direction, then the null hypothesis cannot be rejected no matter how strong the effect is.
A skeptic might question whether the experimenter would really fail to reject the null
hypothesis if the effect were strong enough in the wrong direction. Frequently the most
interesting aspect of an effect is that it runs counter to expectations. Therefore, an
experimenter who committed himself or herself to ignoring effects in one direction may
be forced to choose between ignoring a potentially important finding and using the
techniques of statistical inference dishonestly. One-tailed tests are not used frequently.
Unless otherwise indicated, a test should be assumed to be two-tailed.
Confidence Intervals & Hypothesis Testing
There is an extremely close relationship between confidence intervals and hypothesis
testing. When a 95% confidence interval is constructed, all values in the interval are
considered plausible values for the parameter being estimated. Values outside the
interval are rejected as relatively implausible. If the value of the parameter specified by
the null hypothesis is contained in the 95% interval then the null hypothesis cannot be
rejected at the 0.05 level. If the value specified by the null hypothesis is not in the
interval then the null hypothesis can be rejected at the 0.05 level. If a 99% confidence
interval is constructed, then values outside the interval are rejected at the 0.01 level.
Imagine a researcher wishing to test the null hypothesis that the mean time to respond
to an auditory signal is the same as the mean time to respond to a visual signal. The null
hypothesis therefore is: visual auditory = 0.
Ten subjects were tested in the visual condition and their scores (in milliseconds) were:
355, 421, 299, 460, 600, 580, 474, 511, 550, and 586.
Ten subjects were tested in the auditory condition and their scores were: 275, 320, 278,
360, 430, 520, 464, 311, 529, and 326.
The 95% confidence interval on the difference between means is: 9 visual
auditory 196.
Therefore only values in the interval between 9 and 196 are retained as plausible values
for the difference between population means. Since zero, the value specified by the null
hypothesis, is not in the interval, the null hypothesis of no difference between auditory
and visual presentation can be rejected at the 0.05 level. The probability value for this
example is 0.034. Any time the parameter specified by a null hypothesis is not contained
in the 95% confidence interval estimating that parameter, the null hypothesis can be
rejected at the 0.05 level or less. Similarly, if the 99% interval does not contain the
parameter then the null hypothesis can be rejected at the 0.01 level. The null hypothesis
is not rejected if the parameter value specified by the null hypothesis is in the interval
since the null hypothesis would still be plausible.
However, since the null hypothesis would be only one of an infinite number of values in
the confidence interval, accepting the null hypothesis is not justified. There are many

arguments against accepting the null hypothesis when it is not rejected. The null
hypothesis is usually a hypothesis of no difference. Thus null hypothesis such as:
1 - 2 = 0
1 - 2 = 0
in which the hypothesized value is zero are most common. When the hypothesized value
is zero then there is a simple relationship between hypothesis testing and confidence
intervals:
If the interval contains zero then the null hypothesis cannot be rejected at the stated
level of confidence. If the interval does not contain zero then the null hypothesis can be
rejected.
This is just a special case of the general rule stating that the null hypothesis can be
rejected if the interval does not contain the hypothesized value of the parameter and
cannot be rejected if the interval contains the hypothesized value. Since zero is
contained in the interval, the null hypothesis that 1 - 2 = 0 cannot be rejected at the
0.05 level since zero is one of the plausible values of 1 - 2. The interval contains both
positive and negative numbers and therefore 1 may be either larger or smaller than 2.
None of the three possible relationships between 1 and 2:
1 - 2 = 0,
1 - 2 > 0, and
1 - 2 < 0
can be ruled out. The data are very inconclusive. Whenever a significance test fails to
reject the null hypothesis, the direction of the effect (if there is one) is unknown.
Now, consider the 95% confidence interval:
6 < 1 - 2 15
Since zero is not in the interval, the null hypothesis that 1 - 2 = 0 can be rejected at the
0.05 level. Moreover, since all the values in the interval are positive, the direction of the
effect can be inferred: 1 > 2.
Whenever a significance test rejects the null hypothesis that a parameter is zero, the
confidence interval on that parameter will not contain zero. Therefore either all the
values in the interval will be positive or all the values in the interval will be negative. In
either case, the direction of the effect is known.
Define the Decision Rule and the Region of Acceptance

The decision rule consists of two parts: (1) a test statistic and (2) a range of values, called
the region of acceptance. The decision rule determines whether a null hypothesis is
accepted or rejected. If the test statistic falls within the region of acceptance, the null
hypothesis is accepted; otherwise, it is rejected.
We define the region of acceptance in such a way that the chance of making a Type I
error is equal to the significance level. Here is how that is done:
Given the significance level , find the upper limit (UL) of the range of acceptance.
There are three possibilities, depending on the form of the null hypothesis i. If the null hypothesis is < M: The upper limit of the region of acceptance will be
equal to the value for which the cumulative probability of the sampling distribution is
equal to one minus the significance level. That is, P(x < UL) = 1 - .
ii. If the null hypothesis is = M: The upper limit of the region of acceptance will
be equal to the value for which the cumulative probability of the sampling distribution is
equal to one minus the significance level divided by 2. That is, P(x < UL) = 1 - /2.
iii. If the null hypothesis is > M: The upper limit of the region of acceptance is
equal to plus infinity.
In a similar way, we find the lower limit (LL) of the range of acceptance. Again, there
are three possibilities, depending on the form of the null hypothesis.
i. If the null hypothesis is < M: The lower limit of the region of acceptance is
equal to minus infinity.
ii. If the null hypothesis is = M: The lower limit of the region of acceptance will be
equal to the value for which the cumulative probability of the sampling distribution is
equal to the significance level divided by 2. That is, P(x < LL) = /2
iii. If the null hypothesis is > M: The lower limit of the region of acceptance will
be equal to the value for which the cumulative probability of the sampling distribution is
equal to the significance level. That is, P(x < LL) =
The region of acceptance is defined by the range between LL and UL.
Accept or Reject the Null Hypothesis
Once the region of acceptance is defined, the null hypothesis can be tested against
sample data. Simply compute the test statistic. In this case, the test statistic is the
sample mean. If the sample mean falls within the region of acceptance, the null
hypothesis is accepted; if not, it is rejected.
Other Considerations

When one tests a hypothesis in the real world, other issues may come into play. Here are
some suggestions that may be helpful.
You will need to make an assumption about the sampling distribution of the mean
score. If the sample is relatively large (i.e., greater than or equal to 30), you can assume,
based on the central limit theorem, that the sampling distribution will be roughly
normal. On the other hand, if the sample size is small (less than 30) and if the
population random variable is approximately normally distributed (i.e., has a bellshaped curve), you can transform the mean score into a t-score. The t-score will have a
t-distribution.
Assume that the mean of the sampling distribution is equal to the test value M
specified in the null hypothesis.
In some situations, you may need to compute the standard deviation of the sampling
distribution sx. If the standard deviation of the population is known, then sx = x
sqrt[(1/n) - (1/N)], where n is the sample size and N is the population size. On the other
hand, if the standard deviation of the population is unknown, then
sx = s x sqrt of [(1/n) - (1/N)], where s is the sample standard deviation.
Example 1
An inventor has developed a new, energy-efficient lawn mower engine. He claims that
the engine will run continuously for 5 hours (300 minutes) on a single gallon of regular
gasoline. Suppose a random sample of 50 engines is tested. The engines run for an
average of 295 minutes, with a standard deviation of 20 minutes. Test the null
hypothesis that the mean run time is 300 minutes against the alternative hypothesis
that the mean run time is not 300 minutes. Use a 0.05 level of significance.
Solution
There are four steps in conducting a hypothesis test, as described in the previous
sections. We work through those steps below:
1. Formulate hypotheses
The first step is to state the null hypothesis and an alternative hypothesis.
Null hypothesis: = 300 minutes
Alternative hypothesis: < > 300 minutes
Note that these hypotheses constitute a two-tailed test. The null hypothesis will be
rejected if the sample mean is too big or if it is too small.
2. Identify the test statistic

In this example, the test statistic is the mean run time of the 50 engines in the sample =
295 minutes.
3. Define the decision rule
The decision rule consists of two parts: (1) a test statistic and (2) a range of values, called
the region of acceptance. We already know that the test statistic is a sample mean equal
to 295. All that remains is to describe the region of acceptance; that is, to define the
lower limit and the upper limit of the region. Here is how that is done.
a. Specify the sampling distribution. Since the sample size is large (greater than
or equal to 30), we assume that the sampling distribution of the mean is normal,
based on the central limit theorem.
b. Define the mean of the sampling distribution. We assume that the mean of the
sampling distribution is equal to the mean value that appears in the null
hypothesis - 300 minutes.
c. Compute the standard deviation of the sampling distribution. Here the
standard deviation of the sampling distribution sx is:
sx = x sqrt[(1/n) - (1/N)]
sx = 20 x sqrt[1/50] = 2.83
where s is the sample standard deviation, n is the sample size, and N is the
population size. In this example, we assume that the population size N is very
large, so that the value 1/N is about zero.
4. Find the lower limit of the region of acceptance
Given a two-tailed hypothesis, the lower limit (LL) will be equal to the value for which
the cumulative probability of the sampling distribution is equal to the significance level
divided by 2. That is, P(x < LL) = /2 = 0.05/2 = 0.025. To find this lower limit, we use
the Normal Distribution table. From table, cumulative probability = 0.025, mean = 300,
and standard deviation = 2.83. The calculation tells us that the lower limit is 294.45,
given those inputs.
a. Find the upper limit of the region of acceptance. Given a two-tailed hypothesis,
the upper limit (UL) will be equal to the value for which the cumulative
probability of the sampling distribution is equal to one minus the significance
level divided by 2. That is, P(x < UL) = 1 - /2 = 1 - 0.025 = 0.975. To find this
upper limit, we use the Normal Distribution Table. From table, cumulative
probability = 0.975, mean = 300, and standard deviation = 2.83. The calculation
tells us that the upper limit is 305.55, given those inputs.

b. Thus, we have determined that the region of acceptance is defined by the


values between 294.45 and 305.55.
5. Accept or reject the null hypothesis
The sample mean in this example was 295 minutes. This value falls within the region of
acceptance. Therefore, we cannot reject the null hypothesis that a new engine runs for
300 minutes on a gallon of gasoline.
Example 2
Bon Air Elementary School has 300 students. The principal of the school thinks that the
average IQ of students at Bon Air is at least 110. To prove her point, she administers an
IQ test to 20 randomly selected students. Among the sampled students, the average IQ
is 108 with a standard deviation of 10. Based on these results, should the principal
accept or reject her original hypothesis? Assume a significance level of 0.01.
Solution
There are four steps in conducting a hypothesis test, as described in the previous
sections. We work through those steps below:
1. Formulate hypotheses.
The first step is to state the null hypothesis and an alternative hypothesis.
Null hypothesis: > 110
Alternative hypothesis: < 110
Note that these hypotheses constitute a one-tailed test. The null hypothesis will be
rejected if the sample mean is too small.
2. Identify the test statistic.
In this example, the test statistic is the mean IQ score of the 20 students in the sample.
Thus, the test statistic is the mean IQ score of 108.
3. Define the decision rule. The decision rule consists of two parts: (1) a test statistic
and (2) a range of values, called the region of acceptance. We already know that the test
statistic is a sample mean equal to 108. All that remains is to describe the region of
acceptance; that is, to define the lower limit and the upper limit of the region. Here is
how that is done.
a. Specify the sampling distribution. Since the sample size is small (less than 30),
we assume that the sampling distribution of the mean follows a t-distribution.

b. Define the mean of the sampling distribution. We assume that the mean of the
sampling distribution is equal to the mean value that appears in the null
hypothesis - 110.
c. Find the lower limit of the region of acceptance. Given a one-tailed hypothesis,
the lower limit (LL) will be equal to the value for which the cumulative
probability of the sampling distribution is equal to the significance level. That is,
P(x < LL) = = 0.01.
To find this lower limit, we use the T-Distribution Table. From table, cumulative
probability = 0.01, population mean = 110, standard deviation = 2.16, and
degrees of freedom = 20 - 1 = 19. The calculation tells us that the sample mean is
104.51, given those inputs. This is the lower limit of our region of acceptance.
d. Find the upper limit of the region of acceptance. Since we have a one-tailed
hypothesis in which the null hypothesis states that the IQ is greater than 110, any
big number is consistent with the null hypothesis. Therefore, the upper limit is
plus infinity.
Thus, we have determined that the range of acceptance is defined by the values between
104.51 and plus infinity.
4. Accept or reject the null hypothesis.
The sample mean in this example was an IQ score of 108. This value falls within the
region of acceptance. Therefore, we cannot reject the null hypothesis that the average IQ
score of students at Bon Air Elementary is equal to 110.
Power of a Hypothesis Test
When we conduct a hypothesis test, we accept or reject the null hypothesis based on
sample data. Because of the random nature of sample data, our decision can have four
possible outcomes.

We may accept the null hypothesis when it is true. Thus, the decision is correct.
We may reject the null hypothesis when it is true. This kind of incorrect decision
is called a Type I error.
We may reject the null hypothesis when it is false. Thus, the decision is correct.
We may accept the null hypothesis when it is false. This kind of incorrect decision
is called a Type II error.

The probability of committing a Type I error is called the significance level and is
denoted by . The probability of committing a Type II error is called Beta and is denoted
by . The probability of not committing a Type II error is called the power of the test.
How to Compute the Power of a Test

When a researcher designs a study to test a hypothesis, he/she should compute the
power of the test (i.e., the likelihood of avoiding a Type II error). Here is how to do that:
1. Define the region of acceptance.
(The process for defining the region of acceptance is described in the previous three
lessons. See, for example, Hypothesis Tests of Mean Score, Hypothesis Tests of
Proportion (Large Sample), and Hypothesis Tests of Proportion (Small Sample).)
2. Specify the critical value.
The critical value is an alternative to the value specified in the null hypothesis. The
difference between the critical value and the value from the null hypothesis is called the
effect size. That is, the effect size is equal to the critical value minus the value from the
null hypothesis.
3. Compute power. Assume that the true population parameter is equal to the critical
value, rather than the value specified in the null hypothesis. Based on that assumption,
compute the probability that the sample estimate of the population parameter will fall
outside the region of acceptance. That probability is the power of the test.
Example 1: Power of the Hypothesis Test of a Mean Score
Two inventors have developed a new, energy-efficient lawn mower engine. One inventor
says that the engine will run continuously for 5 hours (300 minutes) on a single gallon
of regular gasoline. Suppose a random sample of 50 engines is tested. The engines run
for an average of 295 minutes, with a standard deviation of 20 minutes. The inventor
tests the null hypothesis that the mean run time is 300 minutes against the alternative
hypothesis that the mean run time is not 300 minutes, using a 0.05 level of significance.
The other inventor says that the new engine will run continuously for only 290 minutes
on a gallon of gasoline. Find the power of the test to reject the null hypothesis, if the
second inventor is correct.
Solution
The steps required to compute power are presented below.
1. Define the region of acceptance.
Earlier, we showed that the region of acceptance for this problem consists of the values
between 294.45 and 305.55.
2. Specify the critical value.
The null hypothesis tests the hypothesis that the run time of the engine is 300 minutes.
We are interested in determining the probability that the hypothesis test will reject the

null hypothesis, if the true run time is actually 290 minutes. Therefore, the critical value
is 290. Another way to express the critical value is through effect size. The effect size is
equal to the critical value minus the hypothesized value. Thus, effect size is equal to 290
- 300 = -10
3. Compute power.
The power of the test is the probability of rejecting the null hypothesis, assuming that
the true population mean is equal to the critical value. Since the region of acceptance is
294.45 to 305.55, the null hypothesis will be rejected when the sampled run time is less
than 294.45 or greater than 305.55.
Therefore, we need to compute the probability that the sampled run time will be less
than 294.45 or greater than 305.55. To do this, we make the following assumptions:

The sampling distribution of the mean is normally distributed. (Because the


sample size is relative large, this assumption can be justified by the central limit
theorem.)
The mean of the sampling distribution is the critical value, 290.
The standard deviation of the sampling distribution is 2.83. The standard
deviation of the sampling distribution was computed in a previous lesson

Given these assumptions, we first assess the probability that the sample run time will be
less than 294.45. This is easy to do, using the Normal Calculator. We enter the following
values into the calculator: value = 294.45; mean = 290, and standard deviation = 2.83.
Given these inputs, we find that the cumulative probability is 0.94207551. This means
the probability that the sample mean will be less than 294.45 is 0.942.
Next, we assess the probability that the sample mean is greater than 305.55. Again, we
use the Normal Calculator. We enter the following values into the calculator: value =
305.55; mean = 290; and standard deviation = 2.83. Given these inputs, we find that the
probability that the sample mean is less than 305.55 (i.e., the cumulative probability) is
0.99999998. Thus, the probability that the sample mean is greater than 305.55 is 1 0.99999998 or 0.00000002.
The power of the test is the sum of these probabilities: 0.94207551 + 0.00000002 =
0.94207553. This means that if the true average run time of the new engine were 290
minutes, we would correctly reject the hypothesis that the run time was 300 minutes
94.2 percent of the time. Hence, the probability of a Type II error would be very small.
Specifically, it would be 1 minus 0.942 or 0.058.
IMPORTANT STATISTICAL DEFINITIONS OF HYPOTHESIS TESTING
Alpha - The significance level of a test of hypothesis that denotes the probability of
rejecting a null hypothesis when it is actually true; In other words, it is the probability of
committing a Type I error.

Alternative Hypothesis - A hypothesis that takes a value of a population parameter


different from that used in the null hypothesis
Beta - The probability of not rejecting a null hypothesis when it actually is false. In
other words, it is the probability of committing a Type II error.
Critical region - The set of values of the test statistic that will cause us to reject the
null hypothesis.
Critical Value - The first (or 'boundary') value in the critical region.
Decision rule - If the calculated test statistic falls within the critical region, the null
hypothesis H0 is rejected. In contrast, if the calculated test statistic does not fall within
the critical region, the null hypothesis is not rejected.
F-distribution - A continuous distribution that has two parameters (df for the
numerator and df for the denominator). It is mainly used to test hypotheses concerning
variances.
F-ratio - In ANOV A, it is the ratio of between column variance to within column
variance.
Hypothesis - An unproven proposition or supposition that tentatively explains a
phenomenon.
Null hypothesis - A statement about a status quo about a population parameter that is
being tested.
One-tailed test - A statistical hypothesis test in which the alternative hypothesis is
specified such that only one direction of the possible distribution of values is considered.
Power of hypothesis test - The probability of rejecting the null hypothesis when it is
false.
Significance level - The value of that gives the probability of rejecting the null
hypothesis when it is true. This gives rise to Type I error.
Test criteria - Criteria consisting of (i) specifying a level of significance ex, (ii)
determining a test statistic, (iii) determining the critical region(s), and (iv) determining
the critical value(s)
Test Statistic - The value of Z or t calculated for a sample statistic such as the sample
mean or the sample proportion.
Two-tail test - A statistical hypothesis test in which the alternative hypothesis is stated
in such a way that it includes both the higher and the lower values of a parameter than
the value specified in the null hypothesis.

Type I error - An error caused by rejecting a null hypothesis that is true.


Type II error - An error caused by failing to reject a null hypothesis that is not true.
SUMMARY
This chapter had given clear picture about the process of Hypothesis Testing and to find
out the details of Null hypothesis and Alternate hypothesis. This lesson also has given
the significance and differences between Type I error and Type II error. The detailed
step-wise procedures were given to perform the hypothesis testing. This chapter also
gives details about the calculation of P value.
KEY TERMS

Hypothesis testing
Null hypothesis
Alternative Hypothesis
Type I error
Type II error
Probability value
One & two tailed tests
Decision rule
Confidence interval and Statistical significance

IMPORTANT QUESTIONS
1. What do you mean by Hypothesis testing?
2. Define: Null hypothesis and Alternative Hypothesis.
3. Write down the steps to be performed for Hypothesis test
4. What are the differences between Type I and Type II error?
5. How will calculate the Probability value?
6. What do you understand by one and two tailed test?
7. Define Decision rule
8. What is the significance of understanding of confidence intervals?

- End of Chapter LESSON - 7

CHI-SQUARE TEST

OBJECTIVES

To learn the significance of Statistical tests.


To find out the details of Chi-square test, F-test and t-test
To learn the detailed procedures for doing above mentioned tests
To understand the concept of Measures of association

STRUCTURE
Bivariate tabular analysis
Chi-square requirements
Computing chi-square
Measure of Association
INTRODUCTION
Chi square is a non-parametric test of statistical significance for bivariate tabular
analysis (also known as crossbreaks). Any appropriately performed test of statistical
significance lets you know the degree of confidence you can have in accepting or
rejecting a hypothesis. Typically, the hypothesis tested with chi square is whether or not
two different samples (of people, texts, whatever) are different enough in some
characteristic or aspect of their behavior that we can generalize from our samples, that
the populations from which our samples are drawn are also different in the behavior or
characteristic.
A non-parametric test, like chi square, is a rough estimate of confidence; it accepts
weaker, less accurate data as input than parametric tests (like t-tests and analysis of
variance, for example) and therefore has less status in the pantheon of statistical tests.
Nonetheless, its limitations are also its- strengths; because chi square is more 'forgiving'
in the data it will accept, it can be used in a wide variety of research contexts.
Chi square is used most frequently to test the statistical significance of results reported
in bivariate tables, and interpreting bivariate tables is integral to interpreting the results
of a chi square test, so we'll take a look at bivariate tabular (crossbreak) analysis.
Bivariate Tabular Analysis

Bivariate tabular (crossbreak) analysis is used when you are trying to summarize the
intersections of independent and dependent variables and understand the relationship
(if any) between those variables. For example, if we wanted to know if there is any
relationship between the biological sex of American undergraduates at a particular
university and their footwear preferences, we might select 50 males and 50 females as
randomly as possible, and ask them, "On average, do you prefer to wear sandals,
sneakers, leather shoes, boots, or something else?" In this example, our independent
variable is biological sex (in experimental research, the independent variable is actively
manipulated by the researcher, for example, whether or not a rat gets a food pellet when
it pulls on a striped bar. In most sociological research, the independent variable is not
actively manipulated in this way, but controlled by sampling for, e.g., males vs. females).
Put another way, the independent variable is the quality or characteristic that you
hypothesize helps to predict or explain some other quality or characteristic (the
dependent variable). We control the independent variable (and as much else as possible
and natural) and elicit and measure the dependent variable to test our hypothesis that
there is some relationship between them. Bivariate tabular analysis is good for asking
the following kinds of questions:
1. Is there a relationship between any two variables in the data?
2. How strong is the relationship in the data?
3. What is the direction and shape of the relationship in the data?
4. Is the relationship due to some Is the relationship due to some intervening variable(s)
in the data?
To see any patterns or systematic relationship between biological sex of undergraduates
at University of X and reported footwear preferences, we could summarize our results in
a table like this:
Table 1: Male and Female Undergraduate Footwear Preferences
Sandals Sneakers

Leather
Boots
shoes

Other

Male
Female

Depending upon how our 50 male and 50 female subjects responded, we could make a
definitive claim about the (reported) footwear preferences of those 100 people.
In constructing bivariate tables, typically values on the independent variable are arrayed
on vertical axis, while values on the dependent variable are arrayed on the horizontal
axis. This allows us to read 'across' from hypothetically 'causal' values on the

independent variable to their 'effects', or values on the dependent variable How you
arrange the values on each axis should be guided "iconically" by your research
question/hypothesis. For example, if values on an independent variable were arranged
from lowest to highest value on the variable and values on the dependent variable were
arranged left to right from lowest to highest, a positive relationship would show up as a
rising left to right line. (But remember, association does not equal causation; an
observed relationship between two variables is not necessarily causal.)
Each intersection/cell-of a value on the independent variable and a value on the
independent variable-reports the result of how many times that combination of values
was chosen/observed in the sample being analyzed. (So you can see that crosstabs are
structurally most suitable for analyzing relationships between nominal and ordinal
variables. Interval and ratio variables will have to first be grouped before they can "fit"
into a bivariate table.) Each cell reports, essentially, how many subjects/observations
produced that combination of independent and dependent variable values? So, for
example, the top left cell of the table above answers the question: "How many male
undergraduates at University of X prefer sandals?" (Answer: 6 out of the 50 sampled.)
Table 1b: Male and Female Undergraduate Footwear Preferences
Sandals Sneakers
Male

17

Female

13

Leather
Boots
shoes
13
9
7

Other
5

16

Reporting and interpreting cross tabs is the most easily done by converting raw
frequencies (in each cell) into percentages of each cell within the values or categories of
the independent variable. For example, in the footwear preferences table above, total
each row, then divide each cell by its own total, and multiply that faction by 100.
Table 1c: Male and Female Undergraduate Footwear Preferences
(percentages)
Sandals Sneakers
Male

12

34

Female

26

10

Leather
Boots
shoes
26
18
14

32

Other

10

50

18

50

Percentages basically standardize cell frequencies as if there were 100


subjects/observations in each category of the independent variable. This is useful for
comparing across values on the independent variable, but that usefulness comes at the
price of a generalization--from the actual number of subjects/observations in that

column in your data to a hypothetical 100 subjects/observations. If the raw row total
was 93, then percentages do little violence to the raw scores; but if the raw total is 9,
then the generalization (on no statistical basis, i.e., with no knowledge of samplepopulation representativeness) is drastic. So you should provide that total N at the end
of each row/independent variable category (for replicability and to enable the reader to
assess your interpretation of the table's meaning).
With this caveat in mind, you can compare the patterns of distribution of
subjects/observations along the dependent variable between the values of the
independent variable: e.g., compare male and female undergraduate footwear
preference. (For some data, plotting the results on a line graph can also help you
interpret the results: i.e., whether there is a positive (/), negative (), or curvilinear (/, /)
relationship between the variables). Table 1c shows that within our sample, roughly
twice as many females preferred sandals and boots as males, and within our sample,
about three times as many men preferred sneakers as women and twice as many men
preferred leather shoes. We might also infer from the 'Other' category that female
students within our sample had a broader range of footwear preferences than did male
students.
Generalizing from Samples to Populations
Converting raw observed values or frequencies into percentages does allow us to see
more easily patterns in the data, but that is all we can see: what is in the data.
Knowing with great certainty the footwear preferences of a particular group of 100
undergraduates at University of X is of limited use to us; we usually want to measure a
sample in order to know something about the larger populations from which our
samples were drawn. On the basis of raw observed frequencies (or percentages) of a
samples behavior or characteristics, we can make claims about the sample itself, but we
cannot generalize to make claims about the population from which we drew our sample,
unless we submit our results to a test of statistical significance. A test of statistical
significance tells us how confidently we can generalize to a larger (unmeasured)
population from a (measured) sample of that population.
How does chi square do this? Basically, the chi square test of statistical significance is a
series of mathematical formulas which compare the actual observed frequencies of some
phenomenon (in our sample) with the frequencies we would expect if there were no
relationship at all between the two variables in the larger (sampled) population. That is,
chi square tests our actual results against the null hypothesis and assesses whether the
actual results are different enough to overcome a certain probability that they are due to
sampling error. In a sense, chi-square is a lot like percentages; it extrapolates a
population characteristic (a parameter) from the sampling characteristic (a statistic)
similarly to the way percentage standardizes a frequency to a total column N of 100. But
chi-square works within the frequencies provided by the sample and does not inflate (or
minimize) the column and row totals.
Chi Square Requirements

As mentioned before, chi square is a nonparametric test. It does not require the sample
data to be more or less normally distributed (as parametric tests like t-tests do),
although it relies on the assumption that the variable is normally distributed in the
population from which the sample is drawn.
But chi square does have some requirements:
1. The sample must be randomly drawn from the population.
As with any test of statistical significance, your data must be from a random sample of
the population to which you wish to generalize your claims.
2. Data must be reported in raw frequencies (not percentages).
You should only use chi square when your data are in the form of raw frequency counts
of things in two or more mutually exclusive and exhaustive categories. As discussed
above, converting raw frequencies into percentages standardizes cell frequencies as if
there were 100 subjects/observations in each category of the independent variable for
comparability. Part of the chi square mathematical procedure accomplishes this
standardizing, so computing the chi square of percentages would amount to
standardizing an already standardized measurement
3. Measured variables must be independent.
Any observation must fall into only one category or value on each variable. In our
footwear example, our data are counts of male versus female undergraduates expressing
a preference for five different categories of footwear. Each observation/subject is
counted only once, as either male or female (an exhaustive typology of biological sex)
and as preferring sandals, sneakers, leather shoes, boots, or other kinds of footwear. For
some variables, no 'other' category may be needed, but often 'other' ensures that the
variable has been exhaustively categorized. (For some kinds of analysis, you may need to
include an "uncodable" category.) In any case, you must include the results for the whole
sample.
4. Values/categories on independent and dependent variables must be mutually
exclusive and exhaustive.
Furthermore, you should use chi square only when observations are independent: i.e.,
no category or response is dependent upon or influenced by another. (In linguistics,
often this rule is fudged a bit. For example, if we have one dependent variable/column
for linguistic feature X and another column for number of words spoken or written
(where the rows correspond to individual speakers/texts or groups of speakers/texts
which are being compared), there is clearly some relation between the frequency of
feature X in a text and the number of words in a text, but it is a distant, not immediate
dependency.)
5. Observed frequencies cannot be too small.

Chi-square is an approximate test of the probability of getting the frequencies you've


actually observed if the null hypothesis were true. It's based on the expectation that
within any category, sample frequencies are normally distributed about the expected
population value. Since (logically) frequencies cannot be negative, the distribution
cannot be normal when expected population values are close to zero-since the sample
frequencies cannot be much below the expected frequency while they can be much
above it (an asymmetric/non-normal distribution). So, when expected frequencies are
large, there is no problem with the assumption of normal distribution, but the smaller
the expected frequencies, the less valid are the results of the chi-square test. We'll
discuss expected frequencies in greater detail later, but for now remember that expected
frequencies are derived from observed frequencies. Therefore, if you have cells in your
bivariate table which show very low raw observed frequencies (5 or below), your
expected frequencies may also be too low for chi square to be appropriately used. In
addition, because some of the mathematical formulas used in chi square use division, no
cell in your table can have an observed raw frequency of 0.
The following minimum frequency thresholds should be obeyed:
for a 1 x 2 or 2 x 2 table, expected frequencies in each cell should be at least 5;
for a 2 x 3 table, expected frequencies should be at least 2;
for a 2 x 4 or 3 x 3 or larger table, if all expected frequencies but one are at least 5 and
if the one small cell is at least 1, chi-square is still a good approximation.
In general, the greater the degrees of freedom (i.e., the more values/categories on the
independent and dependent variables), the more lenient the minimum expected
frequencies threshold. (We'll discuss degrees of freedom in a moment.)
Collapsing Values
A brief word about collapsing values/categories on a variable is necessary. First,
although categories on a variable, especially a dependent variable, may be collapsed,
they cannot be excluded from a chi-square analysis. That is, you cannot arbitrarily
exclude some subset of your data from your analysis. Second, a decision to collapse
categories should be carefully motivated, with consideration for preserving the integrity
of the data as it was originally collected. (For example, how could you collapse the
footwear preference categories in our example and still preserve the integrity of the
original question/data? You can't, since there's no way to know if combining, e.g., boots
and leather shoes versus sandals and sneakers is true to your subjects' typology of
footwear.) As a rule, you should perform a chi square on the data in its uncollapsed
form; if the chi square value achieved is significant, then you may collapse categories to
test subsequent refinements of your original hypothesis.
Computing Chi Square

Lets walk through the process by which a chi square value is computed, using Table 1b
above (renamed Table 1d below).
The first step is to determine our threshold of tolerance for error. That is, what odds are
we willing to accept that we are wrong in generalizing from the results in our sample to
the population it represents? Are we willing to stake a claim on a 50 percent chance that
we're wrong? a 10 percent chance? a five percent chance? 1 percent? The answer
depends largely on our research question and the consequences of being wrong. If
people's lives depend on our interpretation of our results, we might want to take only 1
chance in 100,000 (or 1,000,000) that we're wrong. But if the stakes are smaller, for
example, whether or not two texts use the same frequencies of some linguistic feature
(assuming this is not a forensic issue in a capital murder case!), we might accept a
greater probability--1 in 100 or even 1 in 20 - that our data do not represent the
population we're generalizing about. The important thing is to explicitly motivate your
threshold before you perform any test of statistical significance, to minimize any
temptation for post hoc compromise of scientific standards. For our purposes, we'll set a
probability of error threshold of 1 in 20, or p < .05, for our Footwear study).
The second step is to total all rows and columns.
Table 1d: Male and Female Undergraduate Footwear Preferences: Observed
Frequencies with Row and Column Totals
Sandals Sneakers

Leather
Boots
shoes
13
9

Other

Total

50

Male

17

Female

13

16

50

Total

19

22

20

25

14

100

Remember, that chi square operates by comparing the actual, or observed frequencies in
each cell in the table to the frequencies we would expect if there were no relationship at
all between the two variables in the populations from which the sample is drawn. In
other words, chi square compares what actually happened to what hypothetically would
have happened if all other things were equal' (basically, the null hypothesis). If our
actual results are sufficiently different from the predicted null hypothesis results, we can
reject the null hypothesis and claim that a statistically significant relationship exists
between our variables.
Chi square derives a representation of the null hypothesisthe 'all other things being
equal' scenarioin the following way. The expected frequency in each cell is the product
of that cell's row total multiplied by that cell's column total, divided by the sum total of
all observations. So, to derive the expected frequency of the "Males who prefer Sandals"
cell, we multiply the top row total (50) by the first column total (19) and divide that
product by the sum total 100: (50 X 19)/100 = 9.5. The logic of this is that we are

deriving the expected frequency of each cell from the union of the total frequencies of
the relevant values on each variable (in this case, Male and Sandals), as a proportion of
all observed frequencies (across all values of each variable). This calculation is
performed to derive the expected frequency of each cell, as shown in Table 1e below (the
computation for each cell is listed below:

(Notice that because we originally obtained a balanced male/female sample, our male
and female expected scores are the same. This usually will not be the case.)
We now have a comparison of the observed results versus the results we would expect if
the null hypothesis were true. We can informally analyze this table, comparing observed
and expected frequencies in each cell (Males prefer sandals less than expected), across
values on the independent variable (Males prefer sneakers more than expected, Females
less than expected), or across values on the dependent variable (Females prefer sandals
and boots more than expected, but sneakers and shoes less than expected). But so far,
the extra computation doesn't really add much more information than interpretation of
the results in percentage form. We need some way to measure how different our

observed results are from the null hypothesis. Or, to put it another way, we need some
way to determine whether we can reject the null hypothesis, and if we can, with what
degree of confidence that we're not making a mistake in generalizing from our sample
results to the larger population.
Logically, we need to measure the size of the difference between the pair of observed
and expected frequencies in each cell. More specifically, we calculate the difference
between the observed and expected frequency in each cell, square that difference, and
then divide that product by the difference itself. The formula can be expressed as:
(O - E)2 / E
Squaring the difference ensures a positive number, so that we end up with an absolute
value of differences. If we didn't work with absolute values, the positive and negative
differences across the entire table would always add up to 0. (You really understand the
logic of chi square if you can figure out why this is true.) Dividing the squared difference
by the expected frequency essentially removes the expected frequency from the
equation, so that the remaining measures of observed/expected difference are
comparable across all cells.
So, for example, the difference between observed and expected frequencies for the
Male/Sandals preference is calculated as follows:
1. Observed (6) - Expected (9.5) = Difference (-3.5)
2. Difference (-3.5) squared = 12.25
3. Difference squared (12.25) divided by Expected (9.5) = 1.289
The sum of all products of this calculation on each cell is the total chi square value for
the table. The computation of chi square for each cell is listed below:

(Again, because of our balanced male/female sample, our row totals were the same, so
the male and female observed expected frequency differences were identical. This is
usually not the case)
The total chi square value for Table 1 is 14.026.
Interpreting the Chi Square Value
We now need some criterion or yardstick against which to measure the table's chi square
value, to tell us whether or not it is significant. What we need to know is the probability
of getting a chi square value of a minimum given size even if our variables are not
related at all in the larger population from which our sample was drawn. That is, we
need to know how much larger than 0 (the absolute chi square value of the null
hypothesis) our table's chi square value must be before we can confidently reject the null
hypothesis. The probability we seek depends in part on the degrees of freedom of the
table from which our chi square value is derived.

Degrees of freedom
Mechanically, a table's degrees of freedom (df) can be expressed by the following
formula: df = (r - 1)(c - 1)
That is, a table's degrees of freedom equals the number of rows in the table minus one
multiplied by the number of columns in the table minus one (for 1 x 2 table, df = k - 1,
where k = number of values/categories on the variable).
Degrees of freedom is an issue because of the way in which expected values in each cell
are computed from the row and column totals of each cell. All but one of the expected
values in a given row or column are free to vary (within the total observed-and therefore
expected) frequency of that row or column); once the free to vary expected cells are
specified, the last one is fixed by virtue of the fact that the expected frequencies must
add up to the observed row and column totals (from which they are derived).
Another way to conceive of a table's degrees of freedom is to think of one row and one
column in the table as fixed, with the remaining cells free to vary. Consider the following
visuals (where X = fixed):

df = (r 1) (c 1) = (3 1) (2 1) = 2 x 1 = 2

df = (r 1) (c 1) = (5 1) (3 1) = 4 x 2 = 8
So, for our Table 1,
Table 1: Male and Female Undergraduate Footwear Preferences
Sandals Sneakers
Male

Female

Leather
Boots
shoes
X
X

Other
X

df = (2 - 1) (5 1) = 1 x 4 = 4
In a statistics book, the sampling distribution of chi square (also known as critical values
of chi squares typically listed in an appendix. You read down the column representing
your previously chosen probability of error threshold (e.g., p < 0.5) and across the row
representing the degrees of freedom in your table. If your chi square value is larger than
the critical value in that cell, your data presents a statistically significant relationship
between the variables in your table.
Table 1s chi square value of 14.026 with 4 degrees of freedom, handily clears the related
critical value of 9.49, so we can reject the null hypothesis and affirm the claim that male
and female undergraduates at University of X differ in their (self-reported) footwear
preferences.

Statistical significance does not help you to interpret that nature or explanation of that
relationship; that must be done by other means (including bivariate tabular analysis of
the data). But a statistically significant chi square value does denote the degree of
confidence you may hold that the relationship between variables described in your
results is systematic in the larger population and not attributable to random error.
Statistical significance also does not ensure substantive significance. A large enough
sample may demonstrate a statistically significant relationship between two variables,
but that relationship may be trivially weak one. Statistical significance means only that
the pattern of distribution and relationship between the variables which is found in the
data from a sample can be confidently generalized to the larger population from which
the sample was randomly drawn. By itself it does not ensure that the relationship is
theoretically or practically important, or even very large.
Measures of Association
While the issue of theoretical or practical importance of a statistically significant result
cannot be quantified, the relative magnitude of a statistically significant relationship can
be measured. Chi Square allows you to make decisions about whether there is a
relationship between two or more variables; if the null hypothesis is rejected, we
conclude that there is a statistically significant relationship between the variables. But
we frequently want a measure of the strength of that relationship - an index of degree of
correlation, a measure of the degree of association between the variables represented in
our table (and data). Luckily, several measures of association can be derived from a
tables chi square value.
For tables larger than 2 x 2 (like our Table 1), a measure called 'Cramer's phi' is derived
by the following formula (where N is the total number of observations, and k is the
smaller of the number of rows or columns).
Cramers phi = sqrt [(chi square) / (N x (k-1)]
So, for our Table 1 (2 x 5), Cramers phi will be computed as follows:
1. N (k - 1) = 100 (2 1) = 100
2. Chi square / 100 = 14.026 / 100 = 0.14
3. Square root of 0.14 = 0.37
The product is interpreted as a Pearson r (that is, as a correlation coefficient).
(For 2X2 tables, a measure called 'phi' is derived by dividing the table's chi square value
by N (the total number of observations) and then taking the square root of the product.
Phi is also interpreted as a Pearson r.)

A complete account of how to interpret correlation coefficients is unnecessary for


present purposes. It will suffice to say that r2 is a measure called shared variance.
Shared variance is the portion of the total behavior (or distribution) of the variables
measured in the sample data which is accounted for by the relationship we've already
detected with our chi square. For Table 1, r2 = 0.137, so approximately 14% of the total
footwear preference story is explained/predicted by biological sex.
Computing a measure of association like phi or Cramer's phi is rarely done in
quantitative linguistic analyses, but it is an important benchmark of just 'how much' of
the phenomenon under investigation has been explained. For example, Table 1's
Cramer's phi of 0.37 (r2 = 0.137) means that there are one or more, variables still
undetected which, cumulatively, account for and predict 86% of footwear preferences.
This measure, of course, doesn't begin to address the nature of the relation(s) between
these variables, which is a crucial part of any adequate explanation or theory.
SUMMARY
The above chapter has given the framework for performing key statistical tests like chisquare test. Chi-square is a non parametric test of statistical significance for bivariate
tabular analysis.
KEY WORDS

Bivariate tabular analysis


Chi square test
Measure of association
Degrees of freedom

IMPORTANT QUESTIONS
1. What do you mean by bivariate tabular analysis?
2. What are the statistical applications in chi square tests?
3. How will you calculate the degree of freedom?

- End of Chapter LESSON - 8


T-TEST, F-TEST

OBJECTIVES

To know the significance of statistical list


To find out the details of F-test and T-test
To learn the detailed procedures for doing above mentioned test.

STRUCTURE

F-test
Cumulative probability and F distribution
F-Test for equality
T-test
Two sample t-test for equal means.

F-TEST
The f statistic, also known as an f value, is a random variable that has an F distribution.
The F Distribution
The distribution of all possible values of the f statistic is called an F distribution, with v1
= n1 1 and v2 = n2 1 degrees of freedom.
The curve of the F distribution depends on the degrees of freedom, v1 and v2. When
describing an F distribution, the number of degrees of freedom associated with the
standard deviation in the numerator of the f statistic is always stated first. Thus, f(5,9)
would refer to an F distribution with v1 = 5 and v2 = 9 degrees of freedom; whereas f(9,5)
would refer to an F distribution with v1 = 9 and v2 = 5 degrees of freedom. Note that the
curve represented by f(5,9) would differ from the curve represented by f(9,5).
The F distribution has the following properties:
The mean of the distribution is equal to v1 / (v2 2)
The variance is equal to
v22 (v1 + 2)
---------------------------v1 (v2 2) (v2 4)
Cumulative Probability and the F Distribution
Every f statistic can be associated with a unique cumulative probability. This cumulative
probability represents the likelihood that the f statistic is less than or equal to a specified
value. Here are the steps required to compute an f statistic:

1. Select a random sample of size n1 from a normal population, having a standard


deviation equal to 1
2. Select an independent random sample of size n2 from a normal population, having a
standard deviation equal to 2
The f statistic is the ratio of and s12/12 and s22/22
The following equivalent equations are commonly used to compute an f statistic:
f = [s12 / 12] / [s22 / 22]
f = [s12 x 22] / [s22 x 12]
f = [X12 / v1] / [X22 / v2]
f = [X12 x v2] / [X22 x v1]
where
1 = standard deviation of population 1
s1 = standard deviation of the sample drawn from population 1
2 = standard deviation of population 2
s2 = standard deviation of the sample drawn from population 2
X12 = chi-square static for the sample drawn from population 1
v1 = degrees of freedom for X12
X22 = chi-square statistic for the sample drawn from population 2
v2 = degrees of freedom for X22
Degrees of freedom v1 = n1 - 1, and
Degrees of freedom v2 = n2 1.
Sample Problems
Example

Suppose you randomly select 7 women from a population of women and 12 men from a
population of men. The table below shows the standard deviation in each sample and in
each population.

Compute the f statistic.


Solution
The f statistic can be computed from the population and sample standard deviations,
using the following equation:
f = [s12 / 12] / [s22 / 22]
where,
1 is the standard deviation of population 1
s1 = standard deviation of the sample drawn from population 1
2 = standard deviation of population 2
s2 = standard deviation of the sample drawn from population 2
As you can see from the equation, there are actually two ways to compute an f statistic
from these data. If the women's data appears in the numerator, we can calculate an f
statistic as follows:
f = (35 / 30) / (45 / 50) = 1.66667 / 0.9 = 1.85
On the other hand, if the men's data appears in the numerator, we can calculate an f
statistic as follows:
f = (45 / 50) / (35 / 30) = 0.9 / 1.66667 = 0.54
This example illustrates the importance of specifying the degrees of freedom associated
with the numerator and denominator of an f statistic. This topic is continued in the next
example.

F-Test for Equality of Two Standard Deviations


An F-test (Snedecor and Cochran, 1983) is used to test if the standard deviations of two
populations are equal. This test can be a two-tailed test or a one-tailed test. The twotailed version tests against the alternative that the standard deviations are not equal.
The one-tailed version only tests in one direction, that is, the standard deviation from
the first population is either greater than or less than (but not both) the second
population standard deviation. The choice is determined by the problem. For example, if
we are testing a new process, we may only be interested in knowing if the new process is
less variable than the old process.
We are testing the hypothesis that the standard deviations for sample one and sample
two are equal. The output is divided into four sections.
1. The first section prints the sample statistics for sample one used in the computation of
the F-test.
2. The second section prints the sample statistics for sample two used in the
computation of the F-test
3. The third section prints the numerator and denominator standard deviations, the Ftest statistic value, the degrees of freedom, and the cumulative distribution function
(cdf) value of the F-test statistic. The F-test statistic cdf value is an alternative way of
expressing the critical value. This cdf value is compared to the acceptance interval
printed in section four. The acceptance interval for a two-tailed test is (0,1)
4. The fourth section prints the conclusions for a 95% test since this is the most common
case. Results are printed for an upper one-tailed test. The acceptance interval column is
stated in terms of the cdf value printed in section three. The last column specifies
whether the null hypothesis is accepted or rejected. For a different significance level, the
appropriate conclusion can be drawn from the F-test statistic cdf value printed in
section four. For example, for a significance level of 0.10, the corresponding acceptance
interval become (0.000,0.9000).
The F-test can be used to answer the following questions:
1. Do two samples come from populations with equal standard deviations?
2. Does a new process, treatment, or test reduce the variability of the current process?
T-TEST
We have seen that the central limit theorem can be used to describe the sampling
distribution of the mean, as long as two conditions are met:
1. The sample size must be sufficiently large (at least 30).

2. We need to know the standard deviation of the population, which is denoted by .


But sample sizes are sometimes small and often we do not know the true population
standard deviation. When either of these problems occurs, statisticians rely on the
distribution of the t statistic (the t-score), whose values are given by:
T = [x ] / [s / sqrt(n)]
where,
x = sample mean
= population mean
s = standard deviation of the sample
n = sample size
The distribution of this t statistic is called the t-distribution or the Student tdistribution. The t-distribution can be used whenever samples are drawn from
populations possessing a bell-shaped distribution (i.e., approximately normal).
The t-distribution has the following properties

The mean of the distribution is equal to 0.


The variance is equal to v / (v - 2), where v is the degrees of freedom (see next
section) and v > 2.

Parametric tests provide inferences for making statements about the means of parent
populations. A t-test is commonly used for this purpose. This test is based on the
Student's t statistic. The t statistic assumes that the variable is normally distributed, the
mean is known (or assumed to be known), and the population variance is estimated
from the sample. Assume that the random variable X is normally distributed, with mean
and unknown population variance 2, which is estimated by the sample variance S2.
Recall that the standard deviation of the sample mean, X, is estimated as sx = s/n. Then,
t = (X - ) / sx is distributed with n - 1 degrees of freedom.
The t-distribution is similar to the normal distribution in appearance. Both distributions
are bell-shaped and symmetric. However, as compared to the normal distribution, that
distribution has more area in the tails and less in the center. This is because population
variance 2 is unknown and is estimated by the sample variance S2. Given the
uncertainty in the value of S2, the observed values of t are more variable than those of z.
Thus we must go a large number of standard deviations from 0 to encompass a certain
percentage of values from the t distribution than is the case with the normal
distribution. Yet, as the number of freedom increases, the t distribution approaches the
normal distribution. In large samples of 120 or more, the t distribution and the normal
distribution are virtually indistinguishable. Table 4 in the statistical appendix shows

selected percentiles of the t distribution. Although normality is assumed, the t test is


quite robust to departures from normality.
The procedure for hypothesis testing, for the special case when the t statistic is used, is
as follows:
1. Formulate the null (H0) and the alternative (H1) hypotheses.
2. Select the appropriate formula for the t statistic.
3. Select a significance level, a, for testing H0. Typically, the 0.05 level is selected.
4. Take one or two samples and compute the mean and standard deviation for each
sample.
5. Calculate the t statistic assuming H0 is true.
6. Calculate the degrees of freedom and estimate the probability of getting a more
extreme value of the statistic from Table 4. (Alternatively, calculate the critical value of
statistic.)
7. If the probability computed in step 6 is smaller than the significance level selected in
reject H0. If the probability is larger, do not reject H0. (Alternatively, if the value
calculated t statistic in step 5 is larger than the critical value determined in step 6, reject
H0. If the calculated value is smaller than the critical value, do not reject H0). To reject
H0 does not necessarily imply that H0 is true. It only means that the result is not
significantly different than that assumed by H0.
8. Express the conclusion reached by the t-test in terms of the marketing research
problem.
Degrees of Freedom
There are actually many different t distributions. The particular form of the t
distribution is determined by its degrees of freedom. The degree of freedom refers to the
number of independent observations in a sample.
The number of independent observations is often equal to the sample size minus one.
Hence, the distribution of the t statistic from samples of size 8 would be described by a t
distribution having (8 1) = 7 degrees of freedom. Similarly, a t-distribution having 15
degrees of freedom would be used with a sample of size 16.
The t-distribution is symmetrical with a mean of zero. Its standard deviation is always
greater than 1, although it is close to 1 when there are many degrees of freedom. With
infinite degrees of freedom, the t distribution is the same as the standard normal
distribution.

Sampling Distribution of the Mean


When the sample size is small (< 30), the mean and standard deviation of the sampling
distribution can be described as follows:
x = and x = s * sqrt[1/n 1/N]
where,
x = mean of the sampling distribution
= mean of the population
x = standard error (i.e., the standard deviation of the sampling distribution)
s = standard deviation of the sample
n = sample size
N = population size
Probability and the Student t Distribution
When a sample of size n is drawn from a population having a normal (or nearly normal)
distribution, the sample mean can be transformed into a t score, using the equation
presented at the beginning of this lesson. We repeat that equation here: T = [x ] / [s /
sqrt(n)]
where,
x = sample mean
= population mean
s = standard deviation of the sample
n = sample size
The degrees of freedom = n - 1
Every t score can be associated with a unique cumulative probability. This cumulative
probability represents the likelihood of finding a sample mean less than or equal to x,
given a random sample of size n.
Two-sample F-Test for Equal Means

The two-sample t-test (Snedecor and Cochran, 1989) is used to determine if two
population means are equal. A common application of this is to test if a new process or
treatment is superior to a current process or treatment.
There are several variations on this test:
1. The data may either be paired or not paired. By paired, we mean that there is a oneto-one correspondence between the values in the two samples. That is, if X 1, X2, , Xn
and Y1, Y2... , Yn are the two samples, then Xi corresponds to Yi. For paired samples, the
difference Xi - Yi is usually calculated. For unpaired samples, the sample sizes for the
two samples may or may not be equal. The formulas for paired data are somewhat
simpler than the formulas for unpaired data.
2. The variances of the two samples may be assumed to be equal or unequal. Equal
variances yield somewhat simpler formulas, although with computers this is no longer a
significant issue.
In some applications, you may want to adopt a new process or treatment only if it
exceeds the current treatment by some threshold. In this case, we can state the null
hypothesis in the form that the difference between the two populations means is equal
to some constant (1 2 = d0) where the constant is the desired threshold.
Interpretation of Output
1. We are testing the hypothesis that the population mean is equal for the two samples.
The output is divided into five sections.
2. The first section prints the sample statistics for sample one used in the computation
of the f-test
3. The second section prints the sample statistics for sample two used in the
computation of the t-test.
4. The third section prints the pooled standard deviation, the difference in the means,
the t-test statistic value, the degrees of freedom, and the cumulative distribution
function (cdf) value of the t-test statistic under the assumption that the standard
deviations are equal. The t-test statistic cdf value is an alternative way of expressing the
critical value. This cdf value is compared to the acceptance intervals printed in section
five. For an upper one-tailed test, the acceptance interval is (0,1-), the acceptance
interval for a two-tailed test is (/2, 1- /2), and the acceptance interval for a lower onetailed test is (,1).
5. The fourth section prints the pooled standard deviation, the difference in the means,
the t-test statistic value, the degrees of freedom, and the cumulative distribution
function (cdf) value of the t-test statistic under the assumption that the standard
deviations are not equal. The t-test statistic cdf value is an alternative way of expressing
the critical value. cdf value is compared to the acceptance intervals printed in section

five. For an upper one-tailed test, the alternative hypothesis acceptance interval is (1-,
1), the alternative hypothesis acceptance interval for a lower one-tailed test is (0,), and
the alternative hypothesis acceptance interval for a two-tailed test is (1-/2,1) or (0,/2).
Note that accepting the alternative hypothesis is equivalent to rejecting the null
hypothesis.
6. The fifth section prints the conclusions for a 95% test under the assumption that the
standard deviations are not equal since a 95% test is the most common case." Results
are given in terms of the alternative hypothesis for the two-tailed test and for the onetailed test in both directions. The alternative hypothesis acceptance interval column is
stated in terms of the cdf value printed in section four. The last column specifies
whether the alternative hypothesis is accepted or rejected. For a different significance
level, the appropriate conclusion can be drawn from the t-test statistic cdf value printed
in section four. For example, for a significance level of 0.10, the corresponding
alternative hypothesis acceptance intervals are (0,0.05) and (0.95,1), (0,0.10), and
(0.90,1).
Two-sample f-tests can be used to answer the following questions
1. Is process 1 equivalent to process 2?
2. Is the new process better than the current process?
3. Is the new process better than the current process by at least some pre-determined
threshold amount?
Matrices of t-tests
T-tests for dependent samples can be calculated for long lists of variables, and reviewed
in the form of matrices produced with case wise or pair wise deletion of missing data,
much like the correlation matrices. Thus, the precautions discussed in the context of
correlations also apply to t-test matrices:
a. The issue of artifacts caused by the pair wise deletion of missing data in t-tests and
b. The issue of "randomly" significant test values.
SUMMARY
The above chapter has given the frame work for performing key statistical tests like FTest and T-test. T-test and F-test are parametric tests. T-test is any statistical hypothesis
test in which the test statistic has a Students distribution if the null hypothesis is true.
KEY TERMS

Degrees of Freedom
T test

T distribution
F distribution
Sampling distribution

IMPORTANT QUESTIONS
1. How will you calculate the Degrees of Freedom?
2. Explain the procedures for performing t test?
3. What are the metrics of t test?
4. Explain the procedures for performing f test?
REFERENCE
1. Sumathi, S. and P. Saravanavel - Marketing research and Consumer Behaviour
2. Ferber, R., and Verdoorn, P.J., Research Methods in Business, New York-the
Macmillan Company 1962.
3. Ferber, R., Robert (ed.) Hand Book of Marketing Research. New York McGraw Hill,
Inc. 1948.

- End of Chapter LESSON 9


METHODS OF DATA COLLECTION

OBJECTIVES

To know the different types of Data and the sources of the same.
To learn the different data collection methods and its merits, demerits.
To understand the difference between Questionnaire and Interview Schedule.
To apply the suitable data collection method for the research.

STRUCTURE

Primary data
Secondary data
Interview
Questionnaire

Observation

INTRODUCTION
After defining the research problem and drawing the research design, the important
task of the researcher is Data collection. While deciding on the research method, the
method of data collection to be used for the study also should be planned.
The same of information and the manner in which data are collected could well make a
big difference to the effectiveness of the research project.
In data collection, the researcher should be very clear on what type of data is to be used
for the research. There are two types of data namely primary data and secondary data.
The method of collecting primary and secondary data differ since primary data are to be
originally collected while in case of secondary data, it is merely compilation of the
available data.
PRIMARY DATA
Primary data are generally, information gathered by the researcher for the purpose of
the project at hand. When the data are collected for the first time using experiments and
surveys, the data is known as primary data. So, in case of primary data, it is always the
responsibility of the researcher to decide on further processing of data.
There are several methods of data collection each with its advantages and
disadvantages.
The data collection methods include the following:
1. Interview- Face to face interview, Telephone interview, Computer assigned
interview, Interviews through electronic media
2. Questionnaire - These are personally administered sent through the mail, or
electronically administered
3. Observation - of individuals and events with or without video taping or audio
recording. Hence interviews, questionnaires and observation methods are three main
data collection methods.
Some of the other data collection methods used to collect primary data:
1. Warranty cards
2. Distributor Audits
3. Pantry Audits

4. Consumer Panels
SECONDARY DATA
As already mentioned, it the researcher who decides to collect secondary data for his
research that can be collected through various sources. In the case of secondary data,
the researcher may not face severe problems that are usually associated with primary
data collection.
Secondary data may either be published or unpublished data. Published data may be
available with the following sources:
1. Various publications of the central, state or local governments.
2. Various publications of foreign governments or of international bodies.
3. Technical and Trade Journals.
4. Books, Magazines, Newspapers.
5. Reports and publication from various associations connected with industry and
business.
6. Public records and statistics.
7. Historical documents.
Though there are various sources for secondary data, it is the responsibility of the
researcher that he should make a minute scrutiny of data in order to involve the data
more suitable and adequate.
INTERVIEWS
An interview is a purposeful discussion between two or more people. Interview can help
to gather valid and reliable data. There are several types of interviews.
Types of Interviews
Interviews may be conducted in a very formal manner, using structured and
standardized questionnaire for each respondent.
Interviews also may be conducted informally through unstructured conversation. Based
on formal nature and structure, the interviews are classified as follows:
1. Structured Interviews : These interviews involve the use of a set of predetermined
questions and of highly standardized techniques of recording. So, in this type of
interview a rigid proved method is followed.

2. Semi Structured interviews : These interviews may have a structured


questionnaire but the technique of interviewing may not have a proved method. These
interviews have more scope for discussion and recording of respondent's opinion and
views.
3. Unstructured Interviews : These interviews neither follow a system of predetermined questions nor a standardized technique of recording information. But the
unstructured interviews need an in depth knowledge and greater skill on the part of the
interviewers.
All the three types of interviews may be conducted by the interviewer by asking
questions generally in face to face contact. These interviews may take a form of direct
personal investigation or may be indirect oral investigation. In case of direct personal
investigation, the interviewer has to collect the information personally. So, it is the duty
of interviewer to be there in the spot to meet the respondents to collect data.
When this is not possible the interviewer may have to cross examine others who are
supposed to have knowledge about the problem and here information may be recorded.
Example: Commissions and committee appointed by Government. Depending on the
approaches of the interviewer, the interviews may be classified as:
1. Non-directive interviews : In these types of interviews, the interviewer is very
much free to arrange the form and order of questions. The questionnaire for these kinds
of interviews also may contain open ended questions where in the respondent also feel
free to respond to the questions.
2. Directive interviews : This is also a type of structured interview. In this method a
predetermined questionnaire is used and the respondent is express to limit the answers
only to the question asked. Market surveys and interviews by news paper
correspondents are the suitable examples.
3. Focused interviews : These methods of interviews are in between directive and
non - directive. Here the methods are neither fully standardized nor non-standardized.
Here the objective is to focus the attention of the respondents on a specific issue or
point. Example: A detective questioning a person regarding a crime committed in an
area.
4. In-depth interviews : In these interview methods, the respondents are encouraged
to express his thoughts on the topic of the study. In depth interviews are conducted to
get important aspects of psycho - social situations, which are otherwise not readily
evident.
The major strength of these kinds of interviews is their capacity to uncover the basic and
complete answers to the questions asked.
Advantages of interviews

Despite the variations in interviews techniques the following are the advantages of the
same.
1. Interviews may help to collect more information and also in depth information.
2. The unstructured interviews are more advantages since, there is always an
opportunity for the interviewers to restructure the questions.
3. Since the respondents are contacted for the information, there are always greater
advantages of creating support and collecting personal information also.
4. Interviews help the researcher to collect all the necessary information, here the
evidence no response will be very low.
5. It is also possible for the interviewer to collect additional information about the
environment, name, behavior and attitude of the respondents.
Disadvantages of interviews
1. Interviews are expensive methods especially in case of widely spread geographical
samples are taken.
2. There may be a possibility for the barriers in the case of both interviewer and the
respondent.
3. These methods are also time-consuming especially when the large samples are taken
for the study.
4. There may be a possibility for the respondent to hide the real opinion, so genuine data
may not be collected.
5. Sometimes there will be great difficult in adopting interview methods became fixing
appointment with the respondent itself may not be possible.
Hence, for successful implementation of the interview method, the interviewer should
be carefully selected, trained and briefed.
How to make interviews successful?
1. As mentioned above, the interviewers must be carefully selected and trained properly.
2. The interviewer should have the knowledge of exploring to collect the needed
information from the respondent.
3. Honesty and integrity of the interviewer also determines the outcomes of the
interview.

4. The support with the respondent should be created by the interviewer.


5. Qualities such as politeness, courteousness friendly and conversational are necessary
to make the interview successful.
Telephonic Interviews
A part from all the above the telephonic interviews are also conducted to collect the
data. Respondents are contacted over the phone to gather data. Telephonic interviews
are more flexible in timing - it is faster than other methods and this method is cheaper
method also. For these sorts of interviews no field's staff is required and also the
information can be recorded without causing embarrassment to respondents especially
when very personal questions are asked. But these methods are much restricted to the
respondents who have telephone facility. Possibility for the biased replies is relatively
more and since there is not personal touch by both there is a greater possibility for non
answered questions.
QUESTIONNAIRES
Questionnaires are widely used for data collection in social sciences, research
particularly in surveys. This is been accepted as a reliable not for gathering data from
large, diverse and scattered social groups. Bogardus describes the questionnaire as a list
of questions sent to a number of persons for their answers and which obtains standards
results that can be tabulated and treated statistically.
There are two types of questionnaires, also which are structured and unstructured.
The design of the questionnaire may vary based on the way it is administered. The
questionnaire methods are most extensively used in economic and business surveys.
Structured questionnaire
These contain definite concrete and preordained questions. The answer collected using
this structured questionnaire is very precise and there is no vagueness and ambiguity.
The structured questionnaire may have the following types:
1. Closed - form questionnaire: Questions are set in such a way that leaves only few
alternative answers. Example: Yes / No type questions.
2. Open - ended questionnaire: Here the respondent has the choice of using his own
style, expression of language, length and perception. The respondents are not restricted
in his replies.
Unstructured questionnaire

The questions in this questionnaire are not structured in advance. These sorts of
questionnaire give more scopes for variety of answers. Mainly to conduct interviews
where in different responses are expected, these type of questionnaire are used.
The researcher should be much clear on when to use the questionnaire. These can be
mostly used in case of descriptive and explanatory type of research. Example:
Questionnaire on attitude, opinion, and organizational practices, enable the researcher
to identify and describe the variability.
Advantages of the Questionnaire

Cost involved is low in case of widely spread geographical sample.


It is more appreciable one because it more free from the subjectivity of the
interviews.
Respondents also may find adequate time to give well thought answers.
It is more advantages in the case when respondents are not reachable.

But the rate of return of questionnaire and the fulfillment at needed data for the study
may be doubtful. This can be used only when the respondents are educated and will
know to read the language in which questionnaire is prepared. Possibilities for
ambiguous replies, omission of replies are more. This method is more time consuming.
Before sending the final questionnaire to the respondents it is always more important to
conduct the Pilot study for resting the questionnaire. Pilot study is just a rehearsal of the
main survey such survey conducted with help of experts brings more strength to the
questionnaire.
Data collection through schedules
This method is very much like data collection through questionnaires. Schedules are a
proforma containing set of questions which are filled in by the enumerators who are
specially appointed for this purpose. In this method the enumerators are expected to
perform well and they must be knowledgeable and must possess the capacity of cross
examination in order to find the truth. These methods are usually expensive and are
conducted by bigger and Government organizations.
OBSERVATION
Observation is one of the cheaper and effective techniques of data collection. The
observation is understood as a systematic process of recording the behavioral patterns
of people, objects and occurrence as they are witnessed.
Using this observational method of data collection the following data related to
movements, work habits, the statements mad and meetings conducted,[by human
begins] facial expressions, body language and other emotions such as joy, anger and
sorrowfulness of the human beings can be collected.

Also other environmental factors includes layout, workflow patterns, physical


arrangements can also be noted. In this method of data collection the investigator
collects the data without interacting with the respondents. Example: Instead of asking
about the brands of shirts or the program they watch on just observing their pattern.
The observation would be classified based on the role of researcher as:
1. Non participant observation
The role of the investigator here would be an external agent who sits in a corner without
interacting with the samples which is to be observed.
2. Participant observation
Joins with the group and work along with them in way the work is done but many not be
asking questions related to the research / investigation.
It also can be classified based on the methods as:
1. Structured observation
In this case the researcher may have a predetermined set of categories of activities or
phenomena planned to be studied. Example: To observe the behavior pattern of
individual when he/she go for purchasing to be planned in such a way where the
frequency of purchasing, and the interest during the purchase and the way goods are
preferred / selected. Any researcher would be having a plan on the observation to be
made.
2. Unstructured observation
The researcher may not do the data collection based on the specific ideas. These sorts of
methods are used more qualitative research studies.
Example: To observe the behavior pattern of individual when he/she go for purchasing
to be planned in such a way where the frequency of purchasing, and the interest during
the purchase and the way goods are preferred / selected.
Merits of observational studies

The data collected in this method are generally reliable and they are free from
respondent's bias.
It is easier to observe the busy people rather meeting and collecting the data.

Demerits of observational studies

The observer has to be present in the situation where the data to be collected.
This method is very slow and expensive.

Since the observer is going to collect the data by consuming more time and
observing the sample there is a possibility for biased information.

Survey Method
Survey is a popular method of "fact-finding" study which involves collection of data
directly from the population at a particular time. Survey can be defined as a research
technique in which information may be gathered from a sample of people by use of
questionnaire or interview.
Survey is considered as field study which is conducted in a natural setting which collects
information directly from the respondents. Surveys are been conducted for many
purposes - population census, socio economic surveys, expenditure surveys and
marketing surveys. The purpose of these surveys would be providing information to the
government planners or business enterprises. The surveys are also conducted to explain
phenomena where in causal relationship between two variables to be assessed.
Surveys are been used to compare the demographic groups such as to compare the high
and low income groups, to compare the preference based on age. These surveys are been
conducted in the care of descriptive type of research and in which large samples are
focused.
The surveys are more appropriate in the social and behavioral science. Surveys are more
convened with formulating the hypothesis and testing the relationship between nonmanipulated variables. The survey research requires skillful workers to gather data.
The subjects for surveys may be classified as:
1. Social surveys which include

Demographic characteristics of a group of people


The social environment of people
People's opinion & attitudes

2. Economic surveys

Economic conditions of people


Operations of economic system

Important stages in survey methods


Selecting the universe of the field

Choosing samples from the universe


Deciding on tools used for data collection

Analyzing the data


Other methods of Data Collection
Warranty cards : A type of post card with few focused typed questions may be used by
the dealers / retailers to collect information from the customers. The dealers /
researchers may request customers to fill in the required data.
Distributor Audits : These sorts of data may be collected by distributors to estimate
the market size, market share and seasonal purchasing pattern. This information is
collected by observational methods. Example: Auditing the provisional stores and
collecting data on inventories recorded by copying information from store records.
Pantry Audits : This method is used to estimate the consumption of the basket of the
goods at the consumer level. Here the researcher collects the inventory of types,
quantities and prices of commodities consumed. Thus pantry audit data are recorded
from the consumption of consumer's pantry. An important limitation of pantry audit is
that, sometimes the audit data alone may not be sufficient to identify the consumer's
preferences.
Consumer Panels : An extension of Pantry Audit approach on regular basis is known
as Consumer Panels. A set of consumers are arranged to come to an understanding to
maintain a daily records of their consumption and the same is made available to
researcher on demand. In other words Consumer Panel is a sample of consumers
interviewed repeatedly over a period of time.
Field works : To collect the primary data any researcher or investigator may are
different methods wherein they go to door to door, use telephone to collect data.
SUMMARY
Since there are various methods of data collection, the researcher must select the
appropriate data. Hence, the following factors to be kept in mind by the researcher:
1. Nature, Scope and Object of the Enquiry
2. Availability of funds
3. Time factor
4. Precision required

KEY TERMS

Primary data
Secondary data
Interview
Questionnaire
Schedule
Observation
Survey
Warranty cards
Pantry Audits
Distributor Audits
Consumer Panels

QUESTIONS
1. Describe the different data sources, explaining their advantages and disadvantages.
2. What is bias and how it can be reduced during interviews?
3. "Every data collection method has it own built in biases. Therefore resorting to multi
methods of data collection is only going to compound the biases". How would you
evaluate the statement?
4. Discuss the role technology in data collection.
5. What is your view on using the warranty cards and Distributor audits in data
collection?
6. Differentiate the questionnaire and Interview schedules to decide the best one.
7. Discuss the main purposes for which Survey methods are used.

- End of Chapter LESSON 10


SAMPLING METHODS

OBJECTIVES

To define the terms Sampling, Sample, Population


To describe the various sampling designs

To discuss the importance of Confidence and Precision


To identify and use the appropriate sampling designs for different research
purposes

STRUCTURE

Universe/population
Sampling
Sampling techniques
Importance of sample design and sample size

INTRODUCTION
Sampling Methods is a process of selecting sufficient number of elements from the
population. This is also understood as the process of obtaining information about entire
population by examining only a part of it. So sample can be defined as a subset of the
population. In other words, some, but not all elements of the population would form the
sample.
Basically sampling helps a researcher in variety of ways as follows:

It saves time and money. A sample study usually is less expensive than a
population survey.
It also helps the researcher to obtain the accurate results.
Sampling is only way when the population is very large in size.
It enables to estimate the sampling error so; this assists in obtaining information
and in convening the characteristics of population.

To understand the sampling process, the researcher also should understand the
following terms:
1. Universe / Population
In any of the research, the interest of the researcher is mainly in studying the various
characteristics relating to items or individuals belonging to a particular group. This
group of individuals under study is known as the population or universe.
2. Sampling
A finite subset-selected from a population with the objective of investigating its
properties is called a "sample".
The number of units in the sample is known as the 'sample size'. This is the important
role of any research which enables to draw conclusions about characteristics of the
population.
3. Parameter & Statistics

The statistical constants used for further analysis of Data collected such as Mean (),
Variance (2), Skewness (1), Kurtosis (2), Correlation () can be computed for the
sample drawn from the population.
Sampling is the important part of any research before data collection. So the sampling
process should be done in a careful way to obtain the exact samples and sample size of
the population on which the research is to be done.
Example: A researcher who would like to study the customer satisfaction for a health
drink namely Horlicks should identify the population who are consuming Horlicks. If
the consumers are varying in age, genders all over the state or country, he should be able
to decide to particular consumers are going to be focused. Again, if the number is more
to survey he has to decide on how many individuals he targets for his study.
Hence the effecting sampling process should have the following steps:
Define the population
(Elements, units, extent, and time)

Specify the sample frame


(The mean of representing the elements of population map, city directory)

Specifying the sampling unit


(Sampling unit containing more population elements)

Specifying the sampling method


(The method by which sampling units are to be selected)

Determine the sample size


(The no. of elements of the population to decided)

Sampling plan
(Procedure for selected sampling unit)

Select the sample


(The effective and field work revision for section of samples)
Hence a sample design is a definite plan for obtaining a sample from a given population.
So whenever samples have to be decided for the study, the following can be considered:
- Outline the universe
- Define a sampling unit
- Sampling frame
- Size of the sample
Sampling Techniques
Sampling techniques can be divided into two types:
1. Probability or representative sampling
a. Simple random sampling
b. Stratified random sampling
c. Systematic sampling
d. Cluster sampling
e. Multistage sampling
2. Non probability or judgmental sampling
a. Quota sampling
b. Purposive sampling
Other methods

Snow ball sampling

Spatial sampling
Saturation sampling

1. Probability Sampling
This is a scientific technique of drawing samples form the population according to some
laws of change according to which each unit in the universe has some definite preassigned probability of being selected in the sample.
Simple Random Sampling
In this technique, sample is drawn in such a way that every elements or unit in the
population has an equal and independent chance of being included in the sample.
The unit selected in any draw from the population is not preplanned in population
before making the next draw is known as simple random sampling without replacement.
If the unit is replaced back before making the next draw the sampling plan is called as
simple random sampling with replacement.
Stratified Random Sampling
When the population is heterogeneous with respect to the variable or characteristics
under the study this sampling method is used. Stratification means division into
homogenous layers or groups. Stratified random sampling involves stratifying the given
population into a number of sub-groups or sub-population known as strata.
The characteristics of stratified samples are as follows:

The units within each stratum are as homogenous as possible.


The differences between various strata are as marked as possible.
Each and every unit in the population belongs to one and only one stratum.

The population can be stratified according to geographical, sociological or economic


characteristics. Some of the commonly used stratifying factors are age, sex, income,
occupation, education level, geographic area, economic status etc. To decide on the no.
of samples or items drawn from the different strata, will be wept proportional to the
sizes of the strata.
Example: If pi represents the proportion of the population included in stratum 'i', and
n represents the total sample size, then the number of elements selected from stratum i
is (n - pi)
Example: Suppose we need a sample size of n = 30 to be drawn from a population of
size N= 6000 which is divided into three strata of sizes N1 = 3000, N2 = 1800, N3 =
1200.

Total population = 3000+1800+1200 = 6000

Hence using the proportional allocation, the sample size for different strata are 15, 9,
and 6, which are proportionate to the strata sizes of 3000, 1800, and 1200.
Systematic Sampling
This sampling is a slight variation of simple random sampling in which only the first
sample unit is selected at random while remaining units are selected automatically in a
definite sequence at equal spacing from one another. This kind of sampling is
recommended only when if a complete and up to date list of sampling units is available
and the units are arranged in a systematic order as alphabetical, chronological,
geographical etc.
Systematic sampling can be taken as an improvement over a simple random sampling
since it spreads more evenly over the entire population. This method is one of the easier
and less costly methods of sampling and can be conveniently used in case of large
population.
Cluster Sampling
If the total area of interest happens to be a big one, a convenient way in which a sample
can be taken is to divide the area into a number of smaller non-overlapping areas and
then to randomly select a number of these smaller areas.
In cluster sampling, the total population is divided into a number of relatively small sub
divisions which are themselves clusters of still smaller units and some of these clusters
are randomly selected for inclusion in overall sample. Cluster sampling reduces the cost

by concentrating surveys in clusters. But this type of sampling is less accurate than
random sampling.
Multistage Sampling
It is a further development of the principle of cluster sampling. In case of investigating
the working efficiency of nationalized banks in India, a sample of few banks may be
taken for the purpose of investigation.
Here to select the banks as a first step, the main states in a country are selected from the
states from which the banks to be included for the study will be selected. This represents
the two stage sampling. Even further, from the district certain towns may be selected,
from where the banks will be selected. This may represent three stages sampling.
Even thereafter, instead of taking census from all the banks in all the towns we have
selected, once again banks may be selected randomly for the survey. Hence the random
selection at all levels (various levels) is known as multistage random sampling design.
Sequential Sampling
This is one of the complex sampling designs. The ultimate size of sample in this
technique is not fixed in advance but it is determined according to the mathematical
decision rules on the basis of information yielded in the survey.
This method is adopted when sampling plan is accepted in context of Statistical Quality
Control.
Example: When a lot is to be accepted or rejected on the basis of single sample, it is
known as single sampling; when the decision is to be taken on the basis of two
samples it is known as double sampling, and in case the decision is based on the
more than two samples but the number of samples is certain and decided in advance,
the sampling is known as multi sampling. In case when the number of samples is
more than two but it is neither certain nor decided in advance, this type of system is
often referred to as Sequential Sampling. So in case of Sequential Sampling, one can
go on taking samples one after another as long as one desires to do so.
2. Non-probability Sampling
Quota Sampling
This is stratified-cum-purposive or judgment sampling and thus enjoys the benefits of
both. It aims at making the best use of stratification without incurring the high costs
involved in probabilistic methods. There is considerable saving in time and money as
the simple units may be selected that they are close together. If carefully experienced by
skilled and experienced investigators who are aware of the limitations of judgment
sampling and if proper controls are imposed on the investigators, this sampling method
may give reliable results.

Purposive or Judgment Sampling


A desired number of sampling units are selected deliberately so that only important
items representing the true characteristics of the population are included in the sample.
A major disadvantage of this sampling is that it is highly subjective, since the selection
of the sample depends entirely on the personal convenience and beliefs.
Example: In the case of socio-economic survey on the standard of living from people in
Chennai, if the researcher wants to show that the standard has gone down, he may
include only individuals from the low income stratum of the society in the samples and
exclude people from rich areas.
Other Forms of Sampling
Snow Ball Sampling : This method is used in the cases where information about units
in the population is not available. If a researcher wants to study the problem of the
weavers in a particular region, he may contact the weavers who are known to him. From
them, he may collect the addresses of other weavers in the various parts of the region he
selected. From them again he may collect the information on other known weavers to
him. By repeating like this for several times, he will be able to identify and contact the
majority of weavers from a selected region. He could then draw a sample from this
group. This method is useful only when individuals in the target group have contact with
one another, and also willing to reveal the names of others in the group.
Spatial Sampling : Some populations are not static and moving from place to place
but staying at one place when an event is taking place. In such case the whole population
in a particular place is taken into the sampling and studied.
Example: The number of people living in Dubai may vary depending on many factors.
Saturation Sampling : Sometimes if all members of population is need to be studied
so as to get a picture of entire population. The sampling method that requires a study of
entire population is called Saturation Sampling. This technique is more familiar in Socio
metric studies where in distorted results will be produced even if one person is left out.
Example: In case of analyzing the student's behavior of one particular class room, all
the students in the class room must be examined.
From the above discussion on sampling methods, normally one may resort to simple
random sampling since biasness is generally eliminated in this type of sampling. At the
same time, purposive sampling is considered more appropriate when the universe
happens to be small and a known characteristic of it is to be studied intensively. In
situations where random sampling is not possible then it is advisable to use necessarily
a sampling design other than random sampling.
Determination of Sample size

Determination of appropriate sample size is crucial part of any business research. The
decision on proper sample size tremendously requires the use of statistical theory. When
a business research report is been evaluated, the evaluation start with the question of
'how big is the sample size?'
Having discussed various sampling designs it is important to focus the attention on
Sample Size. Suppose we select a sample size of 30 from the population of 3000 through
a simple random sampling procedure, will we able to generalize the findings to the
population with confidence? So in this case what is the sample size that would be
required to carry out the research?
It is the known fact that larger the sample size, the more accurate the research is. In fact
this is the fact based on the statistics. According to this fact, increasing the sample size
decreases the width of the confidence interval at a given confidence level. When the
standard deviation of the population is unknown, a confidence interval is calculated by
using the formula:
Confidence Interval,
= X KSx
where,
Sx = S / Sqrt(n)
In sum, choosing the appropriate sampling plan is one of the important research design
decisions the researcher has to make. The choice of a specific design will depend broadly
on the goal of research, the characteristics of the population, and considerations of cost.
Issues of Precision and Confidence in determining Sample size
We now need to focus attention on the second aspect of the sampling design issuethe
sample size. Suppose we select 30 people from a population of 3,000 through a simple
random sampling procedure. Will we be able to generalize our findings to the
population with confidence? What is the sample size that would be required to make
reasonably precise generalizations with confidence? What do precision and confidence
mean?
A reliable and valid sample should enable us to generalize the findings from the sample
to the population under investigation. No sample statistic (X, for instance) is going to be
exactly the same as the population parameter (Sx), no matter how sophisticated the
probability sampling design is. Remember that the very reason for a probability design
is to increase the probability that the sample statistics will be as close as possible to the
population parameters.
Precision

Precision refers to how close our estimate is to the true population characteristic.
Usually, we would estimate the population parameter to fall within a range, based on the
sample estimate.
Example: From a study of a simple random sample of 50 of the total 300 employees in
a workshop, we find that the average daily production rate per person is 50 pieces of a
particular product (X = 50). We might then (by doing certain calculations, as we shall
see later, be able to say that the true average daily production of the product (X) would
lie anywhere between 40 and 60 for the population of employees in the workshop. In
saying this, we offer an interval estimate, within which we expect the true population
mean production to be ( = 50 10). The narrower this interval, the greater is the
precision. For instance, if we are able to estimate that the population mean would fall
anywhere between 45 and 55 pieces of production ( = 50 5) rather than 40 and 60 (
= 50 10), then we would have more precision. That is, we would now estimate the
mean to lie within a narrower range, which in turn means that we estimate with greater
exactitude or precision.
Precision is a function of the range of variability in the sampling distribution of the
sample mean. That is, if we take a number of different samples from a population, and
take the mean of each of these, we will usually find that they are all different, are
normally distributed, and have a dispersion associated with them Even if we take only
one sample of 30 subjects from the population, we will still be able to estimate the
variability of the sampling distribution of the sample mean. This variability is called the
standard error, denoted by 'S'. The standard error is calculated by the following formula:
Sx = S / Srqt(n)
where,
S = Standard deviation of the sample
n = Sample size
Sx = Standard error or the extent of precision offered by the sample.
In sum, the closer we want our sample results to reflect the population characteristics,
the greater will be the precision we would aim at. The greater the precision required, the
larger is the sample size needed, especially when the variability in the population itself is
large.
Confidence
Whereas precision denotes how close we estimate the population parameter based on
the sample statistic, confidence denotes how certain we are that our estimates will really
hold true for the population. In the previous example of production rate, we know we
are more .precise when we estimate the true mean production () to fall somewhere
between 45 and 55 pieces, than somewhere between 40 and 60.

In essence confidence reflects the level of certainty with which we can state that our
estimates of the population parameters, based on our sample statistics will hold true.
The level of confidence can range from 0 to 100%. A 95% confidence is the
conventionally accepted level for most business research, most commonly expressed by
denoting the significance level as p = .05. In other words, we say that at least 95 times
out of 100, our estimate will reflect the true population characteristic.
In sum, the sample size n, is a function of
1. The variability in the population
2. Precision or accuracy needed
3. Confidence level desired
4. Type of sampling plan used, for example, sample random sampling versus stratified
random sampling
It thus becomes necessary for researchers to consider at least four points while making
decisions on the sample size needed to do the research:
(1) Much precision is really needed in estimating the population characteristics interest,
that is, what is the margin of allowable error?
(2) How much confidence is really needed, i.e. how much chance can we take of making
errors in estimating the population parameters?
(3) To what extent is there variability in the population on the characteristics
investigated?
(4) What is the cost-benefit analysis of increasing the sample size?
Determining the Sample Size
Now that we are aware of the fact that the sample size is governed by the extent of
precision and confidence desired, how do we determine the sample, retired for our
research? The procedure can be illustrated through an example:
Suppose a manager wants to be 95% confident that the withdrawals in a bank will be
within a confidence interval of $500. Example of a simple of clients indicates that the
average withdrawals made by them have a standard deviation of $3,500. What would be
the sample size needed in this case?
We noted earlier that the population mean can be estimated by using the formula:
= X KSx

Given, S = 3500. Since the confidence level needed here is 95%, the applicable K value is
1.96 (t-table). The interval estimate of $500 will have to encompass a dispersion of
(1.96 x standard error). That is,

The sample size needed in the above is 188. Let us say that this bank has the total
clientele of only 185. This means we cannot sample 188 clients. We can, in this case,
apply the correction formula and see what sample size would be needed to have the
same level of precision and confidence given the fact that we have a total of only 185
clients. The correction formula is as follows:

where,
N = total number of elements in the population = 185
n = sample size to be estimated = ?
Sx = Standard error of estimate of the mean = 255.10
S = Standard deviation of the sample mean = 3500
Applying the correlation formula,
we find that
255.10 = 3500 n 185-n/184
the value of n to be 94.
We would now sample 94 of the total 185 clients.

To understand the impact of precision and/or confidence on the sample size, let us try
changing the confidence level required in the bank withdrawal exercise which needed a
sample size of 188 for a confidence level of 95%. Let us say that the bank manager now
wants to be 99% sure that the expected monthly withdrawals will be within the interval
of $500. What will be the sample size now needed?

The sample has now to be increased 1.73 times (from 188 to 325) to increase the
confidence level from 95% to 99%.It is hence a good idea to think through how much
precision and confidence one really needs, before determining the sample size for the
research project.
So far we have discussed sample size in the context of precision and confidence with
respect to one variable only. However, in research, the theoretical framework has
several variables of interest, and the question arises how one should come up with a
sample size when all the factors are taken into account.
Krejcie and Morgan (1970) greatly simplified size decision by providing a table that
ensures a good decision model. The Table provides that generalized scientific guideline
for sample size decisions. The interested student is advised to read Krejcie and Morgan
(1970) as well as Cohen (1969) for decisions on sample size.

Importance of Sampling Design and Sample Size


It is now possible to see how both sampling design and the sample size are important to
establish the representativeness of the sample for generality. If the appropriate
sampling design is not used, a large sample size will not, in itself, allow the findings to
be generalized to the population. Likewise, unless the sample size is adequate for the
desired level of precision and confidence, no sampling design, however sophisticated,
can be useful to the researcher in meeting the objectives of the study.
Hence, sampling decisions should consider both the sampling design and the sample
size. Too large a sample size, however (say, over 500) could also become a problem in as

much as we would be prone to committing Type II errors. Hence, neither too large nor
too small sample sizes help research projects.
Roscoe (1975) proposes the following rules of thumb for determining
sample size:
1. Sample sizes larger than 30 and less than 500 are appropriate for most research.
2. Where samples are to be broken into sub samples; (male/females, juniors/ seniors,
etc.), a minimum sample size of 30 for each category is necessary.
3. In multivariate research (including multiple regression analyses), the sample size
should be several times (preferably 10 times or more) as large as the number of
variables in the study.
4. For simple experimental research with tight experimental controls (matched pairs,
etc.), successful research is possible with samples as small as 10 to 20 in size.
KEY TERMS
1. "What is Sample Design"? What all are the points to be considered to develop a
sample design?
2. Explain the various sampling methods under probability sampling.
3. Discuss the non probability sampling methods.
4. What are the importance of sample size and sampling design?
5. Discuss the other sampling methods.
6. Explain why cluster sampling is a probability sampling design.
7. What are the advantages and disadvantages of cluster sampling?
8. Explain what precision and confidence are and how they influence sample size.
9. The use of a convenience sample used in organizational research is correct because all
members share the same organizational stimuli and go through almost the same kinds
of experience in their organizational life. Comment.
10. Use of a sample of 5,000 is not necessarily better than one of 500. How would you
react to this, statement?
11. Non-probability sampling designs ought to be preferred to probability sampling
designs in some cases. Explain with an example

- End of Chapter LESSON 11


THE NATURE OF FIELD WORK

OBJECTIVES

To recognize that field work can be performed by many different parties.


To understand the importance of training for new interviewers
To understand the principle tactics of asking questions

STRUCTURE

Interviewing
Training for interviewing
The major principle for asking questions
Probing
Recording the response

INTRODUCTION
A personal interviewer administering a questionnaire door to door, a telephone
interviewer calling from a central location, an observer counting pedestrians in a
shopping mall, and others involved in the collection of data and the supervision of that
process are all Field Workers in the field. The activities of the field workers may vary in
nature. This lesson would help to understand the interview methods in data collection
process of the research and field work management.
Who conducts the field work?
Data collection is rarely carried out by the person who designs the research project.
However, the data collecting stage is crucial, because the research project is no better
than the data collected in the field. Therefore, it is important that the research
administrator select capable people who may be entrusted to collect the data. An irony
of business research is that highly educated and trained individuals design the research,
but the people who collect the data typically have little research training or experience.
Knowing the vital importance of data collected in the field, research administrators
must concentrate on carefully selecting field workers.
INTERVIEWING

Interviewing process is establishing rapport with the, respondent Interviewer bias may
enter in if the field worker's clothing or physical appearance is unattractive or unusual.
Suppose that a male interviewer, wearing a dirty T-shirt, interviews subjects in an
upper-income neighborhood. Respondents may consider the interviewer slovenly and
be less cooperative than they would be with a person dressed more appropriately.
Interviewers and other fieldworkers are generally paid an hourly rate or a per-interview
fee. Often interviewers are part-time workers, housewives, graduate students, secondary
school teachers from diverse backgrounds. Primary and secondary school teachers are
an excellent source for temporary interviewers during the summer, especially when they
conduct interviews outside the school districts where they teach. Teachers' educational
backgrounds and experiences with the public make them excellent candidates for
fieldwork.
TRAINING FOR INTERVIEWERS
The objective of training is to ensure that the data collection instrument is administered
uniformly by all field investigators. The goal of training sessions is to ensure that each
respondent is provided with common information. If the data are collected in a uniform
manner from all respondents, the training session will have been a success. After
personnel are recruited and selected, they must be trained.
Example: A woman who has just sent her youngest child off to first grade is hired by an
interviewing firm. She has decided to become a working mother by becoming a
professional interviewer. The training that she will receive after being selected by a
company may vary from virtually no training to a 3-day program if she is selected by one
of the larger survey research agencies. Almost always there will be a briefing session on
the particular project. Typically, the recruits will record answers on a practice
questionnaire during a simulated training interview.
More extensive training programs are likely to cover the following topics:
1. How to make initial contact with the respondent and secure the interview
2. How to ask survey questions
3. How to probe
4. How to record responses
5. How to terminate the interview
Making Initial Contact and Securing the Interview
Interviewers are trained to make appropriate opening remarks that will convince the
person that his or her cooperation is important.

Example: "Good afternoon, my name is _____ and I'm from a national survey
research company. We are conducting a survey concerning. I would like to get a few of
your ideas".
Much fieldwork is conducted by research suppliers who specialize in data collection.
When a second party is employed, the job of the study designed by the parent firm is not
only to hire a research supplier but also to establish supervisory controls over the field
service.
In some cases a third party is employed. For example, a firm may contact a survey
research firm, which in turn subcontracts the fieldwork to a field se vice. Under these
circumstances it is still desirable to know the problems that might occur in the field and
the managerial practices that can minimize them.
Asking the Questions
The purpose of the interview is, of course, to have the interviewer ask questions and
record the respondent's answers. Training in the art of stating questions can be
extremely beneficial, because interviewer bias can be a source of considerable error in
survey research.
There are five major principles for asking questions:
i. Ask the questions exactly as they are worded in the questionnaire.
ii. Read each question very slowly.
iii. Ask the questions in the order in which they are presented in the questionnaire.
iv. Ask every question specified in the questionnaire.
v. Repeat questions those are misunderstood or misinterpreted.
Although interviewers are generally trained in these procedures, when working in the
field many interviewers do not follow them exactly. Inexperienced interviewers may not
understand the importance of strict adherence to the instructions. Even professional
interviewers take shortcuts when the task becomes monotonous. Interviewers may
shorten questions or rephrase unconsciously when they rely on their memory of the
question rather than reading the question as it is worded. Even the slightest change in
wording can distort the meaning of the question and cause some bias to enter into a
study. By reading the question, the interviewer may be reminded to concentrate on
avoiding slight variations in tone of voice on particular words phases in the question.
PROBING
General training of interviewers should include instructions on how to probe when
respondents give no answer, incomplete answers, or answers that require clarification.

Probing may be needed for two types of situations. First, it is necessary when the
respondent must be motivated to enlarge on, clarify or explain his or her answer. It is
the interviewer's job to probe for complete, unambiguous answers. The interviewer must
encourage the respondent to clarify or expand on answers by providing a stimulus that
will not suggest the interviewer's own ideas or attitudes. The ability to probe with
neutral stimuli is the mark of an experienced interviewer. Second, probing may be
necessary in situations in which the respondent begins to ramble or lose track of the
question. In such cases the respondent must be led to focus on the specific content of
the interview and to avoid irrelevant and unnecessary information.
The interviewer has several possible probing tactics to choose from,
depending on the situation:
i. Repetition of the question: The respondent who remains completely silent may
not have understood the question or may not have decided how to answer it. Mere
repetition may encourage the respondent to answer in such cases. For example, if the
question is "What is there that you do not like about your supervisor?" and the
respondent does not answer, the interviewer may probe: "Just to check; is there
anything you do not like about your supervisor?"
ii. An expectant pause: If the interviewer believes the respondent has more to say,
the "silent probe" accompanied by an expectant look, may motivate the respondent to
gather his or her thoughts and give a complete response Of course, the interviewer must
be sensitive to the respondent so that the silent probe does not become an embarrassed
silence.
iii. Repetition of the Respondent's Reply: Sometimes the interviewer may repeat
the verbatim of the respondent. This may help the respondent to expand the answer.
RECORDING THE RESPONSES
The analyst who fails to instruct fieldworkers in the techniques of recording answers for
one study rarely forgets to do so in the second study. Although the concept of recording
an answer seems extremely simple, mistakes can be made in the recording phase of the
research. All fieldworkers should use the same mechanics of recording.
Example: It may appear insignificant to the interviewer whether she uses a pen or
pencil, but to the editor who must erase and rewrite illegible words, using a pencil is
extremely important.
The rules for recording responses to closed questionnaires vary with the specific
questionnaire. The general rule, however, is to place a check in the box that correctly
reflects the respondent's answer. All too often interviewers don't bother recording the
answer to a filter question because they believe that the subsequent answer will make
the answer to the filter question obvious. However, editors and coders do not know how
the respondent actually answered a question.

The general instruction for recording answers to open-ended-response questions is to


record the answer verbatim, a task that is difficult for most people. Inexperienced
interviewers should be given the opportunity to practice
The Interviewer's Manual of the Survey Research Center provides the instructions on
the recording of interviews. Some of its suggestions:

Recording answers to open-ended-response questions follow


Record the responses during the interview
Use the respondent's own words
Do not summarize or paraphrase the respondent's answer
Include everything that pertains to the question objectives
Include all of your probes

The basics of effective Interviewing


Interviewing is a skilled occupation; not everyone can do it, and even few can do it
extremely well. A good interviewer observes the following principles:
1. Have integrity and be honest.
2. Have patience and tact.
3. Pay attention to accuracy and detail.
4. Exhibit the real enquiry at hand, but keep your own opinions to yourself.
5. Be a good listener.
6. Keep the inquiry and respondent's responses confidential. Respect other's rights
Terminating the Interview
The final aspect of training deals with instructing the interviewers on how to close the
interview. Fieldworkers should not close the interview before pertinent information has
been secured. The interviewer whose departure hasty will not be able to record those
spontaneous comments responds sometimes offer after all formal questions have been
asked. Avoiding hasty departures is also a matter of courtesy.
Fieldworkers should also answer to the best of their ability any quest the respondent,
concerning the nature and purpose of the study. Beat the fieldworker may be required to
re-interview the respondent at some his time, he or she should leave the respondent
with a positive feeling about having cooperated in a worthwhile undertaking. It is
extremely important thank the respondent for his or her cooperation.
FIELDWORK MANAGEMENT

Managers of the field operation select, train, supervise, and control fieldworkers. Our
discussion of fieldwork principles mentioned selection and training. This section
investigates the tasks of the fieldwork manager in greater detail.
Briefing Session for Experienced Interviewers
After interviewers have been trained in fundamentals, and even when they have become
experienced, it is always necessary to inform workers about the individual project. Both
experienced and inexperienced fieldworkers must be instructed on the background of
the sponsoring organization, sampling techniques, asking questions, callback
procedures, and other matters specific to the project.
If there are special instructions - for example, about using show cards or video
equipment or about restricted interviewing times-these should also be covered during
the briefing session. Instructions for handling certain key questions are always
important. For example, the following fieldworker instructions appeared in a survey of
institutional investors who make buy-and sell decisions about stocks for banks, pension
funds, and the like.
A briefing session for experienced interviewers might go like, All interviewers report to
the central office, where the background of the firm and the general aims of the study
are briefly explained. Interviewers are not provided with too much information about
the purpose of the study, thus ensuring that they will not transmit any preconceived
notions to respondents. For example, in a survey about the banks in a community, the
interviewers would be told that the research is a banking study, but not the name of the
sponsoring bank. To train the interviewers about the questionnaire, a field supervisor
conducts an interview with another field supervisor who acts as a respondent. The
trainees observe the interviewing process, after which they each interview and record
the responses of another field supervisor. Additional instructions are given to the
trainees after the practice interview.
Training to Avoid Procedural Errors in Sample Selection
The briefing session also covers the sampling procedure. A number of research projects
allow the interviewer to be at least partially responsible for selection of the sample.
When the fieldworker has some discretion in the selection of respondents, the potential
for selection bias exists. This is obvious in the case of quota sampling, but less obvious
in other cases.
Example: In probability sampling where every nth house is selected, the fieldworker
uses his or her discretion in identifying housing units. Avoiding selection error may not
be as simple as it sounds.
Example: In an older, exclusive neighborhood, a mansion's coach house or servant's
quarters may have been converted into an apartment that should be identified as a
housing unit. This type of dwelling and other unusual housing units (apartments with
alley entrances only, lake cottages, rooming houses) may be overlooked, giving rise to

selection error. Errors may also occur in the selection of random digit dialing samples.
Considerable effort should be expended in training and supervisory control to minimize
these errors.
The activities involved in collecting data in the field may be performed by the
organization needing information, by research suppliers, or by third party field service
organizations. Proper execution of fieldwork is essential for producing research results
without substantial error.
Proper control of fieldwork begins with interviewer selection. Fieldworkers should
generally be healthy, outgoing, and well groomed. New fieldworkers must be trained in
opening the interview, asking the questions, probing for additional information,
recording the responses, and terminating the interview. Experienced fieldworkers are
briefed for each new project so that they are familiar with its specific requirements. A
particular concern of the briefing session is reminding fieldworkers to adhere closely to
the prescribed sampling procedures.
Careful supervision of fieldworkers is also necessary. Supervisors gather and edit
questionnaires each day. They check to see that field procedures are properly followed
and that interviews are on schedule. They also check to be sure that the proper sampling
units are used and that the proper people are responding in the study. Finally,
supervisors check for interviewer cheating and verify a portion of the interviews by reinterviewing a certain percentage of each fieldworker's respondents.
SUMMARY
This paper outlined the importance of training for new interviewers. In this chapter five
major principles for asking questions have been dealt in detail.
KEY TERMS

Field worker
Probing
Field interviewing
Briefing session
Training Interview
Reinterviewing questions

QUESTIONS
1. What qualities should a field worker possess?
2. What is the proper method of asking questions?
3. When should an interviewer probe? Give examples of how probing should be done?
4. How should an Interviewer terminate the interview?

5. What are the qualities of the interviewer that make him more effective?
REFERENCES
1. Ramanuj Majumdar, Marketing Research, Wiley Eastern Limited New Delhi (1991)
2. Cochran, W.G., Sampling Techniques, 2nd ed. New York: John Wiley and Sons.
3. Chaturvedi, J.C., Mathematical Statistics, Agra: Nok Jhonk Karyalaya, 1953.

- End of Chapter LESSON 12


SOURCES OF DATA

Sources of data - primary - secondary data - Questionnaire design: attitude


measurement techniques - motivational research techniques - selection appropriate
statistical techniques - correlation - research.
OBJECTIVES

To explain the difference between secondary and primary data


To discuss the advantages and disadvantages of secondary data
To learn the nature of secondary data
To understand the evaluation of secondary data sources
To learn the sources of secondary data

STRUCTURE

Value of secondary data


Disadvantage of secondary data
Nature and scope of secondary data
Sources of secondary data

INTRODUCTION
The availability of data source is very much needed to solve the problem and there are
many ways by which the data is collected. The task of data collection begins after a
research problem has been defined and research designed is prepared. Thus the data to
be collected can be classified as being either secondary or primary. The determination of
data source is based on three fundamental dimensions as given below:

1. The extent of data already exist in some type,


2. The degree to which the data has been interpreted by someone, and
3. The extent to which the researcher or decision maker understands the reasons and
why the data was collected and researched.
Primary Data is the data gathered and assembled specifically for the project at hand. It
is "finished" raw data, and has yet to receive any type of meaningful interpretation. The
primary data is fresh, and since it is collected for the first time, it happens to be original
in character. On the other hand, Secondary Data is that which has already been collected
by someone else and also passed through statistical process and interpretation.
Secondary data is historical data structure of variables previously collected and
assembled for some research problem other than the current situation.
The sources of primary data tend to be the output of conducting some type of
exploration, descriptive or casual research that employs surveys, experiments and / or
observation as technique of collecting the needed data. The greater insights underlying
primary data will be discussed in the chapter "methods of data collection". The pros &
cons of primary data also discussed with reference to various techniques involved in the
process.
The source of secondary data can be found inside a company at public libraries, and
universities, on World Wide Web (www) sites or purchased from a firm specializing in
providing secondary information. Here, evaluation and source of data are discussed.
THE VALUE / ADVANTAGES OF SECONDARY DATA
More and more companies are interested in using the existing data as a major tool in the
management decisions. As more and more such data become available, many companies
are realizing that they can be used to make sound decisions. Data of this nature are
more readily available, often more highly valid and usually less expensive to secure than
primary data.
"Nowhere in science do we start from scratch" - this quote explains the value of
secondary data. Researchers are able to build on the past research - a body of business
knowledge. The researchers use other's experience and data when it is available as
secondary data. The primary advantage of secondary data is that obtaining data is
almost always less expensive and in addition the data can usually be obtained rapidly.
The major advantage and disadvantages are discussed below.
Advantages of Secondary Data are:
1. It is more economical as the cost of collecting original data is saved. In the collection
of primary data, a good deal of effort is required which includes preparation of data
collection forms, designing and printing of forms, persons appointed to collect data in

turn involves travel plan need to verify and finally data to be tabulated. All these need
large funds which can be utilized elsewhere if secondary data can serve the purpose.
2. Another advantage is that the use of secondary data saves much of the time of the
researcher. This also leads to prompt completion of research project.
3. Secondary data are helpful not only because it is useful but the familiarity with the
data indicates deficiencies and gaps. As a result, the researcher can make the primary
data collection more specific and more relevant to the study.
4. It also helps in gaining new insights to the problem, then can be used to fine tune the
research hypothesis and objectives.
5. Finally, secondary data can be used as a basis of comparison with primary data that
has been collected for this study.
DISADVANTAGES OF SECONDARY DATA
An inherent disadvantage of secondary data is that it is not designed specifically to meet
the researcher's needs. Secondary data quickly becomes outdated in our rapidly
changing environments. Since the purpose of the most of the studies is to predict the
future, the secondary data must be timely. Hence the most common problems with
secondary data are:
1. Outdated information
2. Variation in definition of terms or classifications. The unit of measurement may cause
problems if they are not identical to the researcher's needs. Even though original units
were comparable, the aggregated or adjusted units of measurements are not suitable for
the present study. When the data are reported in a format that does not exactly meet the
researchers needs, the data conversion may be necessary.
3. Another disadvantage of secondary data is that the user has to control over their
accuracy even though it is timely & pertinent, they may be inaccurate.
THE NATURE AND SCOPE OF SECONDARY DATA
Focusing on the particular business or management problem, the researcher needs to
determine whether useful information already exists, of exists how relevant the
information. Since existing information are more widespread than one might expect.
The secondary data exists in three forms:
1. Internal secondary data: The data collected by the individual company for some
purpose and reported periodically. This is also called as primary sources. The primary
sources are original work of research or raw data without interpretation or
pronouncements that represent an official opinion or position - Memos, complete
interviews, speeches, laws, regulations, court decisions, standards, and most

government data, including census, economic and labor data. Primary sources are
always the most authoritative because the information is not filtered. It also includes
inventory records, personnel records, process charts and similar data.
2. External secondary data: It consists of data collected by outside agencies such as
government, trade associations of periodicals. This is called as secondary sources.
Encyclopedia, text books, handbooks, magazine and newspaper articles and most news
crafts are considered to secondary sources. Indeed all reference materials fall into this
category.
3. Computerized data sources: This includes internal and external data usually
collected by specific companies with online information sources. This can be called as
territory sources. These are represented by indexes, bibliography and other finding aids,
e.g., internet search engines.
Evaluation of secondary data sources
The emphasis on secondary data will increase if an attempt is made to establish a set of
procedures to evaluate the secondary data regarding the quality of information obtained
via secondary data sources. Specifically, if secondary data are to be used to assist in the
decision process, then they should be assessed according to the following principles.
1. Purpose: Since most secondary data are collected for purpose other than the one at
hand, the data must be carefully evaluated on how they relate to the current research
objectives. Many times the original collection of data is not consistent with the
particular research study. These inconsistencies usually result from the methods and
units of measurement.
2. Accuracy: When observing secondary data researchers need to keep in mind what
was actually measured. For example, if the actual purchases in a test market were
measured, did they measure the first-time trial purchases or repeat purchases?
Researchers must also asses the generality of the data.
3. Questions like i) was the data collected from certain groups only or randomly? ii)
was the measure developed properly? iii) was the data presented as the total of
responses from all respondents or were they categorized by age, sex or socio economic
status?
4. In addition to the above dimensions, researchers must assess when the data were
collected. This factor not only damages the accuracy of the data but also may be useless
for interpretation. Researchers also must keep in mind that the flaws in the research
design and methods will alter the current research in process.
5. Consistency: When evaluating any source of secondary data, a good strategy is to
seek out multiple sources of the same data to assure consistency. For example, when
evaluating the economic characteristics of a foreign market, a researcher may try to

gather the same information from government sources, private business publications
and specially import or export trade publications.
6. Credibility: Researcher should always question the credibility of the secondary data
source. Technical competence, service quality, reputation, and tracing and expertise of
personnel representing the organization are some measures of credibility.
7. Methodology: The quality of secondary data is only as good as the methodology
employed to gather them. Flaws in methodological procedures could produce results
that are invalid, unreliable or not generalizable beyond the study itself. Therefore,
researchers must evaluate the size and description of the sample, the response date, the
questionnaire, and the overall procedure for collecting the data (telephone, mail, or
personal interview).
8. Bias: Researchers must try to determine the underlying motivation or hidden
agenda, if any, behind the secondary data. It is not uncommon to find many secondary
data sources published to advance the interest of commercial, political or other intersect
groups. Researchers should try to determine if the organization reporting the report is
motivated by certain purpose.

SOURCES OF SECONDARY DATA

A. Internal sources
Generally, internal data consists of sales or cost information. Data of this kind is found
in internal accounting or financial records. The two most useful sources of information
are sales invoices and accounts receivable reports; quarterly sales reports and sales
activity reports are also useful.

The major sources of internal secondary data are given below:


1. Sales invoices
2. Accounts receivable reports
3. Quality sales reports
4. Sales activity reports
5. Other types
a. customer letters
b. Customer comment cards
c. Mail order forms
d. Credit applications
e. Cash register receipts
f. Sales person expense reports
g. Employee exit interviews
h. Warranty cards
i. Post marketing research studies.
B. External Sources
When undertaking the search for secondary data researchers must remember that the
numbers of resources are extremely large. The researcher needs to connect the sources
by common theme.
The key variables most often sought by the researchers are given below:
1. Demographic dimensions
2. Employment characteristics
3. Economic characteristics
4. Competitive characteristics

5. Supply characteristics
6. Regulations characteristics
7. International market characteristics
The external secondary data do not originate in the firm and are obtained from outside
sources. It may be noted that secondary data can be collected from the originating
sources or from secondary sources. For example, the office of economic advisor, GOI is
the originating source for wholesale prices. In contrast a publication such as RBI
bulletin on wholesale price is a secondary source.
These data may be available through Government publications, non-governmental
publications or syndicated services. Some examples are given below:
Government publications
1. Census by Registrar General of India
2. National Income by Central Statistical Organization also statistical abstract, annual
survey of industries.
3. Foreign trade by Director General of Commercial Intelligence.
4. Wholesale price index by Office of Economic Advisor
5. Economic Survey - Dept of Economic Affairs.
6. RBI Bulletin - RBI
7. Agricultural Situation in India - Ministry of Agriculture
8. Indian Labor Year Book - Labor Bureau
9. National Sample Survey Ministry of Planning
Non- government publications
Besides official agencies, there are number of private organizations which bring out
statistics in one form or another on a periodical basis of course industry and trade
associations are important like:
1. Indian Cotton Mills Federation or Confederation of Indian Textile Industry - about
textile industry.
2. Bombay Mill Owners Association - statistics of workers of mills.

3. Bombay Stock Exchange - on financial accounts & ratios.


4. Coffee Board - coffee statistics
5. Coir Board - coir & coir goods
6. Rubber Board - Rubber statistics
7. Federation of Indian Chambers of Commerce & Industry (FICCI)
Syndicated services
Syndicated services are provided by certain organization which collect and tabulate
information on continuous basis. Reports based on marketing information are sent
periodically to subscribers. A number of research agencies offer customized research
services to their clients like consumer research, advertising research etc.
Publication by international organizations
There are several International organizations that publish statistics in their respective
areas.
SUMMARY
In this chapter the importance of secondary data has been outlined. Disadvantage of
secondary data have been dealt in detail in this chapter. Sources of secondary data have
been outlined in this chapter.
KEY TERMS

Primary data
Secondary data
Advantages and disadvantages
Evaluation of secondary data
Sources of secondary data
Governmental publications
Syndicated services

QUESTIONS
1. Discuss the difference between primary and secondary data.
2. Explain the advantages and disadvantages of secondary data.
3. Write short notes on nature and scope of secondary data.
4. How will you evaluate the secondary data sources?

5. Discuss the internal and external sources of secondary data.

- End of Chapter LESSON - 13


QUESTIONNAIRE DESIGN

OBJECTIVES

To recognize the importance and relevance of questionnaire design


To recognize that the type of information will influence the structure of
questionnaire
To understand the role data collection method in designing questionnaire
To understand how to plan and design without mistakes, and improve its layout
To know importance of pretesting.

STRUCTURE

Questionnaire design
Phrasing question
Art of asking question
Layout of traditional questionnaires

Many experts in survey research believe that improving the wording of questions can
contribute far more to accuracy than can improvements in sampling. Experiments have
shown that the range of error due to vague questions or use of imprecise words may be
as high as 20 or 30 percent. Consider the following example, which illustrates the
critical importance of selecting the word with the right meaning. The following
questions differ only in the use of the words should, could, and might:

Do you think anything should be done to make it easier for people to pay doctor
or hospital bills?
Do you think anything could be done to make it easier for people to pay doctor or
hospital bills?
Do you think anything might be done to make it easier for people to pay doctor or
hospital bills?

The results from the matched samples: 82 percent replied something should be done, 77
percent replied something could be done, and 63 percent replied something might be
done. Thus, a 19 percent difference occurred between the two extremes, should and
might. Ironically, this is the same percentage point error as in the Literary Digest Poll,
which is a frequently cited example or error associated with sampling.

The chapter outlines procedure for questionnaire design and illustrates that a little bit of
research knowledge can be a dangerous thing.
A Survey Is Only As Good As the Questions It Asks
Each stage of the business research process is important because of its interdependence
with other stages of the process. However, a survey is only as good as the questions it
asks. The importance of wording questions is easily overlooked, but questionnaire
design is one of the most critical stages in the survey research process.
"A good questionnaire appears as easy to compose as does a good poem. But it is usually
the result of long, painstaking word". Business people who are inexperienced in business
research frequently believe that constructing a questionnaire in a matter of hours.
Unfortunately, newcomers who naively believe that common sense and good grammar
are all that are needed to construct a questionnaire generally learn that their hasty
efforts are inadequate.
While common sense and good grammar are important in question writing, more is
required in the art of questionnaire design. To assume that people will understand the
questions is a common error. People simply may not know what is being asked. They
may be unaware of the product or topic interest, they may confuse the subject with
something else, or the question may not mean the same thing to everyone interviewed.
Respondents may refuse to answer personal questions. Further, properly wording the
questionnaire is crucial, as some problems may be minimized or avoided altogether if a
skilled researcher composes the questions.
QUESTIONNAIRE DESIGN: AN OVERVIEW OF THE MAJOR DECISIONS
Relevance and accuracy are the two basic criteria a questionnaire must meet if it is to
achieve the researcher's purpose. To achieve these ends, a researcher who systematically
plans a questionnaire's design will be required to make several decisions - typically, but
not necessarily, in the order listed below:
1. What should be asked?
2. How should each question be phrased?
3. In what sequence should the questions be arranged?
4. What questionnaire layout will best serve the research objectives?
5. How should the questionnaire be pretested? Does the questionnaire need to be
revised?
What Should Be Asked?

During the early stages of the research process, certain decisions will have been made
that will influence the questionnaire design. The preceding chapters stressed the need to
have a good problem definition and clear objectives for the study. The problem
definition will indicate which type of information must be collected to answer the
manager's questions; different types of questions may be better at obtaining certain
types of information than others. Further, the communication mediums used for data
collection - telephone interview, personal interview, or self-administered survey will
have been determined. This decision is another forward linkage that influences the
structure and content of the questionnaire. The specific questions to be asked will be a
function of the pervious decisions later stages of the research process also have an
important impact o questionnaire wording. For example, determination of the questions
that should be asked will be influenced by the requirements for data analysis. As the
questionnaire is being designed, the researcher should be thinking about the types of
statistical analysis that will be conducted.
Questionnaire Relevancy
A questionnaire is relevant if no unnecessary information is collected and if the
information that is needed to solve the business problem is obtained. Asking the wrong
or an irrelevant question is a pitfall to be avoided. If the task is to pinpoint
compensation problems, for example, questions asking for general information about
morale may be inappropriate. To ensure information relevancy, the researcher must be
specific about data needs, and there should be a rationale for each item of information.
After conducting surveys, many disappointed researchers have discovered that some
important questions were omitted. Thus, when planning the questionnaire design, it is
essential to think about possible omissions. Is information being collected on the
relevant demographic and psychographic variables? Are there any questions that might
clarify the answers to other questions? Will the results of the study provide the solution
to the manager's problem?
Questionnaire Accuracy
Once the researcher has decided what should be asked, the criterion of accuracy
becomes the primary concern. Accuracy means that the information is reliable and valid
while experienced researchers generally believe that one should use simple,
understandable, unbiased, unambiguous, nonirritating words, no step-by-step
procedure to ensure accuracy in question writing can be generalized across projects.
Obtaining accurate answers from respondents is strongly influenced by the researcher's
ability to design a questionnaire that facilitates recall and that will motivate the
respondent to cooperate.
Respondent tend to be most cooperative when, the subjects of the research is
interesting. Also, if questions are not lengthy, difficult to answer, or ego threatening,
there is higher probability of obtaining unbiased answers, question wording and
sequence substantially influence accuracy. These topics are treated in subsequent
sections of this chapter.

PHRASING QUESTIONS
There are many ways to phrase question, and many standard question formats have
been developed in previous research studies. This section presents h classification of
question types and provides some helpful guidelines to researchers who must write
questions.
Open-Ended Response versus Fixed-Alternative Questions
Questions may be categorized as either of two basic types, according to the amount of
freedom respondents are given in answering them. Response questions pose some
problem or topic and ask the respondent to answer in his or her own words. For
example:
What things do you like most about your job?
What names of local banks can you think of offhand?
What comes to mind when you look at this advertisement?
Do you think that there are some ways in which life in the United States is getting
worse? How is that?
If the question is asked in a personal interview, the interviewer may probe for more
information by asking such questions as: Anything else? or Could you tell me more
about your thinking on that? Open-ended response questions are free-answer
questions. They may be contrasted to the fixed-alternative question, sometimes called a
"closed question", in which the respondent is given specific, limited-alternative
responses and asked to choose the one closest to his or her own viewpoint. For example:
Did you work overtime or at more than one job last week?
Yes ____ No _____
Compared to ten years ago, would you say that the quality of most products made in
Japan is higher, about the same, or not as good?
Higher ____ About the same _____ Not as good _____
Open-ended response questions are most beneficial when the researcher is conducting
exploratory research, especially if the range of responses is not known. Open-ended
questions can be used to learn what words and phrases people spontaneously give to the
free-response questions. Respondents are free to answer with whatever is uppermost in
their thinking. By gaining free and uninhibited responses, a researcher may find some
unanticipated reaction toward the topic. As the responses have the "flavor" of the
conversational language that people use in talking about products or jobs, responses to
these questions may be a source for effective communication.

Open-ended response questions are especially valuable at the beginning of an interview.


They are good first questions because they allow respondents up warm up to the
questioning process.
The cost of open-ended response questions is substantially greater than that of fixedalternative questions, because the job of coding, editing and analyzing the data is quite
extensive. As each respondent's answer is somewhat unique, there is some difficulty in
categorizing and summarizing the answers. The process requires an editor to go over a
sample of questions to classify the responses in to some sort of scheme, and then all the
answers are received and coded according to the classification scheme.
Another potential disadvantage of the open-ended response question is that interviewer
may influence the responses. While most instructions state that the interviewer is to
record answers verbatim, rarely can even the best interviewer get every word spoken by
the respondent. There is a tendency for interviewer to take short-cuts in recording
answers - but changing even a few of the respondents' words may substantially influence
the results. Thus, the final answer often is a combination of the respondent's and the
interviewer's ideas rather than the respondent's ideas alone.
The simple-dichotomy or dichotomous-alternative question requires the
respondent to choose one of two alternatives. The answer can be a simple "yes" or "no"
or a choice between "this" and "that". For example:
Did you make any long-distance calls last week?
Yes _____

No _____

Several types of questions provide the respondent with multiple-choice alternatives. The
determinant-choice questions require the respondent to choose one and only one
response from among several possible alternatives. For example:
Please give us some information about your flight. In which section of the aircraft did
you sit?
First Class _______ Business Class ______

Coach Class ______

The frequency-determination question is a determinant-choice question that asks


for an answer about general frequency of occurrence. For example: How frequently do
you watch the MTV television channel?
__ Every day
__ 5-6 times a week
__ 2-4 times a week
__ Once a week

__ Less than a week


__ Never
Attitude rating scales, such as the Likert Scale, Semantic Differential, and Stapel
Scale, are also fixed-alternative questions.
The checklist question allows the respondent to provide multiple answers to a single
question. The respondent indicates past experience, preference, and the like merely by
checking off an item. In many cases the choices are adjectives that describe a particular
object. A typical checklist follows:
Please check which of the following sources of information about investments you
regularly use, if any.
__ Personal advice of your brokers(s)
__ Brokerage newsletters
__ Brokerage research reports
__ Investment advisory service(s)
__ Conversations with other investors
__ Reports on the internet
__ None of these
__ Other (please specify)
Most questionnaires include a mixture of open-ended and closed questions. Each form
has unique benefits; in addition, a change of pace can eliminate respondent boredom
and fatigue.
Phrasing Questions for Self-Administered, Telephone, and Personal
Interview Surveys
The means of data collection (personal interview, telephone, mail, or Internet
questionnaire) will influence the question format and question phrasing. In general,
questions for mail and telephone surveys must be less complex than those utilized in
personal interviews. Questionnaires for telephone and personal interviews should be
written in a conversational style. Consider the following question from a personal
interview:
There has been a lot of discussion about the potential health threat to nonsmokers from
tobacco smoke in public building, restaurants, and business offices. How serious a

health threat to you personally is the inhaling of this secondhand smoke, often called
passive smoking: Is it a very serious health threat, somewhat serious, not too serious,
or not serious at all?
1. Very serious
2. Somewhat serious
3. Not too serious
4. Not serious at all
5. Don't know
THE ART OF ASKING QUESTIONS
In develop a questionnaire, there are no hard-and-fast rules. Fortunately, however,
some guidelines that help to prevent the most common mistakes have "been developed
from research experience.
1. Avoid Complexity: Use Simple, Conversational Language
Words used in questionnaires should be readily understandable to all respondent. The
researcher usually has the difficult task of adopting the conversational language of
people from the lower educational levels without talking down to better-educated
respondents. Remember, not all people have the vocabulary of a college student. A
substantial number of Americans never go beyond high school.
Respondents can probably tell an interviewer whether they are married, single,
divorced, separated, or windowed, but providing their "marital status" may present a
problem. Also, the technical jargon of corporate executives should be avoided when
surveying retailers, factory employees, or industrial users. "Marginal analysis,"
"decision support systems," and other words from the language of the corporate staff
will not have the same meaning to- or be understood by- a store owner / operator in a
retail survey. The vocabulary in following question (from an attitude survey on social
problems) is probably confusing for many respondents:
When effluents from a paper mill can be drunk, and exhaust from factory smokestacks
can be breathed, then humankind will have done a good job in saving the
environment... Don't you agree that what we want is zero toxicity and no effluents?
This lengthy question is also a leading question.
2. Avoid Leading and Loaded Questions

Leading and loaded questions are a major source of bias in question wording. Leading
Questions suggest or imply certain answers. In a study of the dry-cleaning industry, this
question was asked:
Many people are using dry cleaning less because of improved wash-and-wear clothes.
How do you feel wash-and-wear clothes have affected your use of dry-cleaning
facilities in the past 4 years?
_______Use less

______No change

__________Use more

The potential "bandwagon effect" implied in this question threatens the study's validity.
Loaded questions suggest a socially desirable answer or are emotionally charged.
Consider the following:
In light of todays farm crisis, it would be in the public's best interest to have the
federal government require labeling of imported meat.
_____Strongly ____Agree ____Uncertain
Disagree
____Strongly disagree

____

Answers might be different if the loaded portion of the statement, "farm crisis" had
another wording suggesting a problem of less magnitude than a crisis. A television
station produced the following 10-second spot asking for viewer feedback:
We are happy when you like programs on Channel 7. We are sad when you dislike
programs on Channel 7. Write to us and let us know what you think of our
programming.
Most people do not wish to make others sad. This question is likely to elicit only positive
comments. Some answers to certain questions are more socially desirable than others.
For example, a truthful answer to the following classification question might be painful
Where did you rank academically in your high school graduating class?
___Top quarter ___2nd quarter

___3rd quarter

___4th quarter

When taking personality tests, respondents frequently are able to determine which
answers are most socially acceptable, even though those answers do not portray their
true feelings.
3. Avoid Ambiguity: Be a Specific as Possible
Items on questionnaires are often ambiguous because they are too general. Consider
indefinite words such as often, usually, regularly, frequently, many, good, fair, and poor.
Each of these words has many meanings. For one person, frequent reading of Fortune
magazine may be reading six or seven issues a year; for another it may be two issues a

year. The word fair has a great variety of meanings; the same is true for many indefinite
words.
Questions such as the following should be interpreted with care:
How often do you feel that you can consider all of the alternatives before making a
decision to follow a specific course of action?
___Always

___Fairly

___Occasionally

___Seldom ___Never often

In addition to utilizing words like occasionally, this question asks respondents to


generalize about their decision-making behaviour. The question is not specific. What
does consider mean? The respondents may have a tendency to provide stereotyped
"good" management responses rather than to describe their actual behavior. People's
memories are not perfect. We tend to remember the good and forget the bad.
4. Avoid Double- Barreled Items
A question covering several issues at once is referred to as double-barreled and should
always be avoided. It's easy to make the mistake of asking two questions rather than
one. For example:
Please indicate if you agree or disagree with the following statement: "I have called in
sick or left work to golf". Which reason is it - calling in sick or leaving work (perhaps
with permission) to play golf?
When multiple questions are asked in one question, the results may be exceedingly
difficult to interpret. For example, consider the following question from a magazine
survey entitled "How Do You Feel about Being a Woman?":
Between you and your husband, who does the housework (cleaning, cooking,
dishwashing, laundry) over and above that done by any hired help?
I do all of it
I do almost all if it
I do over half of it
We split the work fifty-fifty
My husband does over half of it
The answers to this question do not tell us if the wife cooks and the husband dries the
dishes.
5. Avoid Making Assumptions

Consider the following question:


Should Mary's continue its excellent gift-wrapping program?
___Yes

___ No

The question contains the implicit assumption that people believe the gift-wrapping
program is excellent. By answering yes, the respondent implies that the program is, in
fact, excellent and that things are just fine as they are. By answering no, he or she
implies that the store should discontinue the gift wrapping. The researcher should not
place the respondent in that sort of bind by including an implicit assumption in the
question.
6. Avoid Burdensome Questions That May Tax the Respondent's Memory
A simple fact of human life is that people forget. Researchers writing questions about
past behavior or events should recognize that certain questions may make serious
demands on the respondent's memory. Writing questions about prior events requires a
conscientious attempt to minimize the problem associated with forgetting.
It many situations, respondents cannot recall the answer to a question. For example, a
telephone survey conducted during the 24-hour period following airing of the Super
Bowl might establish whether the respondent watched the Super Bowl and then ask: "Do
you recall any commercials on that program?" If the answer is positive, the interviewer
might ask: "what brands were advertised?" These two questions measure unaided
recall, because they give the respondent no clue as to the brand of interest.
What is the Best Question Sequence?
The order of questions, or the question sequence, may serve several functions for the
researcher. If the opening questions are interesting, simple to comprehend, and easy to
answer, respondents' cooperation and involvement can be maintained throughout the
questionnaire. Asking easy-to-answer questions teaches respondents their role and
builds confidence; they know this is a researcher and not another salesperson posing as
an interviewer. If respondents' curiosity is not aroused at the outset, they can become
disinterested and terminate the interviewer. A mail research expert reports that a mail
survey terminates the interview. A mail research expert reports that a mail survey
among department store buyers drew an extremely poor return. However, when some
introductory questions related to the advisability of congressional action on pending
legislation of great importance to these buyers were placed first on the questionnaire, a
substantial improvement in response rate occurred. Respondents completed all thequestions, not only those in the opening section.
In their attempts to "warm up" respondents toward the questionnaire dependent
researchers frequently ask demographic or classification questions at the beginning of
the questionnaire. This is generally not advisable. It may embarrass or threaten
respondents. It is generally better to ask embarrassing questions at the middle or end of

the questionnaire, after rapport has been established between respondent and
interviewer.
Sequencing specific questions before asking about broader issues is a common cause of
order bias. For example, bias may arise if questions about a specific clothing store are
asked prior to those concerning the general criteria for selecting a clothing store.
Suppose a respondent who indicates in the first portion of a questionnaire that the
shops at a store where parking needs to be improved. Later in the questionnaire, to
avoid appearing inconsistent, she may state that parking is less important a factor than
she really believes it is. Specific questions may thus influence the more general ones.
Therefore, it is advisable to ask general questions before specific questions to obtain the
freest of open ended responses. This procedure, known as the funnel technique, allows
for researcher to understand the respondents frame of reference before asking more
specific questions about the level of the respondents information and the intensity of
his or her opinions.
One advantage of internet surveys is the ability to reduce order bias by having the
computer randomly order questions and/or response alternatives. With complete
randomization, question order is random and respondents see response alternatives in
random positions. Asking a question that doesnt apply to the respondent or that the
respondent is not qualified to answer may be irritating or may cause a biased response
because the respondent wishes to please the interviewer or to avoid embarrassment.
Including a filter question minimizes the chance of asking questions that are
inapplicable. Asking "where do you generally have cheque-cashing problems in Delhi"
may elicit a response even though the respondent has not had any cheque-cashing
problems and may simply wish to please the interviewer with an answer. A filter
question such as:
Do you ever have a problem cashing a cheque in Delhi? ___Yes

___No

would screen out the people who are not qualified to answer.
Another form of filter question, the pivot question, can be used to obtain income
information and other data that respondents may be reluctant to provide. For example,
a respondent is asked.
"Is your total family income over Rs.75,000?" IF NO, ASK...
"Is it over or under Rs.50,000?" IF UNDER, ASK
"Is it over or under Rs.25,000?"
So, the options are
1. Over Rs.75,000
2. Rs.50,001 - Rs.75,000

3. Rs.25,001 - Rs.50,000
4. Under Rs.25,001
Structuring the order of questions so that they are logical will help to ensure the
respondents cooperation and eliminate confusion or indecision. The researcher
maintains legitimacy by making sure that the respondent can comprehend the
relationship between a given question and the overall purpose of the study.
What is the best layout?
Good layout and physical attractiveness are crucial in mail, Internet, and other selfadministered questionnaires. For different reasons it is also important to have a good
layout in questionnaires designed for personal and telephone interviews.
LAYOUT OF TRADITIONAL QUESTIONNAIRES
The layout should be neat and attractive, and the instructions for the interviewer should
be clear. Often money can be spent on an incentive to improve the attractiveness and
quality of the questionnaire.
Mail questionnaires should never be overcrowded. Margins should be of decent size,
white space should be used to separate blocks of print, and any unavoidable columns of
multiple boxed should be kept to a minimum. All boldface capital letters should easy to
follow.
Questionnaires should be designed to appear as brief and small as possible. Sometimes
it is advisable to use a booklet form of questionnaire, rather than a large number of
pages stapled together.
In situations where it is necessary to conserve space on the questionnaire or to facilitate
entering the data into a computer or tabulating the data, a multiple-grid layout may be
used. In this type of layout, a question is followed by corresponding response
alternatives arranged in a grid or matrix format.
Experienced researchers have found that is pays to phrase the title of the questionnaire
carefully. In self-administered and mail questionnaires a carefully constructed title may
by itself capture the respondents interest, underline the important of the research
("Nationwide Study of Blood donors"), emphasize the interesting nature of the study
("Study of Internet Usage"), appeal to the respondents ego ("Survey among Top
Executives"), or emphasize the confidential nature of the study ("A Confidential Survey
among"). The researcher should take steps to ensure that the wording of the title will
not bias the respondent in the same way that a leading question might.
When an interviewer is to administer the questionnaire, the analyst can design the
questionnaire to make the job of following interconnected questions much easier by

utilizing instruction, directional arrows, special question formats, and other tricks of the
trade.
SUMMARY
Many novelists write, rewrite, and revise certain chapters, paragraphs, and even
sentences of their books. The research analyst lives in a similar world. Rarely does one
write only a first draft of a questionnaire. Usually, the questionnaire is tried out on a
group that is selected on a convenience basis and that is similar in makeup to the one
that ultimately will be sampled. Researchers should select a group that is not too
divergent from the actual respondents. (e.g. business students as surrogates for business
people), but it is not necessary to get a statistical sample for protesting. The protesting
process allows the researchers to determine if the respondents have any difficulty
understanding the questionnaire and whether there are any ambiguous or biased
questions. This process is exceedingly beneficial. Making a mistake with 25 or 50
subjects can avert the disaster of administering an invalid questionnaire to several
hundred individuals.
KEY TERMS

Open ended response questions


Fixed alternative questions
Leading question
Loaded question
Double-barreled question
Funnel technique
Filter question
Pretesting

QUESTIONS
1. What is the difference between leading question and loaded question?
2. Design an open end question to measure a reaction to a particular advertisement.
3. Design a complete a questionnaire to evaluate job satisfaction.
4. Develop a checklist to consider in questionnaire construction.

- End of Chapter LESSON 14

MEASUREMENT

OBJECTIVES

To know what is to be measured


To define the operation definition and scale measurement
To distinguish among nominal, ordinal, interval, and ratio scales
To understand the criteria of good measurement
To discuss the various methods of determining reliability
To discuss the various methods of assessing validity

STRUCTURE

Measurement
Scale measurement
Types of scale
Criteria for good measurement

MEASUREMENT
Measurement is an integral part of the modern world. Today we have progressed in the
physical sciences to such as extent that we are now able to measure the rotation of a
distant star, the attitude in micro-inches and so on. Today such a precise physical
measurement is very critical. In many business situations, the majority of the
measurements are applies to things that are much more abstract than attitude or time.
The accurate measurement is essential for effective decision making. The purpose of this
chapter is to provide with a basic understanding of the measurement process and rules
needed for developing sound scale measurements.
In management research, measurement is viewed as the integrative process of
determining the amount (intensity) of information about constructs, concepts or objects
of interest and their relationship to a defined problem or opportunity. It is important to
understand the two aspects of measurement one is construct development, which
provides necessary and precise definition which begins the research process called
problem definition in turn determine what specific data should be collected. Another is
scale measurement means how the information is collected with reference to construct.
In other words, the goal of construct development is to precisely identify and define
what is to be measured including dimensions. In turn, the goal of scale measurement is
to determine how to precisely measure the constructs.
Regardless of whether the researcher is attempting to collect primary data or secondary
data, all data can be logically classified as under.
a) State-of-Being Data

When the problem requires collecting responses that are pertinent to the physical,
demographical or socioeconomic characteristics of individuals, objects or organizations,
the resulting raw data are considered as state-of-being data. This data represent factual
characteristics that can be verified through several sources other than the persons
providing the information.
b) State-of-Mind Data
This represents that mental attributes of individuals that are not directly observable or
available through some other external sources. It exists only within the minds of people.
The researcher has to ask a person to respond the stated questions. Examples are
personality traits, attitudes, feelings, perceptions, beliefs, awareness level, preferences,
images etc.
c) State-of-Behavior Data
This represents an individuals or organizations current observable actions or reactions
or recorded past actions. A person may categorically ask the past behavior. This can be
checked using external secondary sources, but that is very difficult process in terms of
time, effort and accuracy.
d) State-of-Intension Data
This represents individuals or organizations expressed plans of future behavior. Again
this also collected by asking carefully defined questions. Like the above data, this also
very difficult to verify through external, secondary sources, but verification is possible.
With the background information about the type of data which are collected, the
following pages will be very useful in understanding the concepts of scale measurement.
SCALE MEASUREMENT
Scale measurement can be defined as the process of assigning a set of descriptions to
represent the range of possible responses to a question about a particular object or
construct. Scale measurement directly determines the amount of raw data that can be
ascertained from a given questioning or observation method. This attempts to assign
designated degrees of intensity to the responses, which are commonly referred to as
scale points. The researcher can control the amount of raw data that can be obtained
from asking questions by incorporating scale properties or assumption in scale points.
There are four scaling properties that a researcher can use in developing scales namely
assignment, order distance and origin.
1. Assignment (also referred to as description or category property): It is the
researchers employment of unique description to identity each object within a set, e.g.,
the use of numbers, colors, yes & no responses.

2. Order refers to the relative magnitude between the raw responses. It establishes and
creates hierarchical rank-order relationship coming objects, e.g., 1st place is better than
4th place.
3. Distance, is the measurement the express the exact difference between the two
responses. This allows the researcher and respondent to identify, understand, and
accurately express absolute difference between objects, e.g., Family A has 3 children and
Family B has 6 children.
4. Origin, refers to the use of a unique starting as being "true zero" e.g., asking
respondent his or her weight or current age, market share of specific brand.
TYPES OF SCALES
While scaling properties determine the amount of raw data that can be obtained from
any scale design, all questions and the scale measurement can be logically and
accurately classified as one of four basic scale types: nominal, ordinal, integral or ratio.
A scale may be defined as any series of items that are arranged progressively according
to value or magnitude, into which an item can be placed according to its quantification.
The following table represents the relationship between types of scales & scaling
properties:

1. Nominal Scale
In business research, nominal data are probably more widely collected than any other. It
is the simplest type of scale and also the most basic of the four types of scale designs. In
such a scale, the numbers serve as labels to identify persons, objects or events. Nominal
scales are the least powerful of the four data types. They suggest no order or distance
relationship and have no arithmetic origin. This scale allows the researcher only to
categorize the raw responses into mutually exclusive and collectively exhaustive. In the
nominal scale, the only operation is the counting of numbers in each group. An example
of typical nominal scale in business research is the coding of males as 1 and females as 2.

Example 1:
Please indicate your current marital status.
____Married

____Single

___Never married ___Widowed

Example 2:
How do you classify yourself?
____Indian ___American ___Asian ___Black
2. Ordinal Scale
As the name implies, these are ranking scales. Ordinal data include the characteristics of
the nominal data plus an indicator of order, means, this data activates both are
assignment and order scaling properties. The researcher can rank-order the raw
responses into a hierarchical pattern. The use of ordinal data scale implies a statement
of "greater than" or "less than" without stating how much greater or less. Examples of
ordinal data include opinion and preference scales. A typical ordinal scale in business
research asks respondents to rate career opportunities, brands, companies etc., as
excellent, good, fair or poor.
Example 1:
Which of the following one category best describes your knowledge about computers?
1) Excellent 2) Good

3) Basic

4) Little

5) No knowledge

Example 2:
Among the listed below, please indicate top three preference using 1, 2, 3 as your
choice in the respective source provided:
By post
By courier
By telephone
By speed post
By internet
By person

Also to be noted is that the individual ranking can be combined and get a collective
ranking of a group.
3. Interval Scales
The structure of this scale not only show the assignment, order scaling properties but
also the distance property with interval scale, researchers can identity not only some
type of hierarchical order among to raw data but also the specific differences between
the data. The classic example of this scale is the Fahrenheit temperature scale. If a
temperature is 80 degree, it cannot be said that it is twice as hot as 40 degree. The
reason is that 0 degree does not represent the lack of temperature, but a relative point
on the Fahrenheit scale. Similarly, when this scale is used to measure psychological
attributes, the researcher can comment on the magnitude of differences or compare the
average differences but cannot determine the actual strength of attitude toward an
object. However many attitude scales are presumed to be interval scales. Interval scales
are more powerful than nominal and ordinal scales. Also they are quicker to complete
and it is convenient for researcher.
Example 1:
Into which of the following categories does your income fall?
1. Below 5000
2. 5000 - 10,000
3. 10,000 - 15,000
4. 15,000 - 25,000
5. above 25,000
Example 2:
Approximately how long you lived in the current address?
1. Less than 1 year
2. 1-3 years
3. 4-6 years
4. More than 6 years
4. Ratio Scales

This is the only scale that simultaneously activates all four scaling properties. A ratio
scale tends to be the most sophisticated scale in the sense that it allows not only to
identify the absolute differences between each represents but also absolute
comparisons.
Examples of ratio scales are the commonly used physical dimensions such as height,
weight, distance, money value and population counts. It is necessary to remember that
ratio scale structures are designed to allow a "zero" or "true state of nothing" response to
be a valid raw response to the question. Normally, the ratio scale requests that
respondents give a specific singular numerical value as their response, regardless of
whether or not a set of scale points used. The following are the examples of ratio scales:
Example 1:
Please circle the numbers of children below 18 years of ages in your house?
0

Example 2:
In past seven days, how many times did you go to retail shop?
____ number of times
Mathematical and Statistical Analysis of Scales
The type of scale that is utilized in business research will determine the form of
statistical analysis. For example certain operations can be conducted only if a scale, of a
particular nature. The following will show the relationship between scale types and
measures of central tendency and dispersion.

Here,
A - Appropriate
More A - More appropriate
Most A - Most appropriate
IA - Inappropriate
Criteria for good measurement
There are four major criteria for evaluating measurement: reliability, validity, sensitivity
and practicality.
1. Reliability
It refers to the extent to which a scale can reproduce the same measurement results in
repeated trials. Reliability applies to a measure when similar results are obtained
overtime across situations. Broadly defined, reliability is the degree to which measures
are free from error and therefore yield consistent results. As discussed in the earlier
chapter the error in scale measurements leads to lower scale reliability. Two dimensions
underline the concept of reliability: one is repeatability and the other is internal
consistency.
First, the test-retest method involves administrating the same scale or measure to the
same respondents at two separate times to test for stability. If the measure is stable over
time, the test, administered under the same conditions each time, should obtain similar
results. The high stability correlation or consistency between the two measures at time 1
and 2 indicates a high degree of reliability.
The second dimension of reliability concerns the homogeneity of the measure. The
Split-half technique can be used when the measuring tool has many similar
questions or statements to which subjects can respond. The instrument is administered
and the results are separated by item into even and odd numbers or randomly selected
halves. When the two halves are correlated, if the result of the correlation is high, the
instrument is said to be high reliability in internal consistency.
The Spearman-Brown Correction Formula is used to adjust the effect of test
length and to estimate reliability of the whole set. But, this approach may influence the
integral consistency because of the way in which the test is split. In order to overcome
Kuder - Richardson Formula (KR 20) and Cronbach's Coefficient Alpha are
two frequently used examples. The KR 20 is the method from which alpha was
generalized and is used to estimate reliability for dichotomous items. Cronbach's alpha
has the most utility for multi-scale items at the interval level of measurement.

The third perspective on reliability considers how much error may be introduced by
different investigators or different samples of items being studied. In other words, the
researcher creates two similar yet different scale measurements for the given construct.
An example of this is the scoring of Olympic skaters by a panel of judges.
2. Validity
The purpose of measurement is to measure what it intend to measure; but this is not as
simple as it sounds at first. Validity is the ability of a measure to measure what it is
proposed to measure. If it does not measure what it is designated to measure, there will
be problems. To assess the validity there are second ways which are discussed here
under.
Face validity or Content validity, refers to the subjective agreement among
professionals that a scale logically appears to reflect accurately what it intend to
measure. When it appears evident to experts that the measure provides adequate
coverage of the concept, a measure has face validity.
Criterion - Related validity reflects the success of measures used for prediction or
estimation. Criterion validity may be classified as either concurrent validity or
predictive validity, depending on the time sequence in which the 'new' measurement
scale and the criterion measure are correlated. If the new measure is taken at the same
time as the criterion measure and shown to be valid, then it has concurrent validity.
Predictive validity is established when a new measure predicts a future event. These two
measures differ only on the basis of time.
Construct validity is established by the degree to which a measure confirms the
hypotheses generated from theory based on concepts. It implies the empirical evidence
generated by a measure with the theoretical logic. To achieve this validity, the
researcher may use convergent validity (should converge with similar measure) or
discriminant validity (when it has low correlation with the measures of dissimilar
concepts.)
3. Sensitivity
It is an important concept, particularly when changes in attitude or other hypothetical
constructs are under investigation. It refers to an instruments ability to accurately
measure variability in stimuli or responses. A dichotomous response category such as
"agree" or "disagree" does not allow attitude change. But the scale staring from "strongly
agree", "agree", "neither agree nor disagree", "disagree" and "strongly disagree"
increases the sensitivity.
4. Practicality
It can be defined in terms of economy, convenience and interpretability. This means the
scientific requirements for the measurement process is called reliable and valid, while

operational requirements called it as practical where the above mentioned three aspects
are more important.
SUMMARY
This paper outlined the importance of the measurement. Different types of scales have
been dealt in detail. This chapter has given the criteria for a good measurement.
KEY TERMS

Nominal scale
Ordinal scale
Integral scale
Ratio scale
Reliability
Split-half Technique
Spearman-Brown Correction Formula
Kuder - Richardson Formula (KR 20)
Cronbach's Coefficient Alpha
Validity
Face validity
Content validity
Criterion - Related validity - concurrent validity or predictive validity
Construct validity
Sensitivity
Practicality

QUESTIONS
1. What are different types of data in the attitude measurement could be collected?
2. Discuss the measurement scaling properties.
3. Explain different scales of measurement.
4. Is the statistical analysis is based on the type of scale? Explain.
5. What do you mean by good measurement?
6. Explain various methods of reliability and validity.

- End of chapter LESSON 15

ATTITUDE MEASUREMENT AND SCALING TECHNIQUES

OBJECTIVES

To understand the definition of attitude


To learn the techniques for measuring attitudes

STRUCTURE

Techniques for measuring attitude


Physiological measures of attitude
Summated rating method
Numerical scale
Graphic rating scale

ATTITUDE DEFINED
There are many definitions for the term attitude. An attitude is usually viewed as an
enduring disposition to respond consistently in a given manner to various aspects of the
world, including persons, events, and objects. One conception of attitude is reflected in
this brief statement: "Sally loves working at Sam's. She believes it's clean, conveniently
located, and has the best wages in town She intends to work there until she retires. In
this short description are three components of attitude: the affective, the cognitive,
and the behavioral.
The affective component reflects an individual's general feelings or emotion toward
an object. Statements such as "I love my job", "I liked that book, A Corporate Bestiary,
and "I hate apple juice" - reflect the emotional character of attitudes.
The way one feels about a product, a person, or an object is usually tied to ones beliefs
or cognitions. The cognitive component represents one's awareness of and
knowledge about an object. A woman might feel happy about her job because she
"believes that the pay is great" or because she knows "that my job is the biggest
challenge in India."
The third component of an attitude is the behavioral component. Intention and
behavioral expectations are included in this component, which therefore reflects a
predisposition to action.
Techniques for Measuring Attitudes
A remarkable variety of techniques have been devised to measure attitudes part, this
diversity stems from the lack, of consensus about the exact definite of the concept.
Further, the affective, cognitive, and behavioral component an attitude may be
measured by different means. For example, sympathetic nervous system responses may

be recorded using physiological measures to measure affect but they are not good
measures of behavioral intentions. Direct verbal statements concerning affect, belief, or
behavior are utilized to measure behavioral intent. However, attitudes may also be
measured indirectly by using the qualitative explanatory techniques. Obtaining verbal
statements from respondents generally requires that the respondent perform a task such
as ranking, rating, sorting, or making a choice or a comparison.
A ranking task requires that the respondents rank order a small number of items on
the basis of overall preference or some characteristic of the stimulus. Rating asks the
respondents to estimate the magnitude of a characteristic or quality that an object
possesses. Quantitative scores, along a continuum that has been supplied to the
respondents, are used to estimate the strength of the attitude or belief. In other words,
the respondents indicate the position, on a scale, where they would rate the object.
A sorting technique might present respondents with several product concepts,
printed on cards, and require that the respondents arrange the cards into a number of
piles or otherwise classify the product concepts. The choice technique, choosing one
of two or more alternatives, is another type of attitude measurement. If a respondent
chooses one object over another, the researcher can assume that the respondent prefers
the chosen object over the other.
The most popular techniques for measuring attitudes are presented in this chapter.
Physiological Measures of Attitudes
Measures of galvanic skin response, blood pressure, and pupil dilation and other
physiological measures may be utilized to assess the affective component of attitudes.
They provide a means of measuring attitudes without verbally questioning the
respondent. In general, they can provide a gross measure of like or dislike, but they are
not sensitive measures for identifying gradients of an attitude.
Attitude Rating Scales
Using rating scales to measure attitudes is perhaps the most common practice in
business research. This section discusses many rating, scales designed to enable
respondents to report the intensity of their attitudes.
Simple Attitude Scales
In this most basic form, attitude scaling requires that an individual agree or disagree
with a statement or respond to a single question. For example, respondents in a political
poll may be asked whether they agree or disagree with the statement "The president
should run for re-election", or an individual might be asked to indicate whether he likes
or dislikes labor unions. Because this type or self-rating scale merely classifies
respondents into one of two categories, it has only the properties of a nominal scale.
This, of course, limits the type of mathematical analysis that may be utilized with this
basic scale. Despite the disadvantages, simple attitude scaling may be used when

questionnaires are extremely long, when respondents have little education, or for other
specific reasons.
Most attitude theorists believe that attitudes vary along continua. An early attitude
researcher pioneered the view that the task of attitude scaling is to measure the distance
from "good" to "bad", "low" to "high", "like" to "dislike , and so on. Thus the purpose of
an attitude scale is to find an individual's position on the continuum. Simple scales do
not allow for making fine distinctions in attitudes. Several scales have been developed to
help make more precise measurements.
Category Scales
Some rating scales have only two response categories: agree and disagree. Expanding
the response categories provides the respondent more flexibility in the rating task. Even
more information is provided if the categories are ordered according to a descriptive or
evaluative dimension. Consider the questions below:
How often is your supervisor courteous and friendly to you?
___Never

___Rarely ___Often ___Very often

Each of these category scales is a more sensitive measure than a scaled with only two
response categories. Each provides more information.
Wording is an extremely important factor in the usefulness of these scales. Exhibit 14.1
shows some common wordings for category scales.

Summated Ratings Method: The Likert Scale


Business researchers' adaptation of the summated ratings method, developed by Rensis
Likert, is extremely popular for measuring attitudes because the method is simple to
administer. With the Likert scale, respondents indicate their attitudes by checking
how strongly they agree or disagree with carefully constructed statements that range
from very positive to very negative toward the attitudinal object. Individuals generally
choose from five alternatives: strongly agree, agree, uncertain, disagree, and strongly
disagree; but the number of alternatives may range from three to nine.
Consider the following example from a study on mergers and acquisitions:
Mergers and acquisitions provide a faster means of growth than internal expansions.
Strongly Disagree
(1)

Disagree
(2)

Uncertain

Agree

Strongly agree

(3)

(4)

(5)

To measure the attitude, researchers assign scores or weights to the alternative


responses. In thus example, weights of 5, 4, 3, 2, and 1 are assigned to the answers. (The
weights, shown in parentheses, would not be printed questionnaire). Because the
statement used as an example is positive towards the attitude, strong agreement
indicates the most favorable attitudes on the statement and is assigned a weight of 5. If a
negative statement toward the object (such as "Your access to copy machines is limited")
were given the weights would be reversed, and "strongly degree" would be assigned the
weight of 5. A single scale item on a summated rating scale is an ordinal scale.
A Likert scale may include several scale items to form an index. Each statement is
assumed to represent an aspect of a common attitudinal domain For example, Exhibit
14.2 shows the items in a Likert scale to measure attitudes toward a management by
objectives program. The total score is the summation of the weights assigned to an
individual's response.
For example:
Here are some statements that describe how employees might feel about the MBO
(management by objectives, form of management. Please indicate your agreement or
disagreement for each statement. Please encircle the appropriate number to indicate
whether you:
1 - Strongly Agree

2 Agree

3 Neutral

4 Disagree

5 - Strongly Disagree

Circle one and only one answer for each statement. There are no right or wrong
answers to these questions:

In Likert's original procedure, a large number of statements are generated and then an
item analysis is performed. The purpose of the item analysis is to ensure that final
items evoke a wide response and discriminate among those with positive and negative
attitudes. Items that are poor because they lack clarity or elicit mixed response patterns
are eliminated from the final statement list. However, many business researchers do not
follow the exact procedure prescribed by Likert. Hence, many business researches do
not follow the exact procedure prescribed by Likert. Hence, a disadvantage of the Likerttype summated rating method is that it is difficult to know what a single summated
score means. Many patterns of response to the various statements can produce the same
total score. Thus, identical total scores may reflect different "attitudes" because
respondents endorsed different combinations of statements.
Semantic Differential
The semantic differential is a series of attitude scales. This popular attitudemeasurement technique consists of presenting an identification of a company, product,
brand, job, or other concept, followed by a series of seven-point bipolar rating scales.
Bipolar adjectives, such as "good" and "bad", "modern" and "old-fashioned", or "clean"
and "dirty," anchor the beginning and end (or poles) of the scale.
Modern_____:______ :_____:______:_____:_____ :_____ Old-Fashioned
The subject makes repeated judgments of the concept under investigation on each of the
scales.
The scoring of the semantic differential can be illustrated by using the scale bounded by
the anchors "modern" and "old-fashioned." Respondents are instructed to check the
place that indicates the nearest appropriate adjective. From left to right, the scale
intervals are interpreted as extremely modern, very modern, slightly modern, both
modern and old-fashioned, slightly old-fashioned, very old-fashioned, and extremely
old-fashioned. A weight is assigned to each position on the rating scale. Traditionally,
scores are 7, 6, 5, 4, 3, 2, 1, or +3, +2, +1, 0, -1, -2, -3.

Many researchers find it desirable to assume that the semantic differential provides
interval data. This assumption, although widely accepted, has its critics, who argue that
the data have only ordinal properties because the weights are arbitrary. Depending on
whether the data are assumed to be interval or ordinal, the arithmetic mean or the
median is utilized to plot the profile of one concept, product, unit, etc., compared with
another concept, product, or units.
The semantic differential technique was originally developed by Charles Osgood and
others as a method for measuring the meaning of objects or the "semantic space" of
interpersonal experience." Business researchers have found the semantic differential
versatile and have modified it for business applications.
Numerical Scales
Numerical scales have numbers, rather than "semantic space" or verbal descriptions as
response options to identify categories (response positions). If the scale items have five
response positions, the scale is called a 5-point numerical scale; with seven response
positions, it is called a 7-point numerical scale, and so on.
Consider the following numerical scale:
Now that you've had your automobile for about one year, please tell us how satisfied
you are with your Ford Ikon:,
Extremely Satisfied
7

Extremely Dissatisfied
6

This numerical scale utilizes bipolar adjectives in the same manner as the semantic
differential.
Constant-Sum Scale
If a Parcel Service company wishes to determine the importance of the attributes of
accurate invoicing, delivery as promised, and price to organizations that use its service
in business-to-business marketing. Respondents might be asked to divide a constant
sum to indicate the relative importance of the attributes. For example:
Divide 100 points among the following characteristics of a delivery service according
to how important each characteristic is to you when selecting a delivery company.
Accurate invoicing___
Delivery as promised___
Lower price___

The constant-sum-scale works best with respondents with high educational levels. If
respondents follow instructions correctly, the results approximate interval measures. As
in the paired-comparison method, as the number of stimuli increases this technique
becomes more complex.
Stapel Scale
The Stapel scale was originally developed in the 1950s to measure the direction and
intensity of an attitude simultaneously. Modern versions of the scale use a single
adjective as a substitute for the semantic differential when it is difficult to create pairs of
bipolar adjectives. The modified Stapel scale places a single adjective in the center of an
even number of numerical values (for example, ranging from +3 to -3). It measures how
close to or how distant from the adjective a given stimulus is perceived to be.
The advantages and disadvantages of the Stapel scale are very similar to those of the
semantic differential. However, the Stapel scale is markedly easier to administer,
especially over the telephone. Because the Stapel scale does not requires bipolar
adjectives, as does the semantic differential, the Stapel scale is easier to construct.
Research comparing the semantic differential with the Stapel scale indicates that results
from the two techniques are largely the same.
Graphic Rating Scale
A graphic rating scale presents respondents with graphic continuum. The respondents
are allowed to choose any point on the continuum to indicate their attitudes. Typically, a
respondent's score is determined by measuring the length (in millimeters) from one end
of the graphic continuum to the point marked by the respondent. Many researchers
believe scoring in this manner strengthens the assumption that graphic rating scales of
this type are interval scales. Alternatively, the researcher may divide the line into
predetermined scoring categories (lengths) and record respondent's marks accordingly.
In other words, the graphic rating scale has the advantage of allowing the researchers to
choose any interval they wish for purposes of scoring. The disadvantage of the graphic
rating scale is that there are no standard answers.
Thurstone Equal-Appearing Interval Scale
In 1927, Louis Thurstone, an early pioneer in attitude research, developed the concept
that attitudes vary along continua and should be measured accordingly. Construction of
a Thurstone scale is a rather complex process that requires two stages. The first stage is
a ranking operation, performed by judges, who assigns scale values to attitudinal
statements. The second stage consists of asking subjects to respond to the attitudinal
statements.
The Thurstone method is time-consuming and costly. From a historical perspective it is
valuable, but its current popularity is low, because it is rarely utilized in most applied
business research.

Scales Measuring Behavioral Intentions and Expectations


The behavioral component of an attitude involves the behavioral expectations of an
individual toward an attitudinal object. Typically, this represents an intention or a
tendency to seek additional information. Category scales that measure the behavioral
component of an attitude attempt to determine a respondent's "likelihood" of action or
intention to perform some future action, as in the following examples:
How likely is it that you will change jobs in the next six months

I definitely will change.


I probably will change.
I might change.
I probably will not change.
I definitely will not change.

I would write a letter to my congressmen or other government official in support of


this company if it were in a dispute with government.

Extremely likely
Very likely

Somewhat likely
Likely, about 50-50 chance
Somewhat unlikely
Very unlikely
Extremely unlikely

Behavioral Differential
A general instrument, the behavioral differential, has been developed to measure the
behavioral intentions of subjects toward an object or category of objects. As in the
semantic differential, a description of the object to be judged is placed on the top of a
sheet, and the subjects indicate their behavioral intentions toward this object on a series
of scales. For example, one item might be:
A 25-year-old female commodity broker
Would: ______:_____

:_____ :_____ :_____ :_____:Would not

ask this person for advice.


Ranking
People often rank order their preferences. An ordinal scale may be developed by asking
respondents to rank order (from most preferred to lease preferred) a set of objects or
attributes. It is not difficult for respondents to understand the task of rank ordering the
importance of fringe benefits or arranging a set of job tasks according to preference.
Paired Comparisons
The following question is the typical format for asking about paired comparisons.
I would like to know your overall opinion of two brands of adhesive bandages. They
are Curad brand and Band-Aid brand. Overall, which of these two brands - Curad or
Band-Aid - do you think is the better one? Or are both the same?
Curad is better____
Band-Aid is better____
They are the same____
Ranking objects with respects to one attribute is not difficult if only a few concepts or
items are compared. As the number of items increases, the number of comparisons
increases geometrically. If the number of comparisons is too great, respondents may
become fatigued and no longer carefully discriminate among them.
Sorting

Sorting tasks requires that respondents indicate their attitudes or beliefs by arranging
items.
SUMMARY
This chapter describes the technique for measuring attitude. This paper outlined the
importance of the attitude.
KEY TERMS

Attitude
Affective component
Cognitive component
Behavioral component
Ranking
Rating
Category scale
Likert scale
Semantic differential scale
Numerical scale
Constant sum scale
Stapel scale
Graphic rating scale
Paired comparison

QUESTIONS
1. What is an attitude?
2. Distinguish between rating and ranking. Which is a better attitude measurement?
Why?
3. Describe the different methods of scale construction, pointing out the merits and
demerits of each.
4. What advantages do numerical scales have over semantic differential scales?

- End of chapter LESSON 16


STATISTICAL TECHNIQUES

OBJECTIVES

Know the nature of Statistical study


Recognize the importance of Statistics as also its limitations
Differentiate descriptive Statistics from inferential Statistics.

STRUCTURE

Major characteristics of statistics


Description statistics
Inferential statistics
Central tendency of data
Use of different average
Types of frequency distribution
Measure of dispersion

INTRODUCTION
Business researchers edit and code data to provide input that results in tabulated
information that will answer research question. With this input, the results can be
produced statistically and logically. Aspects of Statistics are important if the quantitative
data are to serve their purpose. If Statistics, as a subject, is inadequate and consists of
poor methodology, we would not know the right procedure to extract from the data the
information they contain. On the other hand, if our figures are defective in the sense
that they are inadequate or inaccurate, we would not reach the right conclusions even
though our subject is well developed. With this brief introduction, let us first see how
Statistics has been defined.
Major characteristics of statistics:
1. Statistics are aggregates of facts. This means that a single figure is not Statistics. For
example, national income of a country for a single year is not Statistics but the same for
two or more years is.
2. Statistics are affected by a number of factors. For example, sale of a product depends
on a number of factors such as its price, quality, competition, the income of the
consumers, and so on.
3. Statistics must be reasonably accurate. Wrong figures, if analyzed, will lead to
erroneous conclusions. Hence, it is necessary that conclusions must be based on
accurate figures.
4. Statistics must be collected in a systematic manner. If data are collected in a
haphazard manner, they will not be reliable and will lead to misleading conclusions.

5. Finally, Statistics should be placed in relation to each other. If one collects data
unrelated to each other, then such data will be confusing and will not lead to any logical
conclusions. Data should be comparable over tin e and over space.
Subdivisions in Statistics
The statisticians commonly classify this subject into two broad categories: descriptive
statistics and inferential statistics.
1. Descriptive Statistics
As the name suggests descriptive statistics includes any treatment designed to describe
or summaries the given data, bringing out their important features. These statistics do
not go beyond this. This means that no attempt is made to infer anything that pertains
to more than the data themselves. Thus, if someone compiles the necessary data and
reports that during the financial year 2000-2001, there were 1500 public limited
companies in India of which 1215 earned profits and the remaining 285 companies
sustained losses, his study belongs to the domain of descriptive Statistics. He may
further calculate the average profit earned per company as also average loss sustained
per company. This set of calculations too is a part of descriptive statistics.
Methods used in descriptive statistics may be called as descriptive methods. Under
descriptive methods, we learn frequency distribution, measures of central tendency, that
is, averages, measures of dispersion and skewness.
2. Inferential Statistics
Although descriptive Statistics is an important branch of Statistics and it continues to be
so, its recent growth indicates a shift in emphasis towards the methods of Statistical
inference. A few examples may be given here. The methods of Statistical inference are
required to predict the demand for a product such as tea or coffee for a company for a
specified year or years. Inferential Statistics are also necessary while comparing the
effectiveness of a given medicine in the treatment of any disease.
Again, while determining the nature and extent of relationship between two or more
variables like the number of hours studied by students and their performance in their
examinations, one has to take recourse to inferential Statistics.
Each of these examples is subject to uncertainty on account of partial, incomplete, or
indirect information. In such cases, the Statistician has to judge the merits of all possible
alternatives in order to make the most realistic prediction or to suggest the most
effective medicine or to establish a dependable relationship and the reasons for the
same. In this text, we shall first discuss various aspects of descriptive Statistics. This will
be followed by the discussion on different topics in inferential Statistics. The latter will
understandably be far more comprehensive than the former.
CENTRAL TENDENCY OF DATA

In many frequency distributions, the tabulated values show small frequencies at the
beginning and at the end and very high frequency at the middle of the distribution. This
indicates that the typical values of the variable lie near the central part of the
distribution and other values cluster around these central values. This behavior of the
data about the concentration of the values in the central part of the distribution is called
central tendency of the data. We shall measure this central tendency with the help of
mathematical quantities. A central value which 'enables' to comprehend in a single
effort the significance of the whole is known as Statistical Average or simply Average. In
fact, an average of a statistical series is the value of the variable which is representative
of the entire distribution and, therefore, gives a measure of central tendency.
Measures of Central Tendency
There are three common measures of central tendency
I. Mean
II. Median
III. Mode
The most common and useful measure of central tendency is, however the Mean. In the
following articles the method of calculation of various measures of central tendency will
be discussed. In all such discussion we need a very useful notation known as
Summation.
Choice of a Suitable Average
The different statistical average has different characteristics. There is no all-purpose
average. The choice of a particular average is usually determined by the purpose of
investigation. Within the framework of descriptive statistics, the main requirement is to
know what each average means and then select the one that fulfils the purpose at hand.
The nature of distribution also determines the type of average to be used.
Generally the following points should be kept in mind while making a choice of average
for use
1. Object
The average should be chosen according to the object of enquiry. If all the values in a
series are to be given equal importance, then arithmetic mean will be a suitable choice.
To determine the most stylish or most frequently occurring item mode should be found
out. If the object is to determine an average that would indicate its position or ranking in
relation to all the values, naturally, median should be the choice. If small items are to be
given greater importance than the big items, geometric mean is the best mean.
2. Representative

The average chosen should be such that it represents the basic characteristics of the
distribution.
3. Nature of form of data
If the frequently distribution is symmetrical or nearly symmetrical X, M or Mo may be
used almost interchangeably. If there are open-end class intervals, mean cannot be
calculated definitely. In a closed frequently distribution of unequal class intervals, it is
impossible to determine mode accurately. If there are a few values, it may not be
possible to determine mode. Mean will not give a representative picture, if there are few
extremely large or small values at either end of the array, and yet the great majority of
the values concentrate around a narrow band. In a variable of non-continuous type,
median or mode may give a value that actually exists in the data.
Davis' Test: Arithmetic mean is considered as an appropriate average for use for data
which has a symmetrical distribution or even if it has a moderate degree of asymmetry.
Prof. George Davis has devised a test which is:

If this coefficient works out to be more than +0.20, the distribution is symmetrical
enough to use arithmetic mean.
4. Characteristics of Average
While choosing a suitable average for a purpose, the merits and demerits of various
averages should always be considered and that average which fits into the purpose most
should be preferred over others. The following points should be given due consideration
in the process of selection of an average.
(i) In certain commonly encountered applications, the mean is subject to less sampling
variability than the median or mode.
(ii) Given only the original observations, the median is sometimes easiest to calculate.
Sometimes when there is no strong advantage for the mean, this advantage is enough to
indicate the use of the median.
(iii) Once a frequency distribution has been formed, the mode and the median are mode
quickly calculated than the mean. Moreover, when some classes are open-ended the
mean cannot be calculated from the frequency distribution.
(iv) The median is not a good measure when there are very few possible values for the
observations as with number of children or size of family.

(v) The mode and the median are relatively little affected by 'extreme' observations.
(vi) Calculations of geometric mean and harmonic mean is difficult as it involves the
knowledge of logarithms and reciprocals.
Hence "the justification of employing them (averages) must be determined by an appeal
to all the facts and in the light of the peculiar characteristics of the different types".
Uses of Different Averages
Different averages, due to their inherent characteristics are appropriate in different
circumstances. Thus, their use may be guided by the purpose at hand or circumstances
in which one is. Here a brief discussion is being made of uses of different statistical
averages:
1. Arithmetic Average
The arithmetic average used in the study of a social, economic or commercial problem
like production, income, price, imports, exports, etc. The central tendency if these
phenomena can best be studied by taking out an arithmetic average. Whenever we talk
of an 'average income' or 'average production' or 'average price' we always mean
arithmetic average of all these things. Whenever there is no indication about the type of
the average to be used, arithmetic average is computed.
2. Weighted Arithmetic Average
When it is desirable to give relative importance to the different items of a series,
weighted arithmetic average is computed. If it is desired to compute per capita
consumption of a family, due weights should be assigned to children, males and females.
This average is also useful in constructing numbers. The weighted average should be
used in the following cases:
a) If it is desired to have an average of whole group, which is divided into a
number of sub-classes, widely divergent from each other?
b) When items falling in various sub-classes change in such a way that the
proportion which the items bear among themselves also undergoes a change.
c) When combined average has to be computed.
d) When it is desired to calculate to find an average of ratios, percentages or rates.
3. Median
Median is especially applicable to cases which are not capable of precise quantitative
studies such as intelligence, honesty, etc. It is less applicable in economic or business
statistics, because there is lack of stability in such data.

4. Mode
The utility of mode is being appreciated more and more day by day. In r the sciences of
Biology and Meteorology it has been found to be of great value. In commerce and
industry it is gaining very great importance. Whenever as shop-keeper wants to stock
the goods he sells, he always looks to the modal size of those goods. Model size of shoes,
is of great importance to the businessman dealing in ready-made garments or shoes.
Many problems of production are related with mode. Many business establishments
these days are engaging their attention in keeping statistics of their sale ascertain the
particulars of the modal articles sold.
5. Geometric Mean
Geometric mean can advantageously be used in the construction of index numbers. It
makes the index numbers reversible and gives equal weight to equal ratio of changes.
This average is also useful in measuring the growth of population, because increases in
geometric progression. When there is wide dispersion in a serious, geometric mean is a
useful average.
6. Harmonic Mean
This average is useful in the cases where time, rate and prices are involved. When it is
desired to give the largest weight to the smallest item, this average is used.
Summation Notation ()
The symbol (read: sigma) means summation.
If x1, x2, x3, , xn be the n values of a variable x. Then their sum x1 +x2+ x3+ + xn is
shortly written as

Similarly, the sum w1x1+w2x2+ + wnxn is denoted by

Some important results

There are three types of mean


1. Arithmetic Mean (AM)
2. Geometric Mean (GM)
3. Harmonic Mean (HM)
Of the three means, Arithmetic Mean is most commonly used. In fact, if no specific
mention be made, by Mean we shall always refer to Arithmetic Mean (AM) and calculate
accordingly.
Simple Arithmetic Mean
Definition: The Arithmetic Mean (x ) of a given series of values, say x1, x2, x3, , xn, is
defined as the sum of these values divided by their total number; thus
Weighted Arithmetic Mean is defined by:

Example 1: Find the AM of 3, 6, 24, and 48.


Solution: AM = (3 + 6 + 24 + 48) / 4 = 81/4 = 20.25

Weighted Arithmetic Mean

Definition: x1, x2, x3, , xn be n values of a variable x, and if f1, f2, , fn be their respective
weights (or frequencies), then the weighted arithmetic mean is defined by

where, N = f = total frequency


SKEWNESS
A frequency distribution is said to be symmetrical when the values of the variable
equidistant from their mean have equal frequencies. If a frequency distribution is not
symmetrical, it is said to be asymmetrical or skewed. Any deviation from symmetry is
called Skewness.
In the words of Riggleman and Frisbee: "Skewness is the lack of symmetry. When a
frequency distribution is plotted on a chart, skewness present in the items tends to be
dispersed more on one side of the mean than on the other".
Skewness may be positive or negative. A distribution is said to be positively skewed if
the frequency curve has a longer tail towards the higher values of x, ie, if the 'frequency
curve gradually slopes down towards the high values of x.
For a positively skewed distribution, Mean (M) > Median (Me) > Mode (Mo)
A distribution is said to be negatively skewed if the frequency curve has a longer tail
towards the lower values of x.
For a negatively skewed distribution, Mean (M) < Median (Me) < Mode
(Mo)

For a symmetrical distribution, Mean (M) = Median (Me) = Mode (Mo)

Measures of Skewness
The degree of skewness is measured by its coefficient. The common measures of
skewness are:
1. Pearson's first measure
Skewness = (Mean Mode) / Standard Deviation

2. Pearson's second measure


Skewness = 3 x (Mean - Mode)/ Standard Deviation
3. Bowley's Measure

where Q1, Q2, Q3 are the first, second and third quartiles respectively.
4. Moment Measure
Skewness = m3 / 3 = m3 / m23/2
where m2 and m3 are the second and third central moments and is the S.D.
All the four measure of Skewness defined above are independent of the units of
measurement.
Example:
Calculate the Pearson's measure of Skewness on the basis of Mean, Mode, and Standard
Deviation.

Solution:
According to Pearson's first measure, Skewness = (Mean Mode) / Standard Deviation
Here mid-values of the class-intervals are given. Assuming a continuous series, we
construct the following table:

Types of Frequency Distributions


In general, frequency distributions that form a balanced pattern are called symmetrical
distributions, and those that form an unbalanced pattern are called skewed or
asymmetrical distributions. In a symmetrical distribution frequencies go on increasing
up to a point and then begin to decrease in the same fashion. A special kind of
symmetrical distribution is the normal distribution; the pattern formed is not only
symmetrical but also has the shape of a bell.
In a symmetrical distribution, mean median and mode coincide and they lie at the
centre of distribution. As the distribution departs from symmetry these three values are
pulled apart. A distribution, in which more than half of the area under the curve is to the
right side of the mode, is a positively skewed distribution. In a positively skewed
distribution, its right tail is longer than its left tail. Under such a distribution mean is

greater than the median, and the median is greater than the mode (M > Me > Mo), and
the difference between upper quartile and median is greater than the difference between
median and lower quartile (Q3-Me > Me-Q1). In a negatively skewed distribution, more
than half of the area under the distribution curve is to the left side of the mode. In such a
distribution the elongated tail us to the left and mean is less than the median and
median is less than the mode (M < Me < Mo), and the difference between upper quartile
and median is less than the difference between median and lower quartile (Q3-M < MQ1).The following figures show these facts. The following table will also show these facts
of Position of Average on Various Distributions.

Test of Skewness
In order to find out whether a distribution is symmetrical or skewed, the following facts
should be noticed:
1. Relationship between Averages
If in a distribution mean, median and mode are not identical, then it is a skewed
distribution. The greater is the difference between mean and mode more will be the
skewness in the distribution.
2. Total of Deviations
If the sum of positive deviations from median or mode is equal to the sum of negative
deviation, then is no skewness in the distribution. The extent of difference between the
sums of positive and negative deviations from median or mode will determine the extent
of skewness in the data.
3. The distance of partition values from median

In a symmetrical distribution Q1 and Q2, D1 and D9 and P10 and P90 are equidistant from
median. In an asymmetrical distribution it is not so.
4. The frequencies on either side of the mode
In an asymmetrical distribution, frequencies on either side of the mode are not equal.
5. The curve
When the data are plotted on a graph paper, the curve will not be bell-shaped, or when
cut along a vertical line through the centre, the two halves will not be identical.
Conversely stated, in the absence of skewness in the distribution:
(i) Values of mean, median and mode will coincide.
(ii) Sum of the positive deviations from the median or mode will be equal to the sum of
negative deviations.
(iii) The two quartiles, deciles one and nine, and percentile ten and ninety will be
equidistant from the median.
(iv) Frequencies on the either side of the mode will be equal.
(v) Data when plotted on a graph paper will take a bell-shaped form.
Measures of Skewness
To find out the direction and the extent of symmetry in a series statistical measures of
skewness are calculated, these measures can be absolute or relative. Absolute measures
of skewness tell us the extent of asymmetry and whether it is positive or negative. The
absolute skewness can be known by taking the deference between mean and mode.
Symbolically,
Absolute SK = X - Mo
If the value of mean is greater than the mode (M > Mo) skewness will be positive. In
case the value of mean is less than the mode (M < Mo) skewness will be negative. The
greater is the amount of skewness, the more the mean and mode differ because of the
influence of extreme items. The reason why the difference between mean and mode is
taken for the measure of skewness is that in a symmetrical distribution, both the values
along with median coincide, but in an asymmetrical distribution, there will be a
difference between the mean and mode.
Thus the difference between the mean and the mode, whether positive or negative,
indicates that the distribution is asymmetrical. However such absolute measure of
skewness is unsatisfactory, because:

(1) It cannot be used for comparison of skewness in tow distributions if they are in
different units, because the difference between the mean and the mode will be in terms
of the units of distribution.
(2) The difference between the mean and mode may be more in one series and less in
another, yet the frequency curves of the two distributions may be similarly skewed. For
comparison, the absolute measures of skewness are changed to relative measures, which
are called Coefficient of Skewness.
There are four measures of relative skewness. They are:
1. The Karl Pearson's Coefficient of Skewness
2. The Bowley's Coefficient of Skewness.
3. The Kelly's Coefficient of Skewness
4. Measure of skewness based onmoments. Measures of Skewness
1. The Karl Pearson's Coefficient of Skewness
Karl Pearson has given a formula, for relative measure of Skewness. It is known as Karl
Pearson's Coefficient of Skewness or Pearsonian Coefficient of Skewness. The formula is
based on the difference between the mean and mode divided by the standard deviation.
The coefficient is represented by J

If in a particular frequency distribution, the mode is ill-defined, the coefficient of


Skewness can be determined by the following changed formula.

This is based on the relationship between different averages in a moderately


asymmetrical distribution. In such a distribution:

The Pearsonian coefficient of skewness has the interesting characteristic that it will be
positive when the mean is larger than the mode or median, and it will be negative when
the arithmetic mean is smaller than the mode or median. In a symmetrical distribution,
the value of Pearsonian coefficient of skewness will be zero.
There is no theoretical limit to this measure, however, in practical the value given by this
formula is rarely very high and usually lies between . The direction of the skewness is
given by the algebraic sign of the measure; if it is plus then the skewness is positive, if it
is minus, the skewness is negative. The degree of skewness is obtained by the numerical
figure such as 0.9, 0.4, etc.
Thus this formula gives both the direction as well as the degree of skewness. There is
another relative measure of skewness also based on the position of averages. In this, the
difference between two averages is divided by the mean deviation. The formula is:

These formulas are not very much used in practice, because of demerits of mean
deviation.
DISPERSION
Measures of Dispersion
An average nay give a good idea of the type of data, but it alone can't reveal all the
characteristics of data. It cannot tell us in what manner all the values of the variable are
scattered / dispersed about the average.
Meaning of Dispersion
The Variation or Scattering or Deviation of the different values of a variable from their
average is known as Dispersion. Dispersion indicates the extent to which the values vary
among themselves. Prof. W.I. King defines the term, 'Dispersion' as it is used to indicate

the facts that within a given group, the items differ from another in size or in other
words, there is lack of uniformity in their sizes. The extent of variability in a given set of
data is measured by comparing the individual values of the variable with the average all
the values and then calculating the average of all the individual differences.
Objectives of Measuring Variations
1. To serve as a basis for control of the variability itself.
2. To gauge the reliability of an average
3. To serve as a basis for control of the variability itself
Types of Measures of Dispersion
There are two types of measures of dispersion. The first, which may be referred to as
Distance Measures, describes the spread of data in terms of distance between the values
of selected observations. The second are those which are in terms of an average
deviation from some measure of central tendency.
Absolute and Relative Measures of Dispersion
Measures of absolute dispersion are in the same units as the data whose scatter they
measure. For example, the dispersion of salaries about an average is measured in rupees
and the variation of time requires for workers to do a job is measured in minutes or
hours. Measures of absolute dispersion cannot be used to compare the scatter in one
distribution with that in another distribution when the averages of the distributions
differ in size or the units of measure differ in kind. Measures of relative dispersion show
some measure of scatter as a percentage or coefficient of the absolute measure of
dispersion. They are generally used to compare the scatter in one distribution with the
scatter in other. Relative measure of dispersion is called coefficient of dispersion.
Methods of Measuring Dispersion
There are two meanings of dispersion, as explained above. On the basis of these two
meanings, there are two mathematical methods of finding dispersion, i.e. methods of
limits and methods of moment. Dispersion can also be studied graphically. Thus, the
following are the methods of measuring dispersion:
I. Numerical Methods
1. Methods of Limits
i. The Range
ii. The Inter-Quartile Range

iii. The Percentile Range


2. Methods of Moments
i. The first moment of dispersion or mean deviation
ii. Second moment of dispersion from which standard deviation is
computed
iii. Third moment of dispersion
3. Quartile Deviation
II. Graphic Method
Lorenz Curve
Range
The simplest measure of dispersion is the range of the data. The range is determined by
the two extreme values of the observations and it is simply the differences between the
largest and the smallest value in a distribution. Symbolically,
Range (R) = Largest value (L) - Smallest value (S)
Coefficient of Range (CR) = (Largest Value + Smallest Value) / (Largest
Value - Smallest Value)

Quartile Deviation or Semi-Interquartile Range


Definition: Quartile Deviation (Q) is an absolute measure of dispersion and is defined by
formula:
Q = (Q3 Q1) / 2
Where, Q1 and Q2 are the first (or lower) and the third (or upper) quartiles respectively.
Here Q3 Q1 is the interquartile range and hence quartile deviation is called Semiinterquartile Range.
As it is based only on the Q1 and Q3, it does not take into account the variability of all the
values and hence it is not very much used for practical purposes.

Example:
Find the quartile deviation of the following frequency distribution.
Daily wages:
No of workers:

10-15

15-20

20-25

25-30

30-35

12

18

10

Solution:

N/4 = 50/4 = 12.5 and 3N/4 = 37.5.


By simple interpolation,
(Q1 15) / (20 15) = (12.5 6) / (18 6)
(Q1 15) / 5 = 6.5 / 12
Q1 = (6.5 / 12) x 5 + 15
Q1 = 17.71
Similarly,
(Q3 25) / (30 25) = (37.5 36) / (46 36)
(Q3 25) / 5 = 1.5 / 10
Q3 = 0.15 x 5 + 25
Q3 = 25.75
Hence, Quartile Deviation,

(Q3 Q1) / 2 = (25.75 17.71) / 2 = 4.02


MEAN DEVIATION
(or Average Deviation or Mean Absolute Deviation)
Definition: Mean Deviation of a series of values of a variable is the arithmetic mean of
all the absolute deviations (i.e., difference without regard to sign) from any one of its
averages (Mean, Median or Mode, but usually Mean or Median). It is an absolute
measure of dispersion.
Mean deviation of a set of n values x1, x2,, xn about their AM is defined by

where x = values or mid-value accordingly to data being ungrouped or grouped, and x


bar = Mean

where M = Median and d= x M = Value (or mid value) Median.


Similarly, we can define Mean Deviation about Mode.

Note: The expression |d| is read as mod. d and gives only numerical or absolute value of
d without regard to sign. Thus,
|-3| = 3, |-4| = 4, |0.56| = 0.56
The reason for taking only the absolute and not the algebraic values of the deviation is
that the algebraic sum of the deviations of the value from their mean is zero.
Example:
Find the Mean Deviation about the Arithmetic Mean of the numbers 31, 35, 29, 63, 55,
72, 37.
Solution:
Arithmetic Mean = (31 + 35 + 29 + 63 + 55 + 72 + 37) / 7 = 322/7 = 46
Calculation of absolute Deviations

The required Mean Deviation about the Mean = |d| / n = 104 / 7 = 14.86
Advantages
Mean deviation is based on all the values of the variable and sometimes gives fairly good
result as a measure of dispersion. However the practice of neglecting signs and taking

absolute deviations for the calculation of the mean Deviation seems rather unjustified
and this makes algebraic treatment difficult.
STANDARD DEVIATION
It is the most important absolute measures of Dispersion. Standard Deviation of set
values of a variable is defined as the positive square root of arithmetic mean of the
squares of all deviations of the values from their arithmetic mean. In short it is the
square root of the mean of the squares of deviations from mean.
If x1, x2, ...,xn be a series of values of a variable and x bar their AM, then S.D () is
defined by

The square of Standard Deviation is known as Variance i.e., Variance 2 = (SD)2


SD is often defined as the positive square root of Variance.
Example:
Find the standard deviation of the following numbers: 1, 2, 3, 4, 5, 6, 7, 8, and 9
Solution:
The deviations of the numbers from the AM 5 are respectively -4, -3, -2, -1, 0, 1, 2, 3, and
4
The squares of the deviations from AM are 16, 9, 4, 1, 0, 1, 4, 9, and 16
Therefore,

= (16 + 9 + 4 + 1 + 0 + 1 + 4 + 9 + 16) / 9 = 2.58


Advantages and Disadvantages

Standard Deviation is the most important and widely used among the measures of
dispersion and it possesses almost all the requisites of a good measure of dispersion. It
is rigidly defined and based on all the values of the variable.
It is suitable for algebraic treatment. SD is less affected by sampling fluctuations than
any other absolute measure of dispersion.
SD is difficult to understand. The process of squaring the deviations from their AM and
then taking the square-root of the AM of these squared deviations is a complicated
affair.
The calculation of SD can be made easier by changing the origin and the scale
conveniently.
Relative Measures of Dispersion
Absolute measures expressed as a percentage of a measure of a control tendency gives
relative measures of dispersion. Relative measures are independent of the units of
measurement and hence they are used for the comparison of dispersion of two or more
distributions given in different units.
Co-efficient of Variation
Co-efficient of variation is the first important relative measure of dispersion and is
defined by the following formula:
Co-efficient of variation = Standard Deviation / Mean x 100
Co-efficient of variation is thus the ratio of the Standard deviation to the mean,
expressed as a percentage. In the words of Karl Pearson, Co-efficient of Variation is the
percentage variation in the mean.

Coefficient of Quartile Deviation


Co-efficient of quartile deviation is a relative measure of dispersion and is defined by
Coefficient of Quartile Deviation = Quartile Deviation / Median x 100

Coefficient of Mean Deviation


It is a relative measure of dispersion. Co efficient of Mean Deviation is defined by
Coefficient of Mean Deviation = Mean Deviation / (Mean or Median) x 100

Example:
Find the Mean Deviation about the Median in respect of the following numbers: 46, 79,
26, 85, 39, 65, 99, 29, 56, and 72. Find also the Co efficient of Mean Deviation.
Solution:
By arranging the given numbers in ascending order of magnitude, we obtain 26, 29, 39,
46, 56, 65, 72, 79, 85, and 99.
Median = [(n+1)/2]th value = [11/2]th value = 5.5th value = (5th value + 6th value) / 2 =
(56+65)/2 = 60.5
Absolute deviation of the values from the Median 60.5 is respectively
34.5, 31.5, 21.5, 14.5, 4.5, 4.5, 11.5, 18.5, 24.5, and 38.5
Therefore, Mean Deviation (MD) about the Median
=(34.5+31.5+21.5+14.5+4.5+4.5+11.5+18.5+24.5+38.5)/2 = 20.4
Co-efficient of MD = MD / Median x 100 = 20.4 / 60.5 x 100 = 33.72%
KURTOSIS
Kurtosis in Greek means bulginess. The degree of kurtosis of a distribution is measured
relative to the peakedness of a normal curve. The measure of kurtosis indicates whether
the curve of the frequency distribution is flat or peaked.
Kurtosis is the peakedness of a frequency curve. In two or more distributions having
same average, dispersion and skewdness, one may have higher concentration of values
near the mode, and its frequency curve will show sharper peak than the others. This
characteristic of frequency distribution is known as Kurtosis.
Kurtosis is measured by the coefficient 2, which is defined by the formula
2 = m4 / m22 = m4 / 4 or by 2 = 2 3, where m4 and m2 are the 2nd and 4th control
moments, and = SD.
A distribution is said to be Platy-kurtic, Meso-kurtic, and Lepto-kurtic corresponding to
2 < 3, 2 = 3, or 2 > 3.
Accordingly,
Its Platy-kurtic, if m4 / 4 < 3

Its Meso-kurtic, if m4 / 4 = 3
Its Lepto-kurtic, if m4 / 4 > 3
Karl person in 1905 introduced the terms MESOKURTIC, LEPTOKURTIC, and
PLATYKURTIC. A peaked curve is called "Leptokurtic" and a flat topped curve is termed
"Platykurtic". These are evaluated by comparison with intermediate peaked curve. These
three curves differ widely in regard to convexity.
Example:
Calculate the measures of the following distribution:

Solution:

V1 = fd / N = 0/100 = 0
V2 = fd2 / N = 446/100 = 4.46
V3 = fd3 / N = -36/100 = -0.36
V4 = fd4 / N = 4574/100 = 45.74
4 = V4 4V1V3 + 6V12V2 3V14
= 45.74 4(0)(-0.36) + 6(0)2(4.46) 3(0)4
4 or 2 = V4/ V22 = 45.74 / 4.462 = 2.3
The value of 2 is less than 3, hence the curve is Platykurtic.
SUMMARY
This chapter helps us to know the nature of the statistical study. This chapter recognizes
the importance of statistics and also its limitations. The differences between descriptive
statistics and inferential statistics are dealt in detail.
KEY WORDS

Descriptive Statistics

Inferential Statistics
Mean
Standard Deviation
Median
Mean Deviation
Mode
Arithmetic Mean
Geometric Mean
Averages
Dispersion
Skewness
Mean deviation
Kurtosis

REVIEW QUESTIONS
1. Explain the special features of measures of central tendency.
2. How will you choose an 'average'?
3. What is dispersion? State its objectives. Explain the various types measures of
dispersion.
4. Explain the various methods of measuring dispersion.
5. Differentiate standard deviation from mean deviation.
6. Define 'skewdness'. How will you measure it?
7. Explain the application of averages in research.

- End of Chapter LESSON 17


MEASURES OF RELATIONSHIP

OBJECTIVES
To study simple, partial and multiple correlation and their application in research

STRUCTURE

Measure of relationship
Correlation
Properties of correlation co-efficient
Methods of studying correlations
Application of correlation

MEASURES OF RELATIONSHIP
The following statistical tools measure the relationship between the variables analyzed
in social science research:
(i) Correlation

Simple correlation
Partial correlation
Multiple correlation

(ii) Regression

Simple regression
Multiple regressions

(iii) Association of attributes


CORRELATION
Correlation measures the relationship (positive or negative, perfect) between the two
variables. Regression analysis considers relationship between variables and estimates
the value of another variable, having the value of one variable. Association of Attributes
attempts to ascertain the extent of association between two variables.
Aggarwal Y.P., in his book 'Statistical Methods' has defined coefficient of correlation as,
"a single number that tells us to what extent two variables or things are related and to
what extent variations in one variable go with variations in the other".
Richard Levin in his book, 'Statistics for Management' has defined correlation analysis
as "the statistical tool that we can use to describe the degree to which one variable is
linearly related to another". He has further stated that, "frequently correlation analysis
is used in conjunction with regression analysis to measure how well the regression line
explains the variation of the dependent variable 'r'. Correlation can also be used by itself,
however, to measure the degree of association between two variables".
Srivastava U.K. Shenoy G.V, and Sharma S.C, in their book 'Quantitative Techniques for
Managerial Decision' have stated that, "correlation analysis is the statistical technique
that is used to describe the degree to which one variable is related to another.

Frequently correlation analysis is also used along with the regression analysis to
measure how well the regression line explains the variations of the dependent variable.
The correlation coefficient is the statistical tool that is used to measure the mutual
relationship between the two variables".
Coefficient of correlation is denoted by 'r'.
The sign of 'r' shows the direction of the relationship between the two variables X and Y.
Positive correlation reveals that there is a positive correlation between the two variables.
Negative correlation reveals negative relationship. Levin states that if an inverse
relationship exists that is, if Y decreases as X increases then r will fall between 0
and -1. Likewise, if there is a direct relationship (if Y increases as X increases), then r
will be a value within the range of 0 to 1.
Aggarwal Y.P. has highlighted in his book 'Statistical Methods', the properties of
correlation and factors influencing the size of the correlation coefficient. The details are
given below:
PROPERTIES OF THE CORRELATION COEFFICIENT
The range of correlation coefficient is from -1 through 0 to +1. The values of r = -1 and r
= +1 reveal a case of perfect relationship, though the direction of relationship is negative
in the first case, and positive in the second case.
The correlation coefficient can be interpreted in terms of r2. It is known as 'coefficient of
determination'. It may be considered as the variance interpretation of r2.
Example:
r = 0.5
r2 = 0.5 x 0.5 = 0.25. In terms of percentage, 0.25x100% = 25%
It refers that 25 percent of the variance in Y scores has been accounted for by the
variance in X.
The correlation coefficient does not change if every score in either or both distribution is
increased or multiplied by a constant.
Causality cannot be inferred solely as the basis of a correlation between two variables. It
can be inferred only after conducting controlled experiments.
The direction of the relation is indicated by the sign (+ or -) of the correlation.
The degree of relationship is indicated by the numerical value of the correlation. A value
near 1 indicates a nearly perfect relation, and a value near 0 indicates no relationship.

In a positive relationship both variables tend to change in the same direction: as X


increases, Y also tends to increase.
The Pearson correlation measures linear (straight line) relationship.
A correlation between X and Y should not be interpreted as a cause-effect relationship.
Two variables can be related without one having a direct effect on the other.
FACTORS INFLUENCING THE SIZE OF CORRELATION COEFFICIENT
1. The size of r is very much dependent upon the variability of measured values m the
correlated sample. The greater the variability, the higher will be the correlation,
everything else being equal.
2. The size of r is altered when researchers select extreme groups of subjects in order to
compare these groups with respect to certain behaviors. Selecting extreme groups on
one variable increases the size of r over what would be obtained with more random
sampling.
3. Combining two groups which differ in their mean values on one of the variables is not
likely to faithfully represent the- true situation as far as the correlation is concerned.
4. Addition of an extreme case (and conversely dropping of an extreme case) can lead to
changes in the amount of correlation. Dropping of such a case leads to reduction in the
correlation while the converse is also true.
TYPES OF CORRELATION
a) Positive or Negative
b) Simple, Partial and Multiple
c) Linear and Non-linear
a) Positive and Negative Correlations
Both the variables (X and Y) will vary in the same direction. If variable X increases,
variable Y also will increase; If variable X decreases, variable Y also will decrease. In
negative correlation, the given variables will vary in opposite direction. If one variable
increases, other variable will decrease.
b) Simple, Partial and Multiple Correlations
In simple correlation, relationship between two variables are studied. In partial and
multiple correlations, three or more variables are studied. In multiple correlation, three
or more variables are simultaneously studied. In partial correlation, more than two

variables are studied, but the effect on one variable is kept constant and relationship
between other two variables is studied
c) Linear and Non-Linear Correlations
It depends upon the constancy of the ratio of change between the variables. In linear
correlation the percentage change in one variable will be equal to the percentage change
in another variable. It is not so in non-linear correlation.
METHODS OF STUDYING CORRELATION
a) Scatter Diagram Method
b) Graphic Method
c) Karl Pearson's Coefficient of Correlation
d) Concurrent Deviation Method
e) Method of Least Squares
Karl Pearson's Coefficient of Correlation

Procedure
i. Compute mean of the X series data
ii. Compute mean of the Y series data
iii. Compute deviations of X series from the mean of X. It is denoted as x.
iv. Square the deviations. It is denoted as x2.
v. Compute deviations of Y series from the mean of Y. It is denoted as y.
vi. Square the deviations. It is denoted as y2.
vii. Multiply deviation (X series, Y series) and compute total. It denoted as xy.

The above values can be applied in the formula and correlation can be computed.
Karl Pearson's Coefficient of Correlation (r)

where,
dx = sum of deviations of X series from the assumed mean
dy = sum of deviations of Y series from the assumed mean
dxdy = total of deviations (X and Y series)
dx2 = deviations of X series from assumed mean are squared
dy2 = deviations of Y series from assumed mean are squared
N = Number of items
The above values can be applied in the above formula and correlation can be computed.
Correlation for the grouped data can be computed with the help of the following
formula:

In the above formula, deviations are multiplied by the frequencies. Other steps are the
same.
CALCULATION OF CORRELATION
Raw Score Method

r = 0.7

Deviation Score Method (using Actual Mean)


Calculate Karl Pearsons Coefficient of Correlation from the following data:

Solution
Karl Pearsons Correlation Coefficients:
Year
1985

Index of
Production (X)
100

x = (XXmean)
-4

x2
16

No. of unemployed y = (Y(Y)


Ymean)
15
0

y2

xy

1986
1987
1988
1989
1990
1991
1992

102
104
107
105
112
103
99
X
= 832

-2
0
+3
+1
+8
-1
-5
x
=0

4
0
9
1
64
1
25
x2
= 120

12
13
11
12
12
19
26
Y
= 120

-3
-2
-4
-3
-3
+4
+11
y
=0

= X / N = 832/8 = 104
= Y / N = 120/8 = 15

r = -92 / (120 x 184)1/2 = -0.619


The Correlation between index of production and unemployed is negative.

Calculation of Correlation Coefficient (using Assumed Mean)


Calculate Coefficient of Correlation from the following data:

Solution

9
+6
4
0
16
-12
9
-3
9
-24
16
-4
121
-55
2
y
xy
= 184 = -92

RANK CORRELATION
It is a method of ascertaining co variability or the lack of it between the two variables.
Rank correlation method is developed by the British Psychologist Charles Edward
Spearmen in 1904. Gupta S.P has stated that "the rank correlation method is used when
quantitative measures for certain factors cannot be fixed, but individual in the group can
be arranged in order thereby obtaining for each individual a number of indicating
his/her rank in the group".
The formula for Rank Correlation

Rank Difference Coefficient of Correlation (Case of no ties in ranks)

Rank Correlation,

= 1 (6 x 38) / 5x(52-1) = 1 228/(5x24) = 1 1.9 = -0.9


Relationship between X and Y is very high and inverse. Relationship between Scores on
Test I and Test II is very high and inverse.
Procedure for Assigning Ranks
First rank is given to the student secured highest score. For example, in Test I, student F
is given first rank, as his score is the highest. The second rank is given to the next
highest score. For example, in Test I, student E is given second rank. Student A and G
have similar scores of 20 each and they stand for 6th and 7th ranks. Instead of giving
either 6th or 7th ranks to both the students, the average of the two ranks [average of 6
and 7] 6.5 is given to each of them. The same procedure is followed to assign ranks to
the scores secured by students in Test II.
Calculation of Rank Correlation when ranks are tied
Rank Difference Coefficient of Correlation (in case of ties in ranks)

= 1 [(6x24) / 10(102-1)] = 1 [144/990] = 0.855


APPLICATION OF CORRELATION
Karl Pearson Coefficient of Correlation can be used to assess the extent of relationship
between motivation of export incentive schemes and utilization of such schemes by
exporters.
Motivation and Utilization of Export Incentive Schemes Correlation
Analysis

Opinion scores of various categories of exporters towards motivation and utilization of


export incentive schemes can be recorded and correlated by using Karl Pearson
Coefficient of Correlation and appropriate interpretation may be given based on the
value of correlation.

Testing of Correlation
't' test is used to test correlation coefficient. Height and weight of a random sample of six
adults is given.

It is reasonable to assume that these variables are normally distributed, so the Karl
Pearson Correlation coefficient is the appropriate measure of the degree of association
between height and weight.
r = 0.875
Hypothesis test for Pearson's population correlation coefficient
H0: p = 0 - this implies no correlation between the variables in the population
H1: p > 0 - this implies that there is positive correlation in the population (increasing
height is associated with increasing weight)
5% significance level

= 0.875 x [(62)1/2] / (10.8752) = 0.875 x 2 / 0.234 = 3.61


Table value of 5% significance level
4 degrees of freedom (n-2) = (6-2) = 2.132
Calculated value is more than the table value. Null hypothesis is rejected. There is
significant positive correlation between height and weight.
Partial Correlation
Partial Correlation is used in a situation where three and four variables involved. There
variables such as age, height and weight are given. Here, partial correlation is applied.
Correlation between height and weight can be computed by keeping age constant. Age
may be the important factor influences the strength of relationship between height and
weight. Partial correlation is used to keep constant the effect age. The effect of one

variable is partially out from the correlation between other two variables. This statistical
technique is known as partial correlation.
Correlation between variables x and y is denoted as rxy
Partial correlation is denoted by the symbol r123. This is correlation between variables 1
and 2, keeping 3rd variable constant.

where,
r123 = partial correlation between variables 1 and 2
r12 = correlation between variables 1 and 2
r13 = correlation between variables 1 and 3
r23 = correlation between variables 2 and 3

Multiple Correlation
Three or more variables are involved in multiple correlation. The dependent variable is
denoted by X1 and other variables are denoted by X2, X3 etc. Gupta S. P. has expressed
that "the coefficient of multiple linear correlation is represented by R1 and it is common
to add subscripts designating the variables involved. Thus R1.234 would represent the
coefficient of multiple linear correlation between X1 on the one hand, X2, X3 and X4 on
the other. The subscript of the dependent variable is always to the left of the point".
The coefficient of multiple correlation for r12, r13 and r23 can be expressed as follows:

Coefficient of multiple correlations for R1.23 is the same as R1.32. A coefficient of multiple
correlation lies between 0 and 1. If the coefficient of multiple correlations is 1, it shows
that the correlation is perfect. If it is 0, it shows that there is no linear relationship
between the variables. The coefficients of multiple correlation are always positive in sign
and range from +1 to 0.
Coefficient of multiple determinations can be obtained by squaring R1.23.
Multiple correlation analysis measures the relationship between the given variables. In
this analysis the degree of association between one variable considered as the dependent
variable and a group of other variables considered as the independent variables.
SUMMARY
This chapter outlined the significance in measuring the relationship. This chapter
discuss the factors that affecting correlation. The different applications of correlation
have been dealt in detail.
KEY WORDS

Measures of Relationship
Correlation
Simple correlation
Partial correlation
Multiple correlation
Regression
Simple regression
Multiple regressions
Association of Attributes
Scatter Diagram Method
Graphic Method

Karl Pearson's Coefficient of Correlation


Concurrent Deviation Method
Method of Least Squares Karl Pearson's Coefficient of Correlation

REVIEW QUESTIONS
1. What are the different measures and their significance in measuring Relationship?
2. Discuss the factors affecting Correlation.
3. What are the applications of Correlation?
4. Discuss in detail on different types of Correlation.
REFERENCE BOOKS
1. Robert Ferber, Marketing research, New York: McGraw Hill Inc., 1976.
2. Chaturvedhi, J.C., Mathematical Statistics, Agra: Nok Jhonk Karyalaya, 1953.
3. Emony, C. William, Business Research Methods, Illinois, Irwin, Homewood. 1976.

- End of Chapter LESSON 18


TABULATION OF DATA

STRUCTURE

Table
Relations frequency table
Cross tabulation and stub-and-banner tables
Guideline for cross tabulation

INTRODUCTION
To get meaningful information from the data it is arranged in the tabular form.
Frequency tables, histograms are simple form of tables.

Frequency Tables
Frequency table or frequency distribution is a better way to arrange data. It helps in
compressing data. Though some information is lost, compressed data show a pattern
clearly. For constructing a frequency table, the data are divided into groups of similar
values (class) and then record the number of observations that fall in each group.
Table 1: Frequency table on age-wise classification of respondents

The data of collection days are presented in the following table as a frequency table. The
number of classed can be increased by reducing the size of the class. The choice of class
intervals is mostly guided by practical consideration rather than by rules. Class intervals
are made in such a way that measurements are uniformly distributed over the class and
the interval is not very large. Otherwise, the mid value will either overestimate or
underestimate the measurement.
Relative frequency tables
Frequency is total number of data points that fall within that class. Frequency of each
value can also be expressed as a fraction or percentage of the total number of
observations. Frequencies expressed in percentage terms are known as relative
frequencies. A relative frequency distribution is presented in the table below.
Table 2: Relative frequency table on occupation-wise classification of
respondents

It may be observed that the sum of all relative frequencies is 1.00 or 100% because
frequency of each class has been expressed as a percentage of the total data.

Cumulative frequency tables


Frequency or one-way tables represent the simplest method for analyzing categorical
data. They are often used as one of the exploratory procedures to review how different
categories of values are distributed in the sample.
For example, in a survey of spectator interest in different sports, we could summarize
the respondents' interest in watching football in a frequency table as follows:
Table 3: Cumulative frequency table on Statistics about football watchers

The table above shows the number, proportion, and cumulative proportion of
respondents who characterized their interest in watching football as either (1) Always
interested, (2) Usually interested, (3) Sometimes interested, or (4) Never interested.
Applications
In practically every research project, a first "look" at the data usually includes frequency
tables. For example, in survey research, frequency tables can show the number of males
and females who participated in the survey, the number of respondents from particular
ethnic and racial backgrounds, and so on. Responses on some labeled attitude
measurement scales (e.g., interest in watching football) can also be nicely summarized
via the frequency table. In medical research, one may tabulate the number of patients
displaying specific symptoms; in industrial research one may tabulate the frequency of
different causes leading to catastrophic failure of products during stress tests (e.g.,
which parts are actually responsible for the complete malfunction of television sets
under extreme temperatures?). Customarily, if a data set includes any categorical data,
then one of the first steps in the data analysis is to compute a frequency table for those
categorical variables.
Cross Tabulation and Stub-and-Banner Tables
Managers and researchers frequently are interested in gaining a better understanding of
the differences that exist between two or more subgroups. Whenever they try to identify
characteristics common to one subgroup but not common to other subgroups, (i.e. they

are trying to explain differences between the subgroups). Cross tables are used to
explain the difference between the subgroups.
Cross tabulation is a combination of two (or more) frequency tables arranged such that
each cell in the resulting table represents a unique combination of specific values of
cross tabulated variables. Thus, cross tabulation allows us to examine frequencies of
observations that belong to specific categories on more than one variable.
By examining these frequencies, we can identify relations between cross tabulated
variables. Only categorical variables or variables with a relatively small number of
different meaningful values should be cross tabulated. Note that in the cases where we
do want to include a continuous variable in a cross tabulation (e.g., income), we can first
recode it into a particular number of distinct ranges (e.g. low, medium, high).
Guidelines for Cross Tabulation
The most commonly used method of data analysis is cross tabulation. The following
guidelines will helpful to design proper cross tabulation,
1. The data should be in categorical form
Cross tabulation is applicable to data 1 which both the dependent and the independent
variables appear in categorical form. There are two types of categorical data.
One type (assume type A) consists of variables that can be measured only in classes or
categories. Like marital status, gender, occupation variables can be measured in
categories not quantifiable (i.e. no measurable number).
Another type (say type B) variables, which can be measured in numbers, such as age,
income. For this type the different categories are associated with quantifiable numbers
that show a progression from smaller values to larger values.
Cross tabulation is used on both types of categorical variables. However when
construction across tabulation is done using type B categorical variables, researchers
find it helpful to use several special steps to make such cross tabulations more effective
analysis tools.
1. If certain variable is believed to be influenced by some other variable, the former can
be considered to be a dependent variable and the later is called as independent
variable.
2. Cross tabulate an important dependent variable with one or more
'explaining' independent variables.
Researchers typically cross tabulate a dependent variable of importance to the objectives
of the research project (such as heavy user versus light user or positive attitude versus
negative attitude) with one or more independent variables that the researchers believe

can help explain the variation observed in the dependent variable. Any two variables can
be used in a cross tabulation so long as they both are in categorical form, and they both
appear to be logically related to one another as dependent and independent variables
consistent with the purpose and objectives of the research project.
3. Show percentage in a cross tabulation
In a cross tabulation researchers typically show the percentage as well as the actual
count s of the number of responses falling into the different cells of the table. The
percentages more effectively reveal the relative sizes of the actual counts associated with
the different cells and make it easier for researchers to visualize the patterns of
differences that exist in the data.
Constructing and Interpreting a Cross Tabulation
After drawing the cross table the interpretations has to be drawn from the table. It
should convey the meaning and findings from the table. In management research
interpretations has more value. From the interpretations and findings managers take
decisions.
2x2 Tables
The simplest form of cross tabulation is the 2 by 2 table where two variables are
"crossed," and each variable has only two distinct values. For example, suppose we
conduct a simple study in which males and females are asked to choose one of two
different brands of soda pop (brand A and brand B); the data file can be arranged like
this:

The resulting cross tabulation could look as follows.

Each cell represents a unique combination of values of the two cross tabulated variables
(row variable Gender and column variable Soda), and the numbers in each cell tell us
how many observations fall into each combination of values. In general, this table shows
us that more females than males chose the soda pop brand A, and that more males than
females chose soda B. Thus, gender and preference for a particular brand of soda may be
related (later we will see how this relationship can be measured).
Marginal Frequencies
The values in the margins of the table are simply one-way (frequency) tables for all
values in the table. They are important in that they help us to evaluate the arrangement
of frequencies in individual columns or rows. For example, the frequencies of 40% and
60% of males and females (respectively) who chose soda A (see the first column of the
above table), would not indicate any relationship between Gender and Soda if the
marginal frequencies for Gender were also 40% and 60%; in that case they would simply
reflect the different proportions of males and females in the study. Thus, the differences
between the distributions of frequencies in individual rows (or columns) and in the
respective margins inform us about the relationship between the cross tabulated
variables.
Column, Row, and total Percentages. The example in the previous paragraph
demonstrates that in order to evaluate relationships between cross tabulated variables
we need to compare the proportions of marginal and individual column or row
frequencies. Such comparisons are easiest to perform when the frequencies are
presented as percentages.
Evaluating the Cross Table
Researchers find it useful to answer the following three questions when evaluating cross
tabulation that appears to explain differences in a dependent variable.
1. Does the cross tabulation show a valid or a spurious relationship?
2. How many independent variables should be used in the cross tabulation?
3. Are the differences seen in the cross tabulation statistically significant, or could they
have occurred by chance due to sampling variation?

Each of this is discussed below.


Does the cross tabulation show a valid explanation?
If it is logical to believe that changes in the independent variables can cause changes in
the dependent variables, then the explanation revealed by the cross tabulation is
thought to be a valid one.
Does the cross tabulation show a valid or a spurious relationship?
An explanation is thought to be a spurious one if the implied relationship between the
dependent and independent variables does not seem to be logical.
Example: family size, income seem appear to be logically related to the household
consumption of certain basic food products. However it may not be logical to relate the
number of automobiles owned with the brand of toothpaste preferred, or to relate the
type of family pet with the occupation of the head of the family. If the independent
variable does not logically have an effect or influence on the dependent variable, the
relationship that a cross tabulation seems to show may not be a valid cause and effect
relationship, and therefore may be a spurious relationship.
How many independent variables should be used?
When cross tabulating an independent variable that seems logically related to the
dependent variable, what should researchers do if the results do not reveal a clear-cut
relationship?
Two possible courses of actions are available.
1. Try another cross tabulation, but this time using one of the other independent
variable hypothesized to be important when the study was designed.
2. A preferred course of action is to introduce each additional independent variable
simultaneously with rather than as an alternative to the first independent variable tried
in the cross tabulation. By doing so it is possible to study the interrelationship between
the dependent variable and two or more independent variables.
SUMMARY
The data can be summarized in the form of table. Cross table given the meaning full
information from the raw data. The way of constructing cross tables and interpreting
and evaluating is very important.
KEY WORDS

Class
Frequency

Relations frequency
Cumulative frequency
Marginal frequency
Cross table

REVIEW QUESTIONS
1. Why do we use cross tables?
2. How do you evaluate the cross table?
3. Define the guidelines for constructing the cross table

- End of Chapter LESSON 19


STATISTICAL SOFTWARE

OBJECTIVES

To learn the application of various statistical packages used for the management
research process
To understand the procedures for performing the tests using SPSS

STRUCTURE

Statistical packages
Statistical analysis using SPSS
t-test, F-test, chi-square test, Anova
Factor analysis

Statistical Packages
The following statistical software packages are widely used:

STATA
SPSS
SAS

STATA

Stata, created in 1985 by Statacorp, is a statistical program used by many businesses and
academic institutions around the world. Most of its users work in research, especially in
the fields of economics, sociology, political science, and epidemiology.
Stata's full range of capabilities includes:

Data management
Statistical analysis
Graphics
Simulations
Custom programming

SPSS
The computer program SPSS (originally, Statistical Package for the Social Sciences) was
released in its first version in 1968, and is among the most widely used programs for
statistical analysis in social science. It is used by market researchers, health researchers,
survey companies, government, education researchers' and others. In addition to
statistical analysis, data management (case selection, file reshaping, creating derived
data) and data documentation are features of the base software.
The many features of SPSS are accessible via pull-down menus (see image) or can be
programmed with a proprietary 4GL "command syntax language". Command syntax
programming has the benefits of reproducibility and handling complex data
manipulations and analyses
Solve business and research problems using SPSS for Windows, a statistical and data
management package for analysts and researchers.
SPSS for Windows provides you with a broad range of capabilities for the entire
analytical process. With SPSS, you can generate decision-making information quickly
using powerful statistics, understand and effectively present the results with highquality tabular and graphical output, and share the results with others using a variety of
reporting methods, including secure Web publishing. Results from the data analysis
enable you to make smarter decisions more quickly by uncovering key facts, patterns,
and trends. An optional server version delivers enterprise-strength scalability,
additional tools, security, and enhanced performance.
SPSS can be used for Windows in a variety of areas, including:

Survey and market research and direct marketing


Academia
Administrative research
Medical, scientific, clinical, and social science research
Planning and forecasting
Quality improvement
Reporting and ad hoc decision making

Enterprise-level analytic application development

In particular, apply SPSS statistics software to gain greater insight into the actions,
attributes, and attitudes of peopleth - e customers, employees, students, or citizens.
Add more functionality as you need it
SPSS for Windows is a modular, tightly integrated, full-featured product line for the
analytical process planning, data collecting, data access, data management and
preparation, data analysis, reporting, and deployment. Using a combination of add-on
modules and stand-alone software that work seamlessly with SPSS Base enhances the
capabilities of this statistics software. The intuitive interface makes it, easy to use - yet it
gives you all of the data management, statistics, and reporting methods you need to do a
wide range of analysis.
Gain unlimited programming capabilities
Dramatically increase the power and capabilities of SPSS for Windows by using the
SPSS Programmability Extension. This feature enables analytic and application
developers to extend the SPSS command syntax language to create procedures and
applications - and perform even the most complex jobs within SPSS. The SPSS
Programmability Extension is included with SPSS Base, making this statistics software
an even more powerful solution.
Maximize market opportunities
The more competitive and challenging the business environment, the more you need
market research. Market research is the systematic and objective gathering, analysis,
and interpretation of information. It helps the organization identify problems and
opportunities and allows for better-informed, lower-risk decisions.
For decades, solutions from SPSS Inc. have added value for those involved in market
research. SPSS solutions support the efficient gathering of market research information
through many different methods, and make it easier to analyze and interpret this
information and provide it to decision makers.
We offer solutions both to companies that specialize in providing market research
services and to organizations that conduct their own market research. SPSS market
research solutions help you:

Understand the market perception of the brand


Conduct effective category management
Confidently develop product features
Perform competitive analysis

With this insight, you or the clients can confidently make decisions about developing
and marketing the products and enhancing the brand.

The heart of SPSS market research solution is dimensions product family. Through
Dimensions, the organization can centralize the creation and fielding of surveys in any
mode and in any language, as well as the analysis and reporting phases of the research.
Dimensions data can be directly accessed using SPSS for Windows, which enables the
analysts to use SPSS' advanced statistical and graphing capabilities to explore the survey
data. Add-on modules and integrated stand-alone products extend SPSS' analytical and
reporting capabilities. For example, analyze responses to open-ended survey questions
with SPSS Text Analysis for Surveys.
Maximize the value the organization receives from its Dimensions data by using an
enterprise feedback management (EFM) solution from SPSS. EFM provides you with a
continuous means of incorporating regular customer insight into the business
operations. Engage with current or prospective customers through targeted feedback
programs or by asking questions during naturally occurring events. Then use the
resulting insights to drive business improvement across the organization. SPSS' EFM
solution also enables you to integrate the survey data with transactional and operational
data, so you gain a more accurate, complete understanding of customer preferences,
motivations, and intentions.
Thanks to the integration among SPSS offerings, you can incorporate insights gained
through survey research in the predictive models created by the data mining tools. You
can then deploy predictive insight and recommendations to people and to automated
systems through any of the predictive analytics applications.
SAS
The SAS System, originally Statistical Analysis System, is an integrated system of
software products provided by SAS Institute that enables the programmer to perform:

Data entry, retrieval, management, and mining


Report writing and graphics
Statistical and mathematical analysis
Business planning, forecasting, and decision support
Operations research and project management
Quality improvement
Applications development
Warehousing (extract, transform, load}
Platform independent and remote computing

In addition, the SAS System integrates with many SAS business solutions that enable
large scale software solutions for areas such as human resource management, financial
management, business intelligence, customer relationship management and more.
Statistical analyses using SPSS
Introduction

This section shows how to perform a number of statistical tests using SPSS. Each section
gives a brief description of the aim of the statistical test, when it is used, an example
showing the SPSS commands and SPSS (often abbreviated} output with a brief
interpretation of the output. In deciding which test is appropriate to use, it is important
to consider the type of variables that you have (i.e., whether your variables are
categorical, ordinal or interval and whether they are normally distributed).
Statistical methods using SPSS
One sample t-test
A one sample t-test allows us to test whether a sample mean (of a normally distributed
interval variable) significantly differs from a hypothesized value. For example, using the
data file, say we wish to test whether the average writing score (write) differs
significantly from 50. We can do this as shown below:
t-test
/testval = 50
/variable = write.

The mean of the variable write for this particular sample of students is 52.775, which is
statistically significantly different from the test value of 50. We would conclude that this
group of students has a significantly higher mean on the writing test than 50.
One sample median test

A one sample median test allows us to test whether a sample median differs significantly
from a hypothesized value. We will use the same variable, write, as we did in the one
sample t-test example above, but we do not need to assume that it is interval and
normally distributed (we only need to assume that write is an ordinal variable).
However, we are unaware of how to perform this test in SPSS.
Binomial test
A one sample binomial test allows us to test whether the proportion of successes on a
two-level categorical dependent variable significantly differs from a hypothesized value.
For example, using the, say we wish to test whether the proportion of females (female)
differs significantly from 50%, i.e., from 0.5. We can do this as shown below:
npar tests
/binomial (.5) = female.

a. Based on Z Approximation.
The results indicate that there is no statistically significant difference (p =0.229). In
other words, the proportion of females in this sample does not significantly differ from
the hypothesized value of 50%.
Chi-square goodness of fit
A chi-square goodness of fit test allows us to test whether the observed proportions for a
categorical variable differ from hypothesized proportions. For example, let's suppose
that we believe that the general population consists of 10% Hispanic, 10% Asian, 10%
African American and 70% White folks. We want to test whether the observed
proportions from our sample differ significantly from these hypothesized proportions.
npar test
/chisquare = race
/expected = 10 10 10 70.

a. 0 cells (0%) have expected count less than 5. The minimum expected cell frequency is
20.0
These results show that racial composition in our sample does not differ significantly
from the hypothesized values that we supplied (chi-square with three degrees of
freedom = 5.029, p = 0.170).
Two independent samples t-test
An independent samples t-test is used when you want to compare the means of a
normally distributed interval dependent variable for two independent groups. For
example, using the, say we wish to test whether the mean for write is the same for males
and females.
t-test groups = female(0 1)
/variables = write.

The results indicate that there is a statistically significant difference between the mean
writing score for males and females (t = -3.734, p = 0.000). In other words, females
have a statistically significantly higher mean score on writing (54.99) than males
(50.12).
Chi-square test
A chi-square test is used when you want to see if there is a relationship between two
categorical variables. In SPSS, the chi2 option is used with the tabulate command to
obtain the test statistic and its associated p-value. Let's see if there is a relationship
between the type of school attended (schtyp) and students gender (female).
Remember that the chi-square test assumes that the expected value for each cell is five
or higher. This assumption is easily met in the examples below. However, if this
assumption' is not met in your data, please see the section on Fisher's exact test below.

a. Computed only for a 2x2 table


b. 0 cells (.0%) have expected count less than 5. The minimum expected count is 14.56.

These results indicate that there is no statistically significant relationship between the
type of school attended and gender (chi-square with one degree of freedom = 0.047, p =
0.828).
Let's look at another example, this time looking at the linear relationship between
gender (female) and socio-economic status (ses). The point of this example is that one
(or both) variables may have more than two levels, and that the variables do not have to
have the same number of levels. In this example, female has two levels (male and
female) and ses has three levels (low, medium and high).

a. 0 cells (.0%) have expected count less than 5. The minimum expected count is 21.39.

Again we find that there is no statistically significant relationship between the variables
(chi-square with two degrees of freedom = 4.577, p = 0.101).
Fisher's exact test
The Fisher's exact test is used when you want to conduct a chi-square test but one or
more of your cells have an expected frequency of five or less. Remember that the chisquare test assumes that each cell has an expected frequency of five or more, but the
Fisher's exact test has no such assumption and can be used regardless of how small the
expected frequency is. In SPSS unless you have the SPSS Exact Test Module, you can
only perform a Fisher's exact test on a 2x2 table, and these results are presented by
default. Please see the results from the chi squared example above.
One-way ANOVA

A one-way analysis of variance (ANOVA) is used when you have a categorical


independent variable (with two or more categories) and a normally distributed interval
dependent variable and you wish to test for differences in the means of the dependent
variable broken down by the levels of the independent variable. For example, using the
data file, say we wish to test whether the mean of write differs between the three
program types (prog).

The mean of the dependent variable differs significantly among the levels of program
type. However, we do not know if the difference is between only two of the levels or all
three of the levels. (The F test for the Model is the same as the F test for prog because
prog was the only variable entered into the model. If other variables had also been
entered, the F test for the Model would have been different from prog). To see the
mean of write or each level of program type

From this we can see that the students in the academic program have the highest mean
writing score, while students in the vocational program have the lowest.
Discriminant analysis
Discriminant analysis is used when you have one or more normally distributed interval
independent variables and a categorical dependent variable It is a multivariate
technique that considers the latent dimensions in the independent variables for
predicting group membership in the categorical dependent variable.

Pooled within-groups correlations between discriminating variables and standardized


canonical discriminant functions. Variables ordered by absolute size of correlation
within function.
*Largest absolute correlation between each variable and any discriminant function

Clearly, the SPSS output for this procedure is quite lengthy, and it is beyond the scope of
this page to explain all of it. However, the main point is that two canonical variables are
identified by the analysis, the first of which seems to be more related to program type
than the second.
Factor analysis
Factor analysis is a form of exploratory multivariate analysis that is used to either
reduce the number of variables in a model or to detect relationships among variables.
All variables involved in the factor analysis need to be interval and are assumed to be
normally distributed. The goal of the analysis is to try to identify factors which underlie
the variables. There may be fewer factors than variables, but there may not be more
factors than variables. For our example, let's suppose that we think that there are some
common factors underlying the various test scores. We will include subcommands for
varimax rotation and a plot of the eigenvalues. We will use a principal components
extraction and will retain two factors. (Using these options will make our results
compatible with those from SAS and Stata and are not necessarily the options that you
will want to use).

Communality (which is the opposite of uniqueness) is the proportion of variance of the


variable (i.e., read) that is accounted for by all of the factors taken together, and a very
low communality can indicate that a variable may not belong with any of the factors.
The screen plot may be useful in determining how many factors to retain. From the
component matrix table, we can see that all five of the test scores load onto the first
factor, while all five tend to load not so heavily on the second factor. The purpose of
rotating the factors is to get the variables to load either very high or very low on each
factor. In this example, because all of the variables loaded onto factor 1 and not on factor
2, the rotation did not aid in the interpretation. Instead, it made the results even more
difficult to interpret.
SUMMARY
The statistical packages applied for the management research process are SPSS, SAS
and STATA. This software makes the research process effective. It reduces the time of
doing analysis. Large data also can be easily analyzed using these softwares.
The chapter also given the detailed procedures and interpretation of SPSS output for
statistical tests.
KEY TERMS

SPSS
SAS
STATA
Dependent variable
Independent variable

REVIEW QUESTIONS
1. Define SPSS
2. What do you mean by STATA and SAS

3. List out the application of statistical software to the market research


4. Define dependent variable
5. Define independent variable
6. Differentiate the relative frequency and cumulative frequency with suitable

- End of Chapter -

LESSON 20
ADVANCED DATA TECHNIQUES

OBJECTIVES
To understand the procedures and applications of following statistical analysis o
o
o
o

Discriminant analysis
ANOVA
Multi dimensional Scaling
Cluster analysis

STRUCTURE

Analysis of variance
Condition for ANOVA
ANOVA model
Discriminant analysis
Factor analysis
Cluster analysis
Ward's method

Analysis of Variance (ANOVA)


Analysis of variance (ANOVA) is used to test hypotheses about differences between two
or more means. The t-test based on the standard error of the difference between two

means can only be used to test differences between two means. When there are more
than two means, it is possible to compare each mean with each other mean using t-tests.
However, conducting multiple t-tests can lead to severe inflation of the Type I error rate.
Analysis of variance can be used to test differences among several means for significance
without increasing the Type I error rate. This chapter covers designs with betweensubject variables. The next chapter covers designs with within-subject variables.
The statistical method for testing the null hypothesis that the means of several
populations are equal is analysis of variance. It uses a single factor, fixed -effects model
to compare the effects of one factor (brands of coffee, varieties if residential housing,
types if retail stores) on a continuous dependent variable. In a fixed effects model, the
levels of the factor are established in advance and the results are not generalizable to
other levels of treatment.
Consider a hypothetical experiment on the effect of the intensity of distracting
background noise on reading comprehension. Subjects were randomly assigned to one
of three groups. Subjects in Group 1 were given 30 minutes to read a story without any
background noise. Subjects in Group 2 read the story with moderate background noise,
and subjects in Group 3 read the story in the presence of loud background noise.
The first question the experimenter was interested in was whether background noise has
any effect at all. That is, whether the null hypothesis: 1 = 2 = 3 is true where 1 is the
population mean for the "no noise" condition, 2 is the population mean for the
"moderate noise" condition, and 3 is the population mean for the "loud noise"
condition. The experimental design therefore has one factor (noise intensity) and this
factor has three levels: no noise, moderate noise, and loud noise.
Analysis of variance can be used to provide a significance test of the null hypothesis that
these three population means are equal. If the test is significant, then the null
hypothesis can be rejected and it can be concluded that background noise has an effect.
In a one-factor between subjects ANOVA, the letter "a" is used to indicate the number of
levels of the factor (a = 3 for the noise intensity example). The number of subjects
assigned to condition 1 is designated as n1; the number of subjects assigned to condition
2 is designated by n2, etc.
If the sample size is the same for all of the treatment groups, then the letter "n" (without
a subscript) is used to indicate the number of subjects in each group. The total number
of subjects across all groups is indicated by "N". If the sample sizes are equal, then N =
(a)(n); otherwise,
N = n1 + n2 + ... + na
Some experiments have more than one between-subjects factor. For instance, consider a
hypothetical experiment in which two age groups (8-year olds and 12-year olds) are
asked to perform a task either with or without distracting background noise. The two
factors are age and distraction.

Assumptions
Analysis of variance assumes normal distributions and homogeneity of variance.
Therefore, in a one-factor ANOVA, it is assumed that each of the populations is
normally distributed with the same variance (2). In between-subjects analyses, it is
assumed that each score is sampled randomly and independently. Research has shown
that ANOVA is "robust" to violations of its assumptions.
This means that the probability values computed in an ANOVA are satisfactorily
accurate even if the assumptions are violated. Moreover, ANOVA tends to be
conservative when its assumptions are violated. This means that although power is
decreased, the probability of a Type I error is as low or lower than it would be if its
assumptions were met. There are exceptions to this rule. For example, a combination of
unequal sample sizes and a violation of the assumption of homogeneity of variance can
lead to an inflated Type I error rate.
Conditions for ANOVA
1. The sample must be randomly selected from normal populations
2. The populations should have equal variances
3. The distance from one value to its group's mean should be independent of the
distances of other values to that mean (independence of error).
4. Minor variations from normality and equal variances are tolerable. Nevertheless, the
analyst should check the assumptions with the diagnostic techniques.
Analysis of variance breaks down or partitions total variability into component parts.
Unlike the 't' test, which uses the sample standard deviations, ANOVA uses squared
deviations of the variance so computation of distances of the individual data points from
their own mean or from the grand mean can be summed.
In ANOVA model, each group has its own mean and values that deviate from that mean.
Similarly, all the data points from all of the groups produce an overall grand mean. The
total deviation is the sum of the squared differences between each data point and the
overall grand mean.
The total deviation of any particular data point may be partitioned into between-groups
variance and within-group variance. The between-group variance represents the effect
of the treatment or factor. The difference of between-groups means imply that each
group was treated differently. The treatment will appear as deviations of the sample
means from the grand mean. Even if this were not so, there would still be some natural
variability among subjects and some variability attributable to sampling. The withingroups variance describes the deviations of the data points within each group from the
sample mean. This results from variability among subjects and from random variation.
It is often called error.

When the variability attributable to the treatment exceeds the variability arising from
error and random fluctuations, the viability of the null hypothesis begins to diminish.
And this is exactly the way the test static for analysis of variance works.
The test statistics for ANOVA is the F ratio. It compares the variance from the last two
sources:

To compute the F ratio, the sum of the squared deviations for the numerator and
denominator are divided by their respective degrees of freedom. By dividing, computing
the variances as an average or mean, thus the term mean square. The number of degrees
of freedom for the numerator, the mean square between groups, is one less than the
number of groups (k-1). The degree of freedom for the denominator, the mean square
within groups, is the total number of observations minus the number of groups (n-k).
If the null hypothesis is true, there should be no difference between the populations, and
the ratio should be close to 1. If the population means are not equal, the numerator
should manifest this difference. The F ratio should be greater than 1. The f distribution
determines the size of ratio necessary to reject the null hypothesis for a particular
sample size and level of significance.
ANOVA model
To illustrate reports one way ANOVA, consider the following hypothetical example. To
find out the number one best business school in India, 20 business magnets were
randomly selected and asked to rate the top 3 B-schools. The ratings are given below:
Data

Let's apply the one way ANOVA test on this example.


Step 1: null hypothesis
H0: A1 = A2 = A3
HA: A1 A2 A3
Step 2: statistical test
The F test is chosen because the example has independent samples, accept the
assumptions of analysis of variance and have interval data.
Step 3: significance level
Let a = 0.05 and degree of freedom = [numerator (k-1) = (3-1) = 2], [denominator (n-k)
= (60-3) = 57] = 2, 57

Step 4: calculated value


F = Mean Square between / Mean square within
F= 5822.017/ 205.695 = 28.304 degree of freedom (2, 57)
Step 5: critical test value
From the F-distribution table with degree of freedom (2, 57), = 0.05, the critical value
is 3.16.
Step 6: decision
Since the calculated value is greater than the critical value (28.3 > 3.16), the null
hypothesis is rejected. The conclusion is: there is statistically significant difference
between two or more pairs of means. The following table shows that the p value equals
0.0001. Since the p value (0.0001) is less than the significance level (0.05), this is the
second method for rejecting the null hypothesis.
The ANOVA model summary given in the following table is the standard way of
summarizing the results of analysis of variance. It contains the sources of variation,
degrees of freedom, sum of squares, mean squares and calculated F value. The
probability of rejecting the null hypothesis is computed up to 100 percent -- that is,
the probability value column reports the exact significance for the F ratio being tested.

S = Significantly different at this level. Significance level: 0.05

All data are hypothetical


Figures on One-way analysis of variance plots

Discriminant analysis

Discriminant function analysis is used to determine which variables discriminate


between two or more naturally occurring groups. For example, an educational
researcher may want to investigate which variables discriminate between high school
graduates who decide (1) to go to college, (2) to attend a trade or professional school, or
(3) to seek no further training or education. For that purpose the researcher could
collect data on numerous variables prior to students graduation. After graduation, most
students will naturally fall into one of the three categories. Discriminant Analysis could
then be used to determine which variable(s) are the best predictors of students
subsequent educational choice.
A medical researcher may record different variables relating to patients backgrounds in
order to learn which variables best predict whether a patient is likely to recover
completely (group 1), partially (group 2), or not at all (group 3). A biologist could record
different characteristics of similar types (groups) of flowers, and then perform a
Discriminant function analysis to determine the set of characteristics that allows for the
best discrimination between the types.
Computational Approach
Let us consider a simple example. Suppose we measure height in a random sample of 50
males and 50 females. Females are, on the average, not as tall as males, and this
difference will be reflected in the difference in means (for the variable Height).
Therefore, variable height allows us to discriminate between males and females with a
better than chance probability: if a person is tall, it is likely to be a male, if a person is
short, it is likely to be a female.
We can generalize this reasoning to groups and variables that are less "trivial." For
example, suppose we have two groups of high school graduates: Those who choose to
attend college after graduation and those who do not. We could have measured students'
stated intention to continue on to college one year prior to graduation. If the means for
the two groups (those who actually went to college and those who did not) are different,
then we can say that intention to attend college as stated one year prior to graduation
allows us to discriminate between those who are and are not college bound (and this
information may be used by career counselors to provide the appropriate guidance to
the respective students).
To summarize the discussion so far, the basic idea underlying discriminant function
analysis is to determine whether groups differ with regard to the mean of a variable, and
then to use that variable to predict group membership (e.g., of new cases).
Stepwise Discriminant Analysis
Probably the most common application of Discriminant function analysis is to include
many measures in the study, in order to determine the ones that discriminate between
groups. For example, an educational researcher interested in predicting high school
graduates' choices for further education would probably include as many measures of

personality, achievement motivation, academic performance, etc. as possible in order to


learn which one(s) offer the best prediction.
Model. Put another way, we want to build a "model" of how we can best predict to
which group a case belongs. In the following discussion wc will use the term "in the
model" in order to refer to variables that are included in the prediction of group
membership, and we will refer to variables as being "not in the model" if they are not
included.
Forward stepwise analysis. In stepwise discriminant function analysis, a model of
discrimination is built step-by-step. Specifically, at each step all variables are reviewed
and evaluated to determine which one will contribute most to the discrimination
between groups. That variable will then be included in the model, and the process starts
again.
Backward stepwise analysis. One can also step backwards; in that case all
variables are included in the model and then, at each step, the variable that contributes
least to the prediction of group membership is eliminated. Thus, as the result of a
successful discriminant function analysis, one would only keep the "important"
variables in the model, that is, those variables that contribute the most to the
discrimination between groups.
F to enter, F to remove. The stepwise procedure is "guided" by the respective F to
enter and F to remove values. The F value for a variable indicates its statistical
significance in the discrimination between groups, that is, it is a measure of the extent to
which a variable makes a unique contribution to the prediction of group membership.
Capitalizing on chance. A common misinterpretation of the results of stepwise
discriminant analysis is to take statistical significance levels at face value. By nature, the
stepwise procedures will capitalize on chance because they "pick and choose" the
variables to be included in the model so as to yield maximum discrimination. Thus,
when using the stepwise approach the researcher should be aware that the significance
levels do not reflect the true alpha error rate, that is, the probability of erroneously
rejecting H0 (the null hypothesis that there is no discrimination between groups).
Multi Dimensional Scaling
Multidimensional scaling (MDS) is a set of related statistical techniques often used in
data visualisation for exploring similarities or dissimilarities in data. An MDS algorithm
starts with a matrix of item-item similarities, and then assigns a location of each item in
a low-dimensional space, suitable for graphing or 3D visualisation.
Categorization of MDS
MDS algorithms fall into a taxonomy, depending on the meaning of the input matrix:

Classical multidimensional scaling also often called Metric multidimensional


scaling -- assumes the input matrix is just an item-item distance matrix. Analogous to
Principal components analysis, an eigenvector problem is solved to find the locations
that minimize distortions to the distance matrix. Its goal is to find a Euclidean distance
approximating a given distance. It can be generalized to handle 3-way distance
problems (the generalization is known as DISTATIS).
Metric multidimensional scaling -- A superset of classical MDS that assumes a
known parametric relationship between the elements of the item-to-item dissimilarity
matrix and the Euclidean distance between the items.
Generalized multidimensional scaling (GMDS) -- A superset of metric MDS
that allows for the target distances to be non-Euclidean.
Non-metric multidimensional scaling -- In contrast to metric MDS, non-metric
MDS both finds a non-parametric monotonic relationship between the dissimilarities in
the item-item matrix and the Euclidean distance between items, and the location of each
item in the low-dimensional space. The relationship is typically found using isotonic
regression.
Multidimensional Scaling Procedure
There are several steps in conducting MDS research:
1. Formulating the problem - What brands do you want to compare? How many
brands do you want to compare? More than 20 would be cumbersome. Less than 8 (4
pairs) will not give valid results. What purpose is the study to be used for?
2. Obtaining Input Data - Respondents are asked a series of questions. For each
product pair they are asked to rate similarity (usually on a 7 point Likert scale from very
similar to very dissimilar). The first question could be for Coke/Pepsi for example, the
next for Coke/Hires rootbeer, the next for Pepsi/Dr Pepper, the next for Dr
Pepper/Hires rootbeer, etc. The number of questions is a function of the number of
brands and can be calculated as Q = N (N - 1) / 2 where Q is the number of questions
and N is the number of brands. This approach is referred to as the "Perception data:
direct approach". There are two other approaches. There is the "Perception data:
derived approach" in which products are decomposed into attributes which are rated on
a semantic differential scale. The other is the "Preference data approach" in which
respondents are asked their preference rather than similarity.
3. Running the MDS statistical program - Software for running the procedure is
available in most of the better statistical applications programs. Often there is a choice
between Metric MDS (which deals with interval or ratio level data), and Non-metric
MDS (which deals with ordinal data). The researchers must decide on the number of
dimensions they want the computer to create. The more dimensions, the better the
statistical fit, but the more difficult it is to interpret the results.

4. Mapping the results and defining the dimensions - The statistical program
(or a related module) will map the results. The map will plot each product (usually in
two dimensional space). The proximity of products to each other indicates either how
similar they are or how preferred they are, depending on which approach was used. The
dimensions must be labelled by the researcher. This requires subjective judgment and is
often very challenging. The results must be interpreted.
5. Test the results for reliability and validity - Compute R squared to determine
what proportion of variance of the scaled data can be accounted for by the MDS
procedure. An R-square of 0.6 is considered the minimum acceptable level. Other
possible tests are Kruskal's Stress, split data tests, data stability tests (i.e. eliminating
one brand), and test-retest reliability.
Input Data
The input to MDS is a square, symmetric 1-mode matrix indicating relationships among
a set of items. By convention, such matrices are categorized as either similarities or
dissimilarities, which are opposite poles of the same continuum. A matrix is a similarity
matrix if larger numbers indicate more similarity between items, rather than less. A
matrix is a dissimilarity matrix if larger numbers indicate less similarity. The distinction
is somewhat misleading, however, because similarity is not the only relationship among
items that can be measured and analyzed using MDS. Hence, many input matrices are
neither similarities nor dissimilarities.
However, the distinction is still used as a means of indicating whether larger numbers in
the input data should mean that a given pair of items should be placed near each other
on the map, or far apart. Calling the data "similarities" indicates a negative or
descending relationship between input values and corresponding map distances, while
calling the data "dissimilarities" or "distances" indicates a positive or ascending
relationship.
A typical example of an input matrix is the aggregate proximity matrix derived from a
pile-sort task. Each cell xij of such a matrix records the number (or proportion) of
respondents who placed items i and j into the same pile. It is assumed that the number
of respondents placing two items into the same pile is an indicator of the degree to
which they are similar. An MDS map of such data would put items close together which
were often sorted into the same piles.
Another typical example of an input matrix is a matrix of correlations among variables.
Treating these data as similarities (as one normally would), would cause the MDS
program to put variables with high positive correlations near each other, and variables
with strong negative correlations far apart.
Another type of input matrix is a flow matrix. For example, a dataset might consist of
the number of business transactions occurring during a given period between a set of
corporations. Running this data through MDS might reveal clusters of corporations that
whose members trade more heavily with one another than other than with outsiders.

Although technically neither similarities nor dissimilarities, these data should be


classified as similarities in order to have companies who trade heavily with each other
show up close to each other on the map.
Dimensionality
Normally, MDS is used to provide a visual representation of a complex set of
relationships that can be scanned at a glance. Since maps on paper are two-dimensional
objects, this translates technically to finding an optimal configuration of points in 2dimensional space. However, the best possible configuration in two dimensions may be
a very poor, highly distorted, representation of your data. If so, this will be reflected in a
high stress value. When this happens, you have two choices: you can either abandon
MDS as a method of representing your data, or you can increase the number of
dimensions.
There are two difficulties with increasing the number of dimensions:
The first is that even 3 dimensions are difficult to display on paper and are
significantly more difficult to comprehend. Four or more dimensions render MDS
virtually useless as a method of making complex data more accessible to the
human mind.
The second problem is that with increasing dimensions, you must estimate an
increasing number of parameters to obtain a decreasing improvement in stress.
The result is model of the data that is nearly as complex as the data itself.
On the other hand, there are some applications of MDS for which high dimensionality
are not a problem. For instance, MDS can be viewed as a mathematical operation that
converts an item-by-item matrix into an item-by-variable matrix. Suppose, for example,
that you have a person-by-person matrix of similarities in attitudes. You would like to
explain the pattern of similarities in terms of simple personal characteristics such as
age, sex, income and education. The trouble is that these two kinds of data are not
conformable. The person-by-person matrix in particular is not the sort of data you can
use in a regression to predict age (or vice-versa). However, if you run the data through
MDS (using very high dimensionality in order to achieve perfect stress), you can create a
person-by-dimension matrix which is similar to the person-by-demographics matrix
that you are trying to compare it to.
MDS and Factor Analysis
Even though there are similarities in the type of research questions to which these two
procedures can be applied, MDS and factor analysis are fundamentally different
methods. Factor analysis requires that the underlying data are distributed as
multivariate normal, and that the relationships are linear. MDS imposes no such
restrictions. As long as the rank-ordering of distances (or similarities) in the matrix is
meaningful, MDS can be used. In terms of resultant differences, factor analysis tends to
extract more factors (dimensions) than MDS; as a result, MDS often yields more readily,

interpretable solutions. Most importantly, however, MDS can be applied to any kind of
distances or similarities, while factor analysis requires us to first compute a correlation
matrix. MDS can be based on subjects' direct assessment of similarities between stimuli,
while factor analysis requires subjects to rate those stimuli on some list of attributes (for
which the factor analysis is performed).
In summary, MDS methods are applicable to a wide variety of research designs because
distance measures can be obtained in any number of ways.
Applications
Marketing
In marketing, MDS is a statistical technique for taking the preferences and perceptions
of respondents and representing them on a visual grid. These grids, called perceptual
maps, are usually two-dimensional, but they can represent more than two.
Potential customers are asked to compare pairs of products and make judgments about
their similarity. Whereas other techniques obtain underlying dimensions from
responses to product attributes identified by the researcher, MDS obtains the
underlying dimensions from respondents' judgments about the similarity of products.
This is an important advantage. It does not depend on researchers' judgments. It does
not require a list of attributes to be shown to the respondents. The underlying
dimensions come from respondents judgments about pairs of products. Because of
these advantages, MDS is the most common technique used in perceptual mapping.
The "beauty" of MDS is that we can analyze any kind of distance or similarity matrix.
These similarities can represent people's ratings of similarities between objects, the
percent agreement between judges, the number of times a subjects fails to discriminate
between stimuli, etc. For example, MDS methods used to be very popular in
psychological research on person perception where similarities between trait descriptors
were analyzed to uncover the underlying dimensionality of people's perceptions of traits
(see, for example Rosenberg, 1977). They are also very popular in marketing research, in
order to detect the number and nature of dimensions underlying the perceptions of
different brands or products & Carmone, 1970).
In general, MDS methods allow the researcher to ask relatively unobtrusive questions
("how similar is brand A to brand B") and to derive from those questions underlying
dimensions without the respondents ever knowing what is the researcher's real interest.
Cluster Analysis
The term cluster analysis (first used by Tryon, 1939) encompasses a number of different
algorithms and methods for grouping objects of similar kind into respective categories.
A general question facing researchers in many areas of inquiry is how to organize
observed data into meaningful structures, that is, to develop taxonomies. In other words
cluster analysis is an exploratory data analysis tool which aims at sorting different

objects into groups in a way that the degree of association between two objects is
maximal if they belong to the same group and minimal otherwise. Given the above,
cluster analysis can be used to discover structures in data without providing an
explanation/interpretation. In other words, cluster analysis simply discovers structures
in data without explaining why they exist.
We deal with clustering in almost every aspect of daily life. For example, a group of
diners sharing the same table in a restaurant may be regarded as a cluster of people. In
food stores items of similar nature, such as different types of meat or vegetables are
displayed in the same or nearby locations. There is countless number of examples in
which clustering plays an important role. For instance, biologists have to organize the
different species of animals before a meaningful description of the differences between
animals is possible. According to the modern system employed in biology, man belongs
to the primates, the mammals, the amniotes, the vertebrates, and the animals. Note how
in this classification, the higher the level of aggregation the less similar are the members
in the respective class. Man has more in common with all other primates (e.g., apes)
than it does with the more "distant" members of the mammals (e.g., dogs), etc. For a
review of the general categories of cluster analysis methods, see Joining (Tree
Clustering), Two-way Joining (Block Clustering), and k-Means Clustering.
Cluster Analysis (CA) is a classification method that is used to arrange a set of cases into
clusters. The aim is to establish a set of clusters such that cases within a cluster are more
similar to each other than they are to cases in other clusters.
Cluster analysis is an exploratory data analysis tool for solving classification problems.
Its object is to sort cases (people, things, events, etc) into groups, or clusters, so that the
degree of association is strong between members of the same cluster and weak between
members of different clusters. Each cluster thus describes, in terms of the data collected,
the class to which its members belong; and this description may be abstracted through
use from the particular to the general class or type.
Cluster analysis is thus a tool of discovery. It may reveal associations and structure in
data which, though not previously evident, nevertheless are sensible and useful once
found. The results of cluster analysis may contribute to the definition of a formal
classification scheme, such as taxonomy for related animals, insects or plants; or suggest
statistical models with which to describe populations; or indicate rules for assigning new
cases to classes for identification and diagnostic purposes; or provide measures of
definition, size and change in what previously were only broad concepts; or find
exemplars to represent classes.
Whatever business you're in, the chances are that sooner or later you will run into a
classification problem. Cluster analysis might provide the methodology to help you
solve.
Procedure for Cluster Analysis

1. Formulate the problem - select the variables that you wish to apply the clustering
technique to
2. Select a distance measure - various ways of computing distance:

Squared Euclidean distance - the square root of the sum of the squared
differences in value for each variable
Manhattan distance - the sum of the absolute differences in value for any variable
Chebychev distance - the maximum absolute difference in values for any variable

3. Select a clustering procedure (see below)


4. Decide on the number of clusters
5. Map and interpret clusters - draw conclusions - illustrative techniques like perceptual
maps, icicle plots, and dendrograms are useful
6. Assess reliability and validity - various methods:

repeat analysis but use different distance measure


repeat analysis but use different clustering technique
split the data randomly into two halves and analyze each part separately
repeat analysis several times, deleting one variable each time
repeat analysis several times, using a different order each time

Figure showing three pairs of clusters:


(AB)
(DE)
(FG)

Beyond these we can see that (AB) & (C) and (DE) are more similar to each other than to
(FG).
Hence we could construct the following dendrogram (hierarchical classification).

Note that the clusters are joined (fused) at increasing levels of 'dissimilarity'.
The actual measure of dissimilarity will depend upon the method used. It may be a
similarity measure or a distance measure. Distances between points can be calculated by
using an extension of Pythagorus (these are euclidean distances). These measures of
'dissimilarity' can be extended to more than 2 variables (dimensions) without difficulty.
Clustering Algorithms
Having selected how we will measure similarity (the distance measure) we must now
choose the clustering algorithm, i.e. the rules which govern between which points
distances are measured to determine cluster membership. There are many methods
available, the criteria used differ and hence different classifications may be obtained for
the same data. This is important since it tells us that although cluster analysis may
provide an objective method for the clustering of cases there can be subjectivity in the
choice of method. Five algorithms, available within SPSS, are described.
1.
2.
3.
4.
5.

Average Linkage Clustering


Complete Linkage Clustering
Single Linkage Clustering
Within Groups Clustering
Ward's Method

1. Average Linkage Clustering


The dissimilarity between clusters is calculated using cluster average values; of course
there are many ways of calculating an average. The most common (and recommended if
there is no reason for using other methods) is UPGMA - Unweighted Pair Groups

Method Average. SPSS also provides two other methods based on averages,
CENTROID and MEDIAN. Centroid or UPGMC (Unweighted Pair Groups Method
Centroid) uses the group centroid as the average. The centroid is defined as the centre of
a cloud of points. A problem with the centroid method is that some switching and
reversal may take place, for example as the agglomeration proceeds some cases may
need to be switched from their original clusters.
2. Complete Linkage Clustering (Maximum or furthest-neighbour method):
The dissimilarity between two groups is equal to the greatest dissimilarity between a
member of cluster i and a member of cluster j. This method tends to produce very tight
clusters of similar cases.
3. Single Linkage Clustering (Minimum or nearest-neighbour method): The
dissimilarities between two clusters is the minimum dissimilarity between members
of the two clusters. This method produces long chains which form loose, straggly
clusters. This method has been widely used in numerical taxonomy.

4. Within Groups Clustering


This is similar to UPGMA except clusters are fused so that within cluster variance is
minimized. This tends to produce tighter clusters than the UPGMA method.
5. Ward's Method
Cluster membership is assessed by calculating the total sum of squared deviations from
the mean of a cluster. The criterion for fusion is that it should produce the smallest
possible increase in the error sum of squares.
Cluster analysis is the statistical method of partitioning a sample into homogeneous
classes to produce an operational classification. Such a classification may help:

Formulate hypotheses concerning the origin of the sample, e.g. in evolution


studies
Describe a sample in terms of a typology, e.g. for market analysis or
administrative purposes

Predict the future behaviour of population types. e.g. in modeling economic


prospects for different industry sectors
Optimize functional processes, e.g. business site locations or product design
Assist in identification, e.g. in diagnosing diseases
Measure the different effects of treatments on classes within the population, e.g.
with analysis of variance

SUMMARY
The complete process of generalized hierarchical clustering can be summarized as
follows:
1. Calculate the distance between all initial clusters. In most analyses initial clusters will
be made up of individual cases.
2. Fuse the two most similar clusters and recalculate the distances.
3. Repeat step 2 until all cases are in one cluster.
One of the biggest problems with this Cluster Analysis is identifying the optimum
number of clusters. As the fusion process continues increasingly dissimilar clusters
must be fused, i.e. the classification becomes increasingly artificial. Deciding upon the
optimum number of clusters is largely subjective, although looking at a graph of the
level of similarity at fusion versus number of clusters may help. There will be sudden
jumps in the level of similarity as dissimilar groups are fused.
KEY TERMS

SPSS
Tabulation
Cross-tabulation
ANOVA
Discriminant analysis
Factor analysis
Conjoint analysis
MDS
Cluster analysis

REVIEW QUESTIONS
1. What do you mean by cross-tabulation?
2. Write short notes on statistical packages
3. Explain the step wise procedure for doing Discriminant Analysis.
4. Write short notes on ANOVA.

5. Explain the application of Factor analysis in Marketing.


6. What do you mean by conjoint analysis?
7. Explain the procedure of performing Multi Dimensional Scaling.
8. What are the applications of MDS?
9. Describe the different types of cluster analysis.
10. Explain the marketing situations in which the above said tools will be used.

- End of Chapter LESSON 21


FACTOR ANALYSIS

OBJECTIVES

To learn the basic concepts of Factor.


To understand procedures of performing Factor.
To identify the applications of factor.

STRUCTURE

Evaluation of factor analysis


Steps involved in conducting the factor analysis
Process involved in factor analysis
Output of factor analysis
Limitation of factor analysis

INTRODUCTION
Factor analysis is a general name denoting a class of procedures primarily used for data
reduction and summarization. In marketing research, there may be a large number of
variables, most of which are correlated and which must be reduced to a manageable
level. Relationships among sets of many interrelated variables are examined and
represented in terms of a few underlying factors. For example, store image may be
measured by asking respondents to evaluate stores on a series of items on a semantic

differential scale. The item evaluations may then be analyzed to determine the factors
underlying store image.
In analysis of variance, multiple regression, and discriminant analysis, one variable is
considered as the dependent or criterion variable, and the others as independent or
predictor variables. However, no such distinction is made in factor analysis. Rather,
factor analysis is an interdependence technique in that an entire set of
interdependent relationships is examined.
Factor analysis is used in the following circumstances:
- To identify underlying dimensions, or factors that explain the correlations among a set
of variables. For example, a set of lifestyle statements may be used to measure the
psychographic profiles of consumers. These statements may then be factor analyzed to
identify) the underlying psychographic factors, as illustrated in the department store
example.
- To identify a new, smaller set of uncorrelated variables to replace the original set of
correlated variables in subsequent multivariate analysis (regression or discriminant
analysis). For example, the psychographic factors identified may be used as independent
variables to explaining the differences between loyal and non loyal consumers.
- To identify a smaller set of salient variables from a larger set for use in subsequent
multivariate analysis. For example, a few of the original lifestyle statements that
correlate highly with the identified factors may be used as independent variables to
explain the differences between the loyal and non-loyal users.
Definition
Factor analysis is a class of procedures primarily used for data reduction and
summarization. Factors analysis is an interdependence technique in that an entire set of
interdependent relationships is examined. Factors are defined as an underlying
dimension that explains the correlation among a set of variables
Evolution of Factor Analysis
Charles Spearman first used the factor analysis as a technique of indirect measurement.
When they test human personality and intelligence, a set of questions and tests are
developed for this purpose. They believe that a person gives this set of questions and
tests would respond on the basis of some structure that exists in his mind. Thus, his
responses would form a certain pattern. This approach is based on the assumption that
the underlying structure in answering the questions would be the same in the case of
different respondents.
Even though it is in the field of psychology that factor analysis has its beginning, it has
since been applied to problems in different areas including marketing. Its use has

become far more frequent as a result of the introduction Specialized software packages
such as SPSS, SAS.
Application of Factor Analysis

It can be used in market segmentation for identifying the underlying variables on


which to group the customers. New-car buyers might be grouped based on the
relative emphasis place on economy, convenience, performance, comfort, and
luxury. This might result in five segments: economy seekers, convenience
seekers, performance seekers, fort seekers, and luxury seekers.
In product research, factor analysis can be employed to determine the brand
attributes that influence consumer choice. Toothpaste brand-might be evaluated
in terms of protection against cavities, whiteness of teeth, taste, fresh breath, and
price.
In advertising studies, factor analysis can be used to understand the media
consumption habits of the target market. The users of frozen foods may be heavy
viewers of cable TV, see a lot at movies, and listen to country music.
In pricing studies, it can be used to identify the characteristics of price-sensitive
consumers. For example, these consumers might be methodical, economy
minded, and home centered.

It can bring out the hidden or latent dimensions relevant in the relationships among
product preferences. Factor analysis is typically used to study a complex product or
service in order to identify the major characteristics (or factors) considered to be
important by consumers of the product or service Example: Researchers for an
automobile (two wheeler) company may ask a large sample of potential buyers to report
(using rating scales) the extent of their agreement or disagreement with the number of
statements such as "A motor bikes breaks are its most crucial part", "Seats should be
comfortable for two members". Researchers apply factor analysis to such a set of data to
identity, which factors such as "safety", "exterior styling", "economy of operations" are
considered important by potential customers. If this information is available, it can be
used to guide the overall characteristics to be designed into the product or to identify
advertising themes that potential buyers aid consider important.
Steps Involved In Conducting the Factor Analysis:

Statistics Associated with Factor Analysis


The key statistics associated with factor analysis are as follows:
Bartlett's test of sphericity: A test statistic used to examine the hypothesis that the
variables are uncorrelated in the population. In other words, the population correlation
matrix is an identity matrix; each variable correlates perfectly with itself (r = 1) but has
no correlation with the other variables (r = 0).
Correlation matrix: A lower triangle matrix showing the simple correlations, r,
between all possible pairs of variables included in the analysis. The diagonal elements,
which are all I, are usually omitted.
Communality: The amount of variance a variable shares with all the other variables
being considered. This is also the proportion of variance explained by the common
factors.

Eigenvalue: Represents the total variance explained by each factor.


Factor loadings: Simple correlations between the variables and the factors.
Factor loading plot: A plot of the original variables using the factor loadings as
coordinates.
Factor matrix: Contains the factor loadings of all the variables on all the factors
extracted.
Factor scores: Composite scores estimated for each respondent on the derived
factors.
Kaiser-Meyer-Olkin (KMO) measure of sampling adequacy. An index used to
examine the appropriateness of factor analysis. High values (between 0.5 and 1.0)
indicate factor analysis is appropriate. Values below 0.5 imply that factor analysis may
not be appropriate.
Percentage of variance: The percentage of the total variance attributed to each
factor.
Residuals: The differences between the observed correlations, as given in the input
correlation matrix, and the reproduced correlations, as estimated from the factor
matrix.
Scree plot: A plot of the eigenvalues against the number of factors in order of
extraction. We describe the uses of these statistics in the next section, in the context of
the procedure for conducting factor analysis.
Process Involved In Factor Analysis
Factor analysis applies an advanced form of correlation analysis to the responses to
number of statements. The purpose of this analysis is to determine if the responses to
several of statements are highly correlated .If the responses to three or more statements
are highly correlated, it is believed that the statements measure some factor common to
all of them.
The statements in any one set are highly correlated with each other but are not highly
correlated with the statements in any of the other sets.
For each set of highly correlated statements, the researchers use their own judgment to
determine what the single "theme" or "factor" is that ties the statements together in the
minds of the respondents. For example, regarding the automobile study mentioned
above, researchers may find high correlations among the responses to the following
three statements:

i. Mileage per liter should be high; ii. Maintenance cost should be low; iii. Mileage
should be consistent in all types of roads.
The researcher may then make the judgment that agreement with these set of
statements indicates an underlying concern with the factor of "Economy of operation".
Determine the Method of Factor Analysis
Once it has been determined that factor analysis is an appropriate technique for
analyzing data, an appropriate method must be selected. The approach used to derive
the weights factor score coefficients differentiates the various methods of factor analysis.
The two basic approaches are principal components analysis and common factor
analysis. In principal components analysis, the total variance in the data is considered.
The diagonal of correlation matrix consists of unities, and full variance is brought into
the factor matrix. Principal components analysis is recommended when the
primary concern is to determine the minimum number of factors that will account for
maximum variance in the data for use in subsequent multivariate analysis. The factors
are called principal components.
In common factor analysis, the factors are estimated based only on the common
variance. Communalities are inserted in the diagonal of the correlation matrix. This
method is appropriate when the primary concern is to identify the underlying
dimensions and the common variance is of interest. This method is also known as
principal axis factoring.
Other approaches for estimating the common factors are also available. These include
methods of unweighted least squares, generalized least squares, maximum likelihood,
HA method, and image factoring. These methods are complex and are not
recommended for experienced users.
Determine the Number of Factors
It is possible to compute as many principal components as there are variables, but in
doing so, no parsimony is gained. In order to summarize the information contained in
the original variables, a smaller number of factors should be extracted. The question is,
how many? Several procedures have been suggested for determining the number of
factors. These include a priori determination and approaches based on eigenvalues,
scree plot, percentage variance accounted for, split-half reliability, and significance
tests.
A Prior Determination. Sometimes, because of prior knowledge, the researcher
knows many factors to expect and thus can specify the number of factors to be extracted
beforehand. The extraction of factors ceases when the desired number of factors have
been extracted. Most computer programs allow the user to specify the number of
factors, allowing for an easy implementation of this approach.

Determination Based on Eigenvalues. In this approach, only factors with


eigenvalues greater than 1.0 are retained; the other factors are not included in the
model. An eigenvalue sends the amount of variance associated with the factor. Hence,
only factors with a variance greater than 1.0 are included. Factors with variance less
than 1.0 are no better than a single variable, because, due to standardization, each
variable has a variance of 1.0. If the number of variables is less than 20, this approach
will result in a conservative number of factors.
Determination Based on Scree Plot. A scree plot is a plot of the eigenvalues
against the number of factors in order of extraction. The shape of the plot is used to
determine the number of factors. Typically, the plot has a distinct break between the
steep slope of factors with large eigenvalues and a gradual trailing off associated with
the rest of the factors. This gradual trailing off is referred to as the scree. Experimental
evidence indicates that point at which the scree begins denotes the true number of
factors. Generally, the number of factors determined by a scree plot will be one or a few
more than that determined by the eigenvalue criterion.
Determination Based on Percentage of Variance. In this approach the number
of factors extracted is determined so that the cumulative percentage of variance
extracted by the factors reaches a satisfactory level. What level of variance is satisfactory
depends upon the problem. However, it is recommended that the factors extracted
should account for at least o percent of the variance.
Determination Based on Split-Half Reliability. The sample is split in half and
factor analysis is performed on each half. Only factors with high correspondence of
factor loadings across the two subsamples are retained.
Determination Based on Significance Tests. It is possible to determine the
Statistical significance of the separate eigenvalues and retain only those factors that are
statistically, significant. A drawback is that with large samples (size greater than 200),
many factors likely to be statistically significant, although from a practical viewpoint
many of the count for only a small proportion of the total variance.
Illustration
A manufacturer of motorcycles wanted to know which motorcycle characteristics were
considered very important by the customers. The company identified 100 statements
that related to all characteristics of motorcycles that they believed important. 300
potential customers of motorcycles were selected on a probability basis and were asked
to rate the 100 statements, five of which are listed below. They were then asked to report
on a 5-point scale the extent to which they agreed or disagreed with the statement.

Breaks are the most important parts for motorcycles.


Appearance of motorcycle should be masculine
Mileage per liter should be high
Maintenance cost should be low
Mileage should be consistent in all types of roads.

This resulted in a set of data in which each of 300 individuals gave a response to each of
100 statements. For any given statement, some individuals were found to agree strongly,
some were found to disagree slightly, some neither agreed nor disagreed with the
statement, and so on. Thus, for each statement, there was a distribution of 300
responses on a 5-point scale.
Three Important Measures
There are three important measures used in the factor analysis:
1. Variance
2. Standardized scores of an individual's responses
3. Correlation coefficient.
1. Variance
A factor analysis is somewhat like regression analysis in that it tries to "best fit" factors
to a scatter diagram of the data in such a way that the factors explain the variance
associated with the responses to each statement.
2. Standardized Scores of an Individual's Responses
To facilitate comparisons of the responses from such different scales, researchers
standardize all of the answers from all of the respondents on all statements and
questions.

Thus, an individual's standardized score is nothing more than an actual response


measured in terms of the number of standard deviations (+ or -1) it lies away from the
mean. Therefore, each standardized score is likely to be a value somewhere in the range
of +3.00 and -3.00, with +3.00 typically being equated to the "agree very strongly"
response and -3.00 typically being equated to the "disagree very strongly" response.
3. Correlation Coefficient
The third measure used is the correlation coefficient associated with the standardized
scores of the responses to each pair of statements. The matrix of correlation coefficients
is very important part of factor analysis.
The factor analysis searches through a large set of data to locate two or more sets of
statements, which have highly correlated responses. The responses to the statements in

one set will all be highly correlated with each other, but they will also be quite
uncorrelated with the responses to the statements in other sets. Since the different sets
of statements are relatively uncorrelated with each other, a separate and distinct factor
relative to motorcycles is associated with each set.
It is already noted that Variance is one of the three important measures used in factor
analysis with standardized responses to each statement used in the study.
Factor analysis selects one factor at a time using procedures that "best fit" each other to
the data. The first factor selected is one that fits the data in such a way that it explains
more of the variance in the entire set of standardized scores than any other possible
factor. Each factor selected after the first factor must be uncorrelated with the factors
already selected. This process continues until the procedures cannot find additional
factors that significantly reduce the unexplained variance in the standardized scores
Output of Factor Analysis
Here, only following six statements (or variables) for simplicity will be used to explain
the output of a factor analysis
XI - Mileage per liter should be high
X2 = Maintenance cost should be low
X3 = Mileage should be consistent in all types of roads.
X4 = Appearance of motorcycle should be masculine
X5 = multiple colors should be available
X6 = Breaks are the most important parts for motorcycles
The results a factor analysis of these six statements will appear in the form shown m the
following table which can be used to illustrate the three important outputs from a factor
analysis and how they can be of use to researchers.
Table 1: Factor Analysis Output of the Motorcycle Study

Factor Loadings
The top six rows of the table are associated with the six statements listed above. The
table shows that the factor analysis has identified three factors (F1, F2 and F3) and the
first three columns are associated with those factors. For example' the first factor can be
written as
F1= 0.86 x 1+ 0.84 x 2 + 0.68 x 3 + 0.10 x 4 + 0.06 x 5 + 0.12 x 6
The 18 numbers located in the six rows and three columns are called as factor loadings
and they are one of the 3 useful outputs obtained from a factor analysis. As shown in
table, each statement has a factor loading associated with a specific factor and a specific
statement is simply the correlation between that factor and that statements
standardized response scores.
Thus, table 1 shows that factor 1 is highly correlated with the responses to statement 1
(0.86 correlation) and also with the response to statement 2 (0.84 correlation).
Table 1 shows that statements 1 and 2 are not highly correlated (0.12 and 0.18
respectively) with factor 2. Thus a factor loading is a measure of how well the factor
"fits" the standardized responses to Statement.
Naming the Factors
From the table, it is clear that factor F1 is good fit on the data from statements 1, 2 and
3, but a poor fit on the other statements. This indicates that statements 1, 2 and 3 are
probably measuring the same basic attitude or value system; and it is this finding that
provides the researchers with evidence that factor exists.

By using the knowledge on the industry and the contents of statements 1, 2, and 3
researchers from Motorcycle Company subjectively concluded from these results that
"Economy of operation" was the factor that tied these statements together in the minds
of the respondents.
Researchers next want to know if the 300 respondents participating in the study mostly
agreed with or disagreed with statements 1, 2 and 3. To answer this question, the
researchers had to look at the 300 standardized responses to each of the statements 1, 2
and 3. They found that the means of these responses to each of statements 1, 2 and 3.
They found that the means of these responses were +0.97, + 1.32, and +1.18,
respectively, for statements 1, 2 and 3, indicating that most respondents agreed with the
three statements as per above discussion on "standardized scores". Since a majority of
respondents had agreed with these statements, the researchers also concluded that the
factor of "economy of operation" was important in the minds of potential motorcycle
customers.
Table also shows that F2 is good fit on statements 4 and 5 but a poor fit on the other
statements. This factor is clearly measuring something different from statements 1, 2, 3
and 6. Factor F3 is good fit only on statement 6 and so it is clearly measuring something
not being statements 1 to 5. Researchers again subjectively concluded that the factor
underlying statements 4 and 5 was "Comfort" and that statement 6 was related to
"safety".
Fit between Data and Factor
Researcher has to find how well all of the identified factors fit the data obtained from all
of the respondents on any given statement. Communalities for each statement indicate
the proportion of the variance in the responses to the statement which is explained by
the three identified factors.
For example, Three factors explain 0.89 (89%) of the variance in all of the responses to
statement 5, but only 0.54 (54%) of the variance in all of the responses to statement 3. It
shows that three factors explain 75% or more of the variance associated with statements
1, 2, 4, 5 and 6, but only about half of statement 3's variance .Researchers can use these
communalities to make a judgment about for most of the variance associated with each
of the six statements in this example, the three factors fit the data quite well.
How well any given factor fits the data from all of the respondents on all of the
statements can be determined by eigenvalue. There is an eigenvalue associated with
each of the factors.
When a factor's eigenvalue is divided by the number of statements used in factor
analysis the resulting figure is the proportion of the variance in the entire set of
standardized response scores which is explained by that factor i.e. each eigenvalue is
divided by 6 the number of statements. For example factor F1 explains 0.3226 (32%)
of the variance of the standardized response scores from all of the respondents on all six
statements. By adding these figures for the three factors, three factors together explain

0.3226+0.3090+0.1391 = 0.7707 (77.07%) of the variance in the entire set of response


data. This figure can be used as a measure of how well, over all, the identified factors fit
the data. In general, a factor analysis that accounts for 60-70% or more of the total
variance can be considered a good fit to the data.
Limitations
The utility of this technique largely depends to a large extent on the judgment of the
researcher. He has to make number of decisions as to how the factor analysis will come
out. Even with a given set of decisions, different results will emerge from different
groups of respondents, different mixes of data as also different ways of getting data. In
other words, factor analysis is unable to give a unique solution or result.
As any other method of analysis, a factor analysis will be of little use if the appropriate
variable has not been measured, or if the measurements are inaccurate, or if the
relationships in the data are non linear
In the view of ongoing limitations, the exploratory nature of factor analysis becomes
clear. As Thurston mentions, the use of factor analysis should not be made where
fundamental and fruitful concepts are already well formulated and tested. It may be
used especially in those domains where basic and fruitful concepts are essentially
lacking and where crucial experiments have been difficult to conceive.
SUMMARY
This chapter has overview of factor analysis in detail. Factor analysis is used to find
latent variable or factor among observed variables. With factor analysis you can produce
a small number of factors frame large number of variables. The induced factor can also
be used for further analysis.
KEY WORDS

Factor analysis
Communalities
Factor loading
Correlation matrix
Eigen value

IMPORTANT QUESTIONS
1. What are the applications of factor analysis?
2. What is the significance of factor loading in factor analysis?
3. What do you mean by Eigen value?

- End of Chapter LESSON 22


CONJOINT ANALYSIS

STRUCTURE

Conjoint analysis
Basics of conjoint analysis
Steps involved in conjoint analysis
Application of conjoint analysis

Conjoint analysis, also called multi attribute compositional models, is a statistical


technique that originated in mathematical psychology and was developed by marketing
professor Paul Green at the Wharton School of the University of Pennsylvania. Today it
is used in many of the social sciences and applied sciences including marketing, product
management, and operations research. The objective of conjoint analysis is to determine
what combination of a limited number of attributes is most preferred by respondents. It
is used frequently in testing customer acceptance of new product designs and assessing
the appeal of advertisements. It has been used in product positioning, but there are
some problems with this application of the technique. Recently, new alternatives such as
Genetic Algorithms have been used in market research.
The Basics of Conjoint Analysis
The basics of conjoint analysis are easy to understand. It should only take about 20
minutes to introduce this topic so you can appreciate what conjoint analysis has to offer.
In order to understand conjoint analysis, let's look at a simple example. Suppose you
wanted to book an airline flight and you had a choice of spending Rs.400 or Rs.700 for a
ticket. If this were the only consideration then the choice is clear: the lower priced ticket
is preferable. What if the only consideration in booking a flight was sitting in a regular
or extra-wide seat? If seat size was the only consideration then you would probably
prefer an extra-wide seat. Finally, suppose you can take either a direct flight which takes
three hours or a flight that stops once and takes five hours. Virtually everyone would
prefer the direct flight.
Conjoint analysis attempts to determine the relative importance consumers attach to
salient attributes and the utilities they attach to the levels of attributes. This information
is derived from consumers' evaluations of brands, or brand profiles composed of these
attributes and their levels. The respondents are presented with stimuli that consist of
combinations of attribute levels. They are asked to evaluate these stimuli in terms of

their desirability. Conjoint procedures attempt to assign values to the levels of each
attribute, so that the resulting values or utilities attached to the stimuli match, as closely
as possible the input evaluations provided by the respondents. The underlying
assumption is that any set of stimuli, such as products, brands, or stores, is evaluated as
a bundle of attributes.
Conjoint Analysis is a technique that attempts to determine the relative importance
consumers attach to salient attributes and the utilities they attach to the levels of
attributes.
In a real purchase situation, however, consumers do not make choices based on a single
attribute like comfort. Consumers examine a range of features or attributes and then
make judgments or trade-offs to determine their final purchase choice. Conjoint analysis
examines these trade-offs to determine the combination of attributes that will be most
satisfying to the consumer, in other words, by using conjoint analysis a company can
determine the optimal features for their product or service. In addition, conjoint
analysis will identify the best advertising message by identifying the features that are
most important in product choice.
Like multidimensional scaling, conjoint analysis relies on respondents' subjective
evaluations. However, in MDS the stimuli are products or brands. In conjoint analysis,
the stimuli are combinations of attribute levels determined by the researcher. The goal
in MDS is to develop a spatial map depicting the stimuli in a multidimensional
perceptual or preference space. Conjoint analysis, on the other hand, seeks to develop
the part-worth or utility functions describing the utility consumers attach to the levels of
each attribute. The two techniques are complementary.
In sum, the value of conjoint analysis is that it predicts what products or services people
will choose and assesses the weight people give to various factors that underlie their
decisions. As such, it is one of the most powerful, versatile and strategically important
research techniques available.
Statistics and Terms Associated with Conjoint Analysis
The important statistics and terms associated with conjoint analysis include:

Part-worth functions. Also called utility functions, these describe the utility
consumers attach to the levels of each attribute.
Relative importance weights. Indicate which attributes are important in
influencing consumer choice. These weights are estimated.
Attribute levels. Denote the values assumed by the attributes.
Full profiles. Full profiles or complete profiles of brands are constructed in
terms of all the attributes by using the attribute levels specified by the design.
Pair-wise tables. The respondents evaluate two attributes at a time until all
the required pairs of attributes have been evaluated.
Cyclical designs. Designs employed to reduce the number of paired
comparisons.

Fractional factorial designs: Designs employed to reduce the number of


stimulus profiles to be evaluated in the full-profile approach.
Orthogonal arrays. A special class of fractional designs that enable the
efficient estimation of all main effects.
Internal validity: This involves correlations of the predicted evaluations for
the holdout or validation stimuli with those obtained from the respondents.

Conducting Conjoint Analysis


The following chart lists the steps in conjoint analysis. Formulating the problem
involves identifying the salient attributes and their levels. These attributes and levels are
used for constructing the stimuli to be used in a conjoint evaluation task.
Formulate the Problem

Construct the Stimuli

Decide on the Form of Input Data

Select a Conjoint Analysis Procedure

Interpret the Results

Assess Reliability and Validity

Steps involved in conjoint analysis


The basic steps are:

Select features to be tested


Show product feature combinations to potential customers
Respondents rank, rate, or choose between the combinations

Input the data from a representative sample of potential customers into a


statistical software program and choose the conjoint analysis procedure. The
software will produce utility functions for each of the features.
lncorporate the most preferred features into a new product or advertisement.

Any number of algorithms may be used to estimate utility functions. The original
methods were monotonic analysis of variance or linear programming techniques, but
these are largely obsolete in contemporary marketing research practice. Far more
popular are the Hierarchical Bayesian procedures that operate on choice data. These
utility functions indicate the perceived value of the feature and how sensitive consumer
perceptions and preferences are to changes in product features.
A Practical Example of Conjoint Analysis
Conjoint analysis presents choice alternatives between products/services defined by sets
of attributes. This is illustrated by the following choice: would you prefer a flight with
regular seats, that costs Rs.400 and takes 5 hours, or a flight which costs Rs.700, has
extra-wide seats and takes 3 hours?
Extending this, we see that if seat comfort, price and duration are the only relevant
attributes, there are potentially eight flight choices.

Given the above alternatives, product 4 is very likely the most preferred choice, while
product 5 is probably the least preferred product. The preference for the other choices is
determined by what is important to that individual.
Conjoint analysis can be used to determine the relative importance of each attribute,
attribute level, and combinations of attributes. If the most preferable product is not
feasible for some reason (perhaps the airline simply cannot provide extra-wide seats and
a 3 hour arrival time at a price of Rs400) then the conjoint analysis will identify the next
most preferred alternative. If you have other information on travelers, such as
background demographics, you might be able to identify market segments for which

distinct products may be appealing. For example, the business traveller and the vacation
traveller may have very different preferences which could be met by distinct flight
offerings.
You can now see the value of conjoint analysis. Conjoint analysis allows the researcher
to examine the trade-offs that people make in purchasing a product. This allows the
researcher to design products/services that will be most appealing to a specific market.
In addition, because conjoint analysis identifies important attributes, it can be used to
create advertising messages that will be most persuasive.
In evaluating products, consumers will always make trade-offs. A traveller may like the
comfort and arrival time of a particular flight, but reject purchase due to the cost. In this
case, cost has a high utility value. Utility can be defined as a number which represents
the value that consumers place on an attribute. In other words, it represents the relative
"worth" of the attribute. A low utility indicates less value; a high utility indicates more
value.
The following figure presents a list of hypothetical utilities for an individual consumer:

Based on these utilities, we can make the following conclusions:

This consumer places a greater value on a 3 hour flight (the utility is 42) than on
a 5 hour flight (utility is 22).
This consumer does not differ much in the value that he or she places on comfort.
That is, the utilities are quite close (12 vs.15).
This consumer places a much higher value on a price of Rs.400 than a price of
Rs.700.
The preceding example depicts an individual's utilities. Average utilities can be
calculated for all consumers or for specific subgroups of consumers.

These utilities also tell us the extent to which each of these attributes drives the decision
to choose a particular flight. The importance of an attribute can be calculated by
examining the range of utilities (that is, the difference between the lowest and highest
utilities) across all levels of the attribute. That range represents the maximum impact
that the attribute can contribute to a product.
Using the hypothetical utilities presented earlier, we can calculate the relative
importance of each of the three attributes. The range for each attribute is given below:

Duration: Range = 20 (42-22)


Comfort: Range = 3 (15-12)
Cost: Range = 56 (61-5)

These ranges tell us the relative importance of each attribute. Cost is the most important
factor in product purchase as it has the highest range of utility values. Cost is followed in
importance by the duration of the flight. Based on the range and value of the utilities, we
can see that seat comfort is relatively unimportant to this consumer. Therefore,
advertising which emphasizes seat comfort would be ineffective. This person will make
his or her purchase choice based mainly on cost and then on the duration of the flight.
Marketers can use the information from utility values to design products and/or services
which come closest to satisfying important consumer segments. Conjoint analysis will
identify the relative contributions of each feature to the choice process. This technique,
therefore, can be used to identify market opportunities by exploring the potential of
product feature combinations that are not currently available.
Choice Simulations
In addition to providing information on the importance of product features, conjoint
analysis provides the opportunity to conduct computer choice simulations. Choice
simulations reveal consumer preference for specific products defined by the researcher.
In this case, simulations will identify successful and unsuccessful flight packages before
they are introduced to the market!
For example, let's say that the researcher defined three flights as follows:

The conjoint simulation will indicate the percentage of consumers that prefer each of
the three flights. The simulation might show that consumers are willing to travel longer
if they can pay less and are provided a meal. Simulations allow the researcher to
estimate preference, sales and share for new flights before they come to market.

Simulations can be done interactively on a microcomputer to quickly and easily look at


all possible options. The researcher may, for example, want to determine if a price
change of Rs.50, Rs.100, or Rs.150 will influence consumer's choice. Also, conjoint will
let the researcher look at interactions among attributes. For example, consumers may be
willing to pay Rs.50 more for a flight on the condition that they are provided with a hot
meal rather than a snack.
Data Collection
Respondents are shown a set of products, prototypes, mock-ups or pictures. Each
example is similar enough that consumers will see them as close substitutes, but
dissimilar enough that respondents can clearly determine a preference. Each example is
composed of a unique combination of product features. The data may consist of
individual ratings, rank-orders, or preferences among alternative combinations. The
latter is referred to as "choice based conjoint" or "discrete choice analysis".
In order to conduct a conjoint analysis, information must be collected from a sample of
consumers. This data can be conveniently collected in locations such as shopping
centers or by the Internet. In the previous example, data collection could take place at a
booth located in an airport or in the office of a travel agent.
A sample size of 400 is generally sufficient to provide reliable data for consumer
products or services. Data collection involves showing respondents a series of cards that
contain a written description of the product or service. If a consumer product is being
tested then a picture of the product can be included along with a written description. A
typical card examining the business traveller might look like the following:

Readers might be worried at this point about the total number of cards that need to be
rated by a single respondent. Fortunately, we are able to use statistical manipulations to
cut down on the number of cards. In a typical conjoint study, respondents only need to
rate between 10-20 cards.
This data would be input to the conjoint analysis. Utilities can then be calculated and
simulations can be performed to identify which products will be successful and which
should be changed. Price simulations can also be conducted to determine sensitivity of
the consumer to changes in prices.
A wide variety of companies and service organizations have successfully used conjoint
analysis
A conjoint analysis was developed using a number of attributes such as saving on energy
bills, efficiency rating of equipment, safety record of energy source, and dependability of
energy source. The conjoint analysis identified that cost savings and efficiency were the
main reasons for converting appliances to gas. The third most important reason was
cleanliness of energy source. This information was used in marketing campaigns in
order to have the greatest effect.
A natural gas utility used conjoint analysis to evaluate which advertising message would
be most effective in convincing consumers to switch from other energies to natural gas.
Previous research failed to discover customer's specific priorities - it appeared that the
trade-offs that people made were quite subtle.
Advantages

Able to use physical objects


Measures preferences at the individual level
Estimates psychological tradeoffs that consumers make when evaluating several
attributes together

Disadvantages

Only a limited set of features can be used because the number of combinations
increases very quickly as more features are added.
Information gathering stage is complex.
Difficult to use for product positioning research because there is no procedure for
converting perceptions about actual features to perceptions about a reduced set
of underlying features
Respondents are unable to articulate attitudes toward new categories

Applications of conjoint analysis


Conjoint analysis has been used in marketing for a variety of purposes, including:

Determining the relative importance of attributes in the consumer choice


process. A standard output from conjoint analysis consists of derived relative
importance weights for all the attributes used to construct the stimuli used in the
evaluation task. The relative importance weights indicate which attributes are
important in influencing consumer choice.
Estimating market share of brands that differ in attribute levels. The utilities
derived from conjoint analysis can be used as input into a choice simulator to
determine the share of choices, and hence the market share, of different brands.
Determining the composition of the most-preferred brand. The brand features
can be varied in terms of attribute levels and the corresponding utilities
determined. The brand features that yield the highest utility indicate the
composition of the most-preferred brand.
Segmenting the market based on similarity of preferences for attribute levels. The
part worth functions derived for the attributes may be used as a basis for
clustering respondents to arrive at homogenous preference segments.

Applications of conjoint analysis have been made in consumer goods, industrial goods,
financial and other services. Moreover, these applications have spanned all areas of
marketing. A recent survey of conjoint analysis reported applications in the areas of new
product/concept identification, competitive analysis, pricing, market segmentation,
advertising, and distribution.
SUMMARY
This chapter has given over view conjoint analysis in detail. Conjoint analysis, also
called multi-attribute compositional models. Today it is used in many of the social
sciences and applied sciences including marketing, product management, and
operations research.
KEY TERMS

Kaiser-Meyer-Olkin (KMO) measure of sampling adequacy


Residuals
Conjoint analysis

IMPORTANT QUESTIONS
1. What are the applications of conjoint analysis?
2. Explain the procedure of performing Conjoint Analysis with one practical example.
REFERENCE BOOKS
1. Robert Ferber, Marketing Research, New York: McGraw Hill, Inc. 1948
2. Dennis, Child, The Essentials of Factor Analysis, New York 1973

3. Cooley, William, W., and Lohnes, Paul R., Multivariate Data Analysis, New York:
John Wiley and Sons. 1971.

- End of Chapter -

LESSON 23
APPLICATION OF RESEARCH TOOLS

OBJECTIVES

o
o
o

To identify the areas of management in which Research tools can be used


To understand the application of various research tools in the following domains
of management:
Marketing Management
Operations Management
Human Resources Management

STRUCTURE

Application of marketing research


Concept of market potential
Techniques of perceptual mapping
Limitation of Marketing Research
Methods of Forecasting
Statistical Methods

INTRODUCTION
Research Methodology is becoming a necessary tool for its application of all functional
areas of management such as Marketing, Human Resources and Operations
Management. There is an increasing realisation of the importance of research
methodology in various quarters. This is reflected in the increasing use of research
methodology in various domains of management. Here, a brief description of typical
application of research methodology is given below:
APPLICATIONS OF MARKETING RESEARCH

Applications of marketing research can be divided into two broad areas:

Strategic
Tactical

Among the strategic areas, marketing research applications would be demand


forecasting, sales forecasting, segmentation studies, identification of target markets for
a given product, and positioning strategies identification.
In the second area of tactical applications, we would have applications such as product
testing, pricing research, advertising research, promotional research, distribution and
logistics related research. In other words, it would include research related to all the 'P's
of marketing: how much to price the product, how to distribute it, whether to package it
in one way or another, what time to offer a service, consumer satisfaction with respect to
the different elements of the marketing mix (product, price, promotion, distribution),
and so on. In general, we would find more tactical applications than strategic
applications because these areas can be fine-tuned more easily, based on the marketing
research findings.
Obviously, strategic changes are likely to be fewer than tactical changes. Therefore, the
need for information would be in proportion to the frequency of changes.
The following list is a snapshot of the kind of studies that have actually been done in
India:

A study of consumer buying habits for detergents-frequency, pack size, effect of


promotions, brand loyalty and so forth.
To find out the potential demand for ready-to-eat chapattis in Mumbai city.
To determine which of the three proposed ingredients tulsi, coconut oil, or
neem that the consumer would like to have in a toilet soap
To find out what factors would affect the sales of Flue Gas Desulphurization
equipment {industrial pollution control equipment)
To find out the effectiveness of the advertising campaign for a car brand
To determine brand awareness and brand loyalty for a branded PC (Personal
Computer)
To determine appropriate product mix, price level, and target market for a new
restaurant
To find the customer satisfaction level among consumers of an Internet service
provider
To determine factors which influenced consumers in choosing a brand of cellular
phone handset
To find out the TV viewing preferences of the target audience in specific time
slots in early and late evenings

As the list shows, marketing research tackles a wide variety of subjects. The list is only
indicative, and the applications of marketing research in reality can be useful for almost

any major decision related to marketing. The next sections discuss some typical
application areas.
Concept Research
During a new product launch, there would be several stages-for example, concept
development, concept testing, prototype development and testing, test marketing in a
designated city or region, estimation of total market size based on the test marketing,
and then a national rollout or withdrawal of the product based on the results.
The first stage is the development of a concept and its testing. The concept for a new
product may come from several sources-the idea may be from a brain storming session
consisting of company employees, a focus group conducted among consumers, or the
brainwave of a top executive. Whatever may be its source, it is generally researched
further through what is termed as concept testing, before it goes into prototype or
product development stages.
A concept test takes the form of developing a description of the product, its benefits,
how to use it, and so on, in about a paragraph, and then asking potential consumers to
rate how much they like the concept, how much they would be willing to pay for the
product if introduced, and similar questions. As an example, the concept statement for a
fabric softener may read as follows.
This fabric softener cum whitener is to be added to the wash cycle in a machine or to the
bucket of detergent in which clothes are soaked. Only a few drops of this liquid will be
needed per wash to whiten white clothes and also soften them by eliminating static
charge. It will be particularly useful for woolens, undergarments and babys or childrens
clothes. It will have a fresh fragrance, and will be sold in handy 200 ml, bottles to last
about a month. It can also replace all existing blues with the added benefit of a softener.
This statement can be used to survey existing customers of 'blues' and whiteners, and we
could ask customers for their reactions on pack size, pricing, colour of the liquid, ease of
use, and whether or not they would buy such a product. More complex concept tests can
be done using Conjoint Analysis where specific levels of price or product/service
features to be offered are pre-determined and reactions of consumers are in the form of
ratings given to each product concept combining various features. This is then used to
make predictions about which product concepts would provide the highest utility to the
consumer, and to estimate market shares of each concept. The technique of Conjoint
Analysis is discussed with an example in Part II of the book.
Product Research
Apart from product concepts, research helps to identify which alternative packaging is
most preferred, or what drives a consumer to buy a brand or product category itself, and
specifics of satisfaction or dissatisfaction with elements of a product. These days, service
elements are as important as product features, because competition is bringing most
products on par with each other.

An example of product research would be to find out the reactions of consumers to


manual cameras versus automatic cameras. In addition to specific likes or dislikes for
each product category, brand preferences within the category could form a part of the
research. The objectives may be to find out what type of camera to launch and how
strong the brand salience for the sponsor's brand is.
Another example of product research could be to find out from existing users of
photocopiers (both commercial and corporate), whether after-sales service is
satisfactory, whether spare parts are reasonably priced and easily available, and any
other service improvement ideas - for instance, service contracts, leasing options or buybacks and trade-ins.
The scope of product research is immense, and includes products or brands at various
stages of the product life cycle-introduction, growth, maturity, and decline. One
particularly interesting category of research is into the subject of brand positioning. The
most commonly used technique for brand positioning studies (though not the only one)
is called Multidimensional Scaling. This is covered in more detail with an example and
case studies in Part II as a separate chapter.
Pricing Research
Pricing is an important part of the marketing plan. In the late nineties in India, some
interesting changes have been tried by marketers of various goods and services. Newer
varieties of discounting practices including buy-backs, exchange -offers, and straight
discounts have been offered by many consumer durable manufacturers-notably AKAI
and AIWA brands of TVs. Most FMCG (fast moving consumer goods)
manufacturers/marketers of toothpaste, toothbrush, toilet soap, talcum powder have
offered a variety of price-offs or premium-based offers which affect the effective
consumer price of a product.
Pricing research can delve into questions such as appropriate pricing levels from the
customers' point of view, or the dealer's point of view. It could try to find out how the
current price of a product is perceived, whether it is a barrier for purchase, how a brand
is perceived with respect to its price and relative to other brands' prices (price
positioning). Here, it is worth remembering that price has a functional role as well as a
psychological role. For instance, high price may be an indicator of high quality or high
esteem value for certain customer segments. Therefore, questions regarding price may
need careful framing and careful interpretation during the analysis.
Associating price with value is a delicate task, which may require indirect methods of
research at times. A bland question such as - "Do you think the price of Brand A of
refrigerators is appropriate?" may or may not elicit true responses from customers. It is
also not easy for a customer to articulate the price he would be willing to pay for
convenience of use, easy product availability, good after-sales service, and other
elements of the marketing mix. It may require experience of several pricing-related
studies before one begins to appreciate the nuances of consumer behaviour related to
price as a functional and psychological measure of the value of a product offering.

An interesting area of research into pricing has been determining price elasticity at
various price points for a given brand through experiments or simulations. Price
framing, or what the consumer compares (frames) price against, is another area of
research. For example, one consumer may compare the price of a car against an
expensive two-wheeler (his frame of reference), whereas another may compare it with
an investment in the stock market or real estate. Another example might be the interest
earned from a fixed deposit, which serves as a benchmark for one person before he
decides to invest in a mutual fund, whereas for another, the investment may be a
substitute for buying gold, which earns no interest. In many cases, therefore, it is the
frame of reference used by the customer which determines 'value' for him of a given
product. There are tangible as well as intangible (and sometimes not discernible)
aspects to a consumer's evaluation of price. Some of the case studies at the end of Part I
include pricing or price-related issues as part of the case.
Distribution Research
Traditionally, most marketing research focuses on consumers or buyers. Sometimes this
extends to potential buyers or those who were buyers but have switched to other brands.
But right now, there is a renewed interest in the entire area of logistics, supply chain,
and customer service at dealer locations. There is also increasing standardisation from
the point of view of brand building, in displays at the retail level, promotions done at the
distribution points. Distribution research focuses on various issues related to the
distribution of products including service levels provided by current channels, frequency
of sales persons visits to distribution points, routing/transport related issues for
deliveries to and from distribution points throughout the channel, testing of new
channels, channel displays, linkages between displays and sales performance, and so on.
As an example, a biscuit manufacturer wanted to know how it could increase sales of a
particular brand of biscuits in cinema theatres. Should it use existing concessionaires
selling assorted goods in theatres, or work out some exclusive arrangements? Similarly,
a soft drink manufacturer may want to know where to set up vending machines.
Potential sites could include roadside stalls, shopping malls, educational institutions,
and cinema theatres. Research would help identify factors that would make a particular
location a success.
In many service businesses where a customer has to visit the location, it becomes very
important to research the location itself. For example, a big hotel or a specialty
restaurant may want to know where to locate themselves for better visibility and
occupancy rates. Distribution research helps answer many of these questions and
thereby make better marketing decisions.
Advertising Research
The two major categories of research in advertising are:
1. Copy
2. Media

Copy Testing
This is a broad term that includes research into all aspects of advertising-brand
awareness, brand recall, copy recall (at various time periods such as day after recall,
week after recall), recall of different parts of the advertisement such as the headline for
print ads, slogan or jingle for TV ads, the star in an endorsement and so on. Other
applications include testing alternative ad copies (copy is the name given to text or
words used in the advertisement, and the person in the advertising agency responsible
for writing the words is known as the copy writer) :or a single ad, alternative layouts (a
layout is the way all the elements of, the advertisement are laid out in a print
advertisement) with the same copy, testing of concepts or storyboards (a storyboard is a
scene-by-scene drawing of a TV commercial which is like a rough version before the ad
is actually shot on film) of TV commercials to test for positive/negative reactions, and
many others. Some of these applications appear in our discussion of Analysis of
Variance (ANOVA) in Part II and some case studies elsewhere in the book.
A particular class of advertising research is known as Tracking Studies. When an
advertising campaign is running, periodic sample surveys known as tracking studies can
be conducted to evaluate the effect of the campaign over a long period of time such as
six months or one year, or even longer. This may allow marketers to alter the advertising
theme, content, media selection or frequency of airing / releasing advertisements and
evaluate the effects. As opposed to a snapshot provided by a one-time survey, tracking
studies may provide a continuous or near-continuous monitoring mechanism. But here,
one should be careful in assessing the impact of the advertising on sales, because other
factors could change along with time. For example, the marketing programmes of the
sponsor and the competitors could vary over time. The impact on sales could be due to
the combined effect of several factors.
Media Research
The major activity under this category is research into viewership of specific television
programmes on various TV channels. There are specialised agencies like A.C. Nielsen
worldwide which offer viewership data on a syndicated basis (i.e., to anyone who wants
to buy the data). In India, both ORG-MARG and IMRB offer this service. They provide
people meter data with brand names of TAM and INTAM which is used by advertising
agencies when they draw up media plans for their clients. Research could also focus on
print media and their readership. Here again, readership surveys such as the National
Readership Survey (NRS) and the Indian Readership Survey (IRS) provide syndicated
readership data. These surveys are now conducted almost on a continuous basis in India
and are helpful to find out circulation and readership figures of major print media. ABC
(Audit Bureau of Circulations) is an autonomous body which provides audited figures on
the paid circulation (number of copies printed and sold) of each newspaper and
magazine, which is a member of ABC.
Media research can also focus on demographic details of people reached by each
medium, and also attempt to correlate consumption habits of these groups with their
media preferences. Advertising research is used at all stages of advertising, from

conception to release of ads and thereafter to measure advertising effectiveness based


on various parameters. It is a very important area of research for brands that rely a lot
on advertising. The top rated programmes in India are usually cricket matches and film
based programmes.
Sales Analysis by Product
Sales analysis by product will enable a company to identify its strong or weak products.
It is advisable to undertake an analysis on the basis of a detailed break-up of -products
such as product variation by size, colour, etc. This is because if an analysis is based on a
broad break-up, it may not reveal important variations.
When a company finds that a particular product is doing poorly, two Options are open
to it. One is, it may concentrate on that product to ensure improved sales. Or
alternatively, it may gradually withdraw the product and eventually drop it altogether.
However, it is advisable to decide on the latter course on the basis of additional
information such as trends in the market share, contribution margin, and effect of sales
volume on product profitability, etc. In case the product in question has
complementarity with other items sold by the company, the decision to abandon the
product must be made with care and caution.
Combining sales analysis by product with that by territory will further help in providing
information on which products are doing better in which areas.
Sales Analysis by Customers
Another way to analyse sales data is by customers. Such an analysis would normally
indicate that a relatively small number of customers accounts for a large proportion of
sales. To put it differently: a large percentage of customers accounts for a relatively
small percentage of aggregate sales. One may compare the data with the proportion of
time spent on the customers, i.e. the number of sales calls. An analysis of this type will
enable the company to devote relatively more time to those customers who collectively
account for proportionately larger sales.
Sales analysis by customer can also be combined with analysis both by area and product.
Such an analysis will prove to be more revealing. For example, it may indicate that in
some areas sales are not increasing with a particular type of customer though they have
grown fast in other areas. Information of this type will be extremely useful to the
company as it identifies the weak spots where greater effort is called for.
Sales Analysis by Size of Order
Sales analysis by size of order may show that a large volume of sales is accompanied by
low profit and vice versa. In case cost accounting data are available by size of order, this
would help in identifying sales where the costs are relatively high and the company is
incurring a loss. Sales analysis by size of order can also be combined with that by
products, areas and types of customers. Such a perceptive analysis would reveal useful

information to the company and enable it to make a more rational and effective effort in
maximizing its return from sales.
THE CONCEPT OF MARKET POTENTIAL
Market potential has been defined as "the maximum demand response possible for a
given group of customers within a well-defined geographic area for a given product or
service over a specified period of time under well-defined competitive and
environmental conditions".
We will elaborate this comprehensive definition. First, market potential is the maximum
demand response under certain assumptions. It denotes a meaningful boundary
condition on ultimate demand.
Another condition on which the concept of market potential depends is a set of relevant
consumers of the product or service. It is not merely the present consumer who is to be
included but also the potential consumer as maximum possible demand is to be
achieved. Market potential will vary depending on which particular group of consumers
is of interest.
Further, the geographic area for which market potential is to be determined should be
well-defined. It should be divided into mutually exclusive subsets of consumers so that
the management can assign a sales force and supervise and control the activities in
different territories without much difficulty.
Another relevant aspect in understanding the concept of market potential is to clearly
know the product or service for which market potential is to be estimated. Especially in
those cases where the product in question can be substituted by another, it is desirable
to have market potential for the product class rather than that particular product. For
example, tea is subjected to a high degree of cross-elasticity of demand with coffee.
It is necessary to specify the time period for which market potential is to be estimated.
The time period should be so chosen that it coincides with planning periods in a firm.
Both short and long-time periods can be used depending on the requirements of the
firm.
Finally, a clear understanding of environmental and competitive conditions relevant in
case of a particular product or service is necessary if market potential is to be useful.
What is likely to be the external environment? What is likely to be the nature and extent
of competition? These are relevant questions in the context of any estimate of market
potential since these are the factors over which the firm has no control.
It may be emphasised that market potential is not the same thing as sales potential and
sales forecast. It is only when "a market is saturated can the industry sales forecast be
considered equivalent to market potential". Such a condition is possible in case of well
established and mature products. Generally, the industry sales forecast will be less than
the market potential. Likewise, a company's sales forecast will be less than its sales

potential. The former is a point estimate of the future sales, while the latter represents a
boundary condition which the sales might reach in an ideal situation. "In the latter
sense, sales potential is to a firm what market potential is to an industry or product
class: both represent maximum demand response and are boundary conditions".
Brand Positioning
Brand positioning is a relatively new concept in marketing. The concept owes its origin
to the idea that each brand occupies a particular space in the consumer's mind,
signifying his perception of the brand in question in relation to other brands. While
product or brand positioning has been defined by various authors in different ways, the
underlying meaning conveyed through these definitions seems to be the same. Instead
of giving several definitions, we may give one here. According to Green and Tull,
"Brand positioning and market segmentation appear to be the hallmarks of today's
marketing research. Brand (or service) positioning deals with measuring the perceptions
that buyers hold about alternative offerings".
From this definition it is evident that the term 'position' reflects the essence of a brand
as perceived by the target consumer in relation to other brands. In view of this the
management's ability to position its product or brand appropriately in the market can be
a major source of company's profits. This seems to be an important reason for the
emergence of product or brand positioning as a major area in marketing research.
Components of Positioning
Positioning comprises four components. The first component is the product class or the
structure of the market in which a company's brand will compete. The second
component is consumer segmentation. One cannot think of positioning a brand without
considering the segment m which it is to be offered. Positioning and segmentation are
inseparable. The third component is the consumers perception of the company's brand
in relation to those of the competitors. Perceptual mapping is the device by which the
company can know this. Finally, the fourth component positioning is the benefit offered
by the company's brand. A consumer can allot a position in his mind to a brand only
when it is beneficial to him. The benefits may be expressed as attributes or dimensions
in a chart where brands are fitted to indicate the consumer's perceptions.
As perceptual maps are used to indicate brand positioning, blank spaces in such maps
show that a company can position its brand in one or more of such spaces.
TECHNIQUES FOR PERCEPTUAL MAPPING
There are a number of techniques for measuring product positioning. Some of these
which are important are:

Factor analysis
Cluster analysis Multi-dimensional scaling.

We will not go into the detailed mechanism of these techniques. All the same, we will
briefly explain the techniques.
Image profile analysis
This technique is the oldest and most frequently used for measuring the consumers
perceptions of competitive brands or services. Normally, a 5 or 7 point numerical scale
used. A number of functional and psychological attributes are selected. The respondent
is asked to show his perception of each brand in respect of each attribute on the 5 or 7
point scale.
It will be seen that the figures provides some insight as to which brands are competing
with each other and on what attribute(s). This technique has some limitations. First, if
the number of brands is large, it may not be possible to plot all the brands in a single
figure. Second, there is an implicit assumption in this technique that all attributes are
equally important and independent of each other. This is usually not true. However, this
limitation can be overcome by using the technique of factor analysis.
Factor analysis
As regards factor analysis, it may be pointed out that its main object is to reduce a large
number of variables into a small number of factors or dimensions. In Chapter 17, two
examples have been given to illustrate the use of factor analysis. The discussion also
brings out some major limitations of the method.
Cluster analysis
Cluster analysis is used to classify consumers or objects into a small number of mutually
exclusive and exhaustive groups. With, the help of cluster analysis it is possible to
separate brands into clusters or groups so that the brand within a cluster is similar to
other brands belonging to the same cluster and is very different from brands included in
other clusters. This method has been discussed in Chapter
Multi-dimensional scaling
Multi-dimensional scaling too has been discussed in Chapter 17, pointing out how
perceptual maps can be developed on the basis of responses from consumers. In this
connection, two illustrations of perceptual maps were given. The first illustration related
to select Business Schools based on hypothetical data On the basis of two criteria, viz.
how prestigious and quantitative an MBA course is different Business Schools have been
shown in the map. It will be seen that the MBA course of Business School 'C is extremely
different from that offered by Business School Q Points which are close to each other
indicate similarity of the MBA courses in the students perception. The second
illustration related to four brands of washing soaps based on a survey data from
Calcutta. This is a non-attribute based example where a paired comparison for four
high- and-medium-priced detergents Surf, Sunlight, Gnat and Key was undertaken. As
mentioned there, Sunlight and Surf are closest and Surf and Key are farthest. In other

words the first two brands are most similar and the remaining two are most dissimilar.
How the points in the figures for the four brands have been arrived at has been
explained at length in that chapter and so is not repeated here.
Subroto Sengupta has discussed at length product positioning in his book. While
explaining different techniques of product positioning, he has shown how the concept of
positioning can be used to improve the image of the concerned product or brand. He has
given a number of examples covering a wide variety of products such as coffee, soft
drinks, washing soaps, toilet soaps, shampoos and magazines As Sengupta points out
the perceptual maps of product class also indicate holes or vacant positions in the
market. These open spaces can be helpful to the management in suggesting new product
opportunities as also possibilities for repositioning of old products. While it is true that
the management does get the clues on preferred attributes of the product in question, it
is unable to know all the relevant features of the new product such as its form, package
and price. This problem can be overcome through the application of the conjoint
analysis. In addition, Sengupta has discussed some research studies in respect of
advertising positioning.
We now give a detailed version of a study indicating how a brand which was putting up a
poor performance in the market was repositioned. As a result, it improved its image and
contributed to increased market share and profits.
When to do Marketing Research?
Marketing research can be done when:
There is an information gap which can be filled by doing research.
The cost of filling the gap through marketing research is less than the cost of
taking a wrong decision without doing the research.
o The time taken for the research does not delay decision-making beyond
reasonable limits.
o A delay can have many undesirable effects, like competitors becoming aware of
strategies or tactics being. Contemplated, consumer opinion changing between
the beginning and end of the study, and so forth.
o
o

LIMITATIONS OF MARKETING RESEARCH


It must be kept in mind that marketing research, though very useful most of the time, is
not the only input for decision-making. For example, many small businesses work
without doing marketing research, and some of them are quite successful. It is obviously
some other model of informal perceptions about consumer behaviour, needs, and
expectations that is at work in such cases. Many businessmen and managers base their
work on judgment, intuition, and perceptions rather than numerical data.
There is a famous example in India, where a company commissioned a marketing
research company to find out if there was adequate demand for launching a new
camera. This was in pre-liberalised India, of the early 1980s. The finding of the research

study was that there was no demand, and that the camera would not succeed, if
launched. The company went ahead and launched it anyway, and it was a huge success.
The camera was Hot Shot. It was able to tap into the need of consumers at that time for
an easy-to-use camera at an affordable price.
Thus marketing research is not always the best or only source of information to be used
for making decisions. It works best when combined with judgment, intuition,
experience, and passion. For instance, even if marketing research were to show there
was demand for a certain type of product, it still depends on the design and
implementation of the appropriate marketing plans to make it succeed. Further,
competitors could take actions which were not foreseen when marketing research was
undertaken. This also leads us to conclude that the time taken for research should be the
minimum possible, if we expect the conditions to be dynamic, or fast changing.
Differences in Methodology
The reader may be familiar with research studies or opinion polls conducted by different
agencies showing different results. One of the reasons why results differ is because the
methodology followed by each agency is usually different. The sampling method used,
the sample size itself, the representativeness of the population, the quality of field staff
who conduct interviews, and conceptual skills in design and interpretation all differ
from agency to agency. Minor differences are to be expected in sample surveys done by
different people, hut major differences should be examined for the cause, which will
usually lead us to the different methodologies adopted by them. Based on the credibility
of the agency doing the research and the appropriateness of the methodology followed,
the user decides which result to rely upon. A judgment of which methodology is more
appropriate for the research on hand comes from experience of doing a variety of
research.
To summarise, it is important to understand the limitations of marketing research, and
to use it in such a way that we minimise its limitations.
Complementary Inputs for Decision-Making
Along with marketing research, marketing managers may need to look into other
information while making a decision. For example, our corporate policy may dictate that
a premium image must be maintained in all activities of our company. On the other
hand, marketing research may tell us that consumers want a value-for-money product.
This creates a dilemma for the basic corporate policy, which has to be balanced with
consumer perception as measured by marketing research.
Other inputs for decision-making could be growth strategies for the brand or product,
competitors' strategies, and regulatory moves by the government and others. Some of
these are available internally-for example, corporate policy and growth plans may be
documented internally. Some other inputs may come from a marketing intelligence cell
if the company has one. In any case, marketing decisions would be based on many of
these complementary inputs, and not on the marketing research results alone.

Secondary and Primary Research


One of the most basic differentiations is between secondary and primary research.
Secondary research is any information we may use, but which has not been specifically
collected for the current marketing research. This includes published sources of data,
periodicals, newspaper reports, and nowadays, the Internet. It is sometimes possible to
do a lot of good secondary research and get useful information. But marketing research
typically requires a lot of current data which is not available from secondary sources.
For example, the customer satisfaction level for a product or brand may not be reported
anywhere. The effectiveness of a particular advertisement may be evident from the sales
which follow. But why people liked the advertisement may not be obvious, and can only
be ascertained through interviews with consumers. Also, the methodology for the
secondary data already collected may be unknown, and therefore we may be unable to
judge the reliability and validity of the data.
Primary research is what we will be dealing with throughout this book. It can be defined
as research which involves collecting information specifically for the study on hand,
from the actual sources such as consumers, dealers or other entities involved in the
research. The obvious advantages of primary research are that it is timely, focused, and
involves no unnecessary data collection, which could be a wasted effort. The
disadvantage could be that it is expensive to collect primary data. But when an
information gap exists, the cost could be more than compensated by better decisions,
which are taken with the collected data.
Thus Research has become a significant element of successful Marketing System. The
following section speaks about application of Research in Operation/Production
Management domains.
Applications of Research Methodology in Operations Management
Methods of Estimating Current Demand
There are two types of estimates of current demand which may be helpful to a company.
These are: total market potential and territory potential. "Total market potential is the
maximum amount of sales that might be available to all the firms in an industry during
a given period under a given level- of industry marketing effort and given environmental
conditions".
Symbolically, total market potential is:
Q=n x q x p
where
Q = total market potential
n = number of buyers in the specific product/market under the given assumptions

q = quantity purchased by an average buyer


p = price of an average unit
Of the three components n, q, and p in the above formula, the most difficult component
to estimate is q. One can start with a broad concept of q, gradually reducing it. For
example, if we are thinking of readymade shirts for home consumption, we may first
take the total male population eliminating that in rural areas. From the total male urban
population, we may eliminate the age groups which are not likely to buy readymade
shirts. Thus, the number of boys below 20 may be eliminated. Further eliminations on
account of low income may be made. In this way we can arrive at the 'prospect pool' of
those who are likely to buy shirts.
The concept of market potential is helpful to the firm as it provides a benchmark against
which actual performance can be measured. In addition, it can be used as a basis for
allocation decisions regarding marketing effort.
The estimate of total market potential is helpful to the company when it is in a dilemma
whether to introduce a new product or drop an existing one. Such an estimate will
indicate whether the prospective market is large enough to justify the company's
entering it.
Since it is impossible for a company to have the global market exclusively to itself, it has
to select those territories where it can sell its products well. This means that companies
should know the territorial potentials so that they can select markets most suited to
them, channelise their marketing effort optimally among these markets and also
evaluate their sale performance in such markets.
There are two methods for estimating territorial potentials: (i) market-buildup method,
and (ii) index-of-buying-power method. In the first method, several steps are involved.
First, identify all the potential buyers for the product in each market. Second, estimate
potential purchases by each potential buyer. Third, sum up the individual figures in step
(ii) above. However, in reality the estimation is not that simple as it is difficult to
identify all potential buyers. When the product in question is an industrial product,
directories of manufacturers of a particular product or group of products are used.
Alternatively, the Standard Industrial Classification of Manufacturers of a particular
product or group of products is used.
The second method involves the use of a straight forward index. Suppose a textile
manufacturing company is interested in knowing the territorial potential for its cloth in
a certain territory. Symbolically,
Bi = 0.5Yi + 0.2ri + 0.3Pi
where
Bi = percentage of total national buying power in territory i

Yi = percentage of national disposable personal income originating in territory i


ri = percentage of national retail sales in territory i
Pi = percentage of national population living in territory i
It may be noted that such estimates indicate potential for the industry as a whole rather
than for individual company. In order to arrive at a company potential, the concerned
company has to make certain adjustments in the above estimate on the basis of one or
more other factors that have not been covered in the estimation of territorial potential.
These factors could be the company's brand share, lumber of salespersons, number and
type of competitors, etc.
Forecasting Process
After having described the methods of estimating the current demand, we now turn to
forecasting. There are five steps involved in the forecasting process. These are
mentioned below.
First, one has to decide the objective of the forecast. The marketing researcher should
know as to what will be the use of the forecast he is going to make.
Second, the time period for which the forecast is to be made should be selected. Is the
forecast short-term, medium-term or long-term? Why should a particular period of
forecast be selected?
Third, the method or technique of forecasting should be selected. One should be clear
as to why a particular technique from amongst several techniques should be used.
Fourth, the necessary data should be collected. The need for specific data will depend
on the forecasting technique to be used.
Finally, the forecast is to be made. This will involve the use of computational
procedures.
In order to ensure that the forecast is really useful to the company, there should be good
understanding between management and research. The management should clearly
spell out the purpose of the forecast and how it is going to help the company. It should
also ensure that the researcher has a proper understanding of the operations of the
company, its environment, past performance in terms of key indicators and their
relevance to the future trend. If the researcher is well-informed with respect to these
aspects, then he is likely to make a more realistic and more useful forecast for the
management.
Methods of Forecasting

The methods of forecasting can be divided into two broad categories, viz. subjective or
qualitative methods and objective or quantitative methods. These can be further divided
into several methods. Each of these methods is discussed below.
I. Subjective / Qualitative Methods
There are four subjective methods - field sales force, jury of executives, users'
expectations and delphi. These are discussed here briefly.
a. Field sales force
Some companies ask their salesmen to indicate the most likely sales for a specified
period in the future. Usually the salesman is asked to indicate anticipated sales for each
account in his territory. These forecasts are checked by district managers who forward
them to the company's head office. Different territory-forecasts are then combined into
a composite forecast at the head office. This method is more suitable when a short-term
forecast is to be made as there would be no major changes in this short period affecting
the forecast. Another advantage of this method is that it involves the entire sales force
which realises its responsibility to achieve the target it has set for itself. A major
limitation of this method is that sales force would not take an overall or broad
perspective and hence may overlook some vital factors influencing the sales. Another
limitation is that salesmen may give somewhat low figures in their forecasts thinking
that it may be easier for them to achieve those targets. However, this can be offset to a
certain extent by district managers who are supposed to check the forecasts.
b. Jury of executives
Some companies prefer to assign the task of sales forecasting to executives instead of a
sales force. Given this task each executive makes his forecast for the next period. Since
each has his own assessment of the environment and other relevant factors, one forecast
is likely to be different from the other. In view of this it becomes necessary to have an
average of these varying forecasts. Alternatively, steps should be taken to narrow down
the difference in the forecasts. Sometimes this is done by organising a discussion
between the executives so that they can arrive at a common forecast. In case this is not
possible, the chief executive may have to decide which of these forecasts is acceptable as
a representative one.
This method is simple. At the same time, it is based on a number of different viewpoints
as opinions of different executives are sought. One major limitation of this method is
that the executives' opinions are likely to be influenced in one direction on the basis of
general business conditions.
c. Users' expectations
Forecasts can be based on users' expectations or intentions to purchase goods and
services. It is difficult to use this method when the number of users is large. Another
limitation of this method is that though it indicates users' 'intentions' to buy, the actual

purchases may be far less at a subsequent period. It is most suitable when the number of
buyers is small such as in case of industrial products.
d. The Delphi method
This method too is based on the experts' opinions. Here, each expert has access to the
same information that is available. A feedback system generally keeps them informed of
each others' forecasts but no majority opinion is disclosed to them. However, the experts
are not brought together. This is to ensure that one or more vocal experts do not
dominate other experts.
The experts are given an opportunity to compare their own previous forecasts with those
of the others and revise them. After three or four rounds, the group of experts arrives at
a final forecast.
The method may involve a large number of experts and this may delay the forecast
considerably. Generally it involves a small number of participants ranging from 10 to
40.
It will be seen that both the jury of executive opinion and the Delphi method are based
on a group of experts. They differ in that in the former, the group of experts meet,
discuss the forecasts, and try To arrive at a commonly agreed forecast while in the latter
the group of experts never meet. As mentioned earlier, this is to ensure that no one
person dominates the discussion thus influencing the forecast. In other words, the
Delphi method retains the wisdom of a group and at the same time reduces the effect of
group pressure. An approach of this type is more appropriate when long-term forecasts
are involved.
In the subjective methods, judgments are an important ingredient. Before attempting a
forecast, the basic assumptions regarding environmental conditions as also competitive
behaviour must be provided people involved in forecasting. An important advantage of
subjective methods is that they are easily understood. Another advantage is that the cost
involved in forecasting is quite low.
As against these advantages, subjective methods have certain limitations also. One
major limitation is the varying perceptions of people involved in forecasting. As a result,
wide variance is found in forecasts. Subjective methods are suitable when forecasts are
to be made for highly technical products which have a limited number of customers.
Generally, such methods are used for industrial products. Also when cost of forecasting
is to be kept to a minimum, subjective methods may be more suitable.
II. Objective / Quantitative or Statistical Methods
Based on statistical analysis, these methods enable the researcher to make forecasts on a
more objective basis. It is difficult to make a wholly accurate forecast for there is always
an element of uncertainty regarding the future'. Even so, statistical methods are likely to
be more useful as they are more scientific and hence more objective.

a. Time Series
In time-series forecasting, the past sales data are extrapolated as a linear or a curvilinear
trend. Even if such data are plotted on a graph, one can extrapolate for the desired time
period. Extrapolation can be made with the help of statistical techniques.
It may be noted that time-series forecasting is most suitable to stable situations where
the future ids will largely be an extension of the past. Further, the past sales data should
have distinctive ids from the random error component for a time-series forecasting to be
suitable.
Before using the time-series forecasting one has to decide how far back in the past one
can go. It may be desirable to use the more recent data as conditions might have been
different in the remote past. Another issue pertains to weighting of time-series data. In
other words, should equal weight be given to each time period or should greater
weightage be given to more recent data? Finally, should data be decomposed into
different components, viz. trend, cycle, season and error? We now discuss methods, viz.
moving averages, exponential smoothing and decomposition of time series.
b. Moving average
This method uses the last 'n' data points to compute a series of average in such a way
that each time latest figure is used and the earliest one dropped. For example, when we
have to calculate a five monthly moving average, we first calculate the average of
January, February, March, April and May by adding the figures of these months, and
dividing the sum by five. This will give one figure. In the next calculation, the figure for
June will be included and that for January dropped thus giving a new average. Thus a
series of averages is computed. The method is called as 'moving' average as it uses w
data point each time and drops the earliest one.
In a short-term forecast, the random fluctuations in the data are of major concern. One
method of minimizing the influence of random error is to use an average of several past
data points. This is achieved by the moving average method. It may be noted that in a
12-month moving average, the effect of seasonality is removed from the forecast as data
points for every season are included before computing the moving average.
c. Exponential smoothing
A method which has been receiving increasing attention in recent years is known as
exponential smoothing. It is a type of moving average that 'smoothens' the time-series.
When a large number of forecasts are to be made for a number of items, exponential
smoothing is particularly suitable as it combines the advantages of simplicity of
computation and flexibility.
d. Time-series decomposition

This method consists of measuring the four components of a time-series (i) trend, (ii)
cycle, (iii) season, and (iv) erratic movement.
(i) The trend component indicates long-term effects on sales that are caused by
such factors as income, population, industrialisation and technology. The time
period a trend function varies considerably from product to product However, it
is usually taken as any period in excess of the time period required for a business
cycle (which averages at 4-5 years).
(ii) The cyclical component indicates some sort of a periodicity in the general
economic activity. When data are plotted, they yield a curve with peaks and
troughs, indicating increases and falls in the trend series with a certain
periodicity. A careful study of the impact of a business cycle must be made on the
sale of each product. Cyclical forecasts are likely to be more accurate for the longterm than for the short term.
(iii) The seasonal component reflects changes in sales levels due to factors such as
weather, festivals, holidays, etc. There is a consistent pattern of sales for period
within a year.
(iv) Finally, the erratic movements in data arise on account of events such as
strikes, lockouts, price wars, etc. The decomposition of time-series enables
identification of the error component from the trend, cycle and season which are
systematic components.
Casual or Explanatory Methods
Causal or explanatory methods are regarded as the most sophisticated methods of
forecasting sales. These methods yield realistic forecasts provided relevant data are
available on the- major variables influencing changes in sales. There are three distinct
advantages of causal methods. First, turning points in sales can be predicted more
accurately by these methods than by time-series methods. Second, the use of these
methods reduces the magnitude of the random component far more than it may be
possible with the time-series methods. Third, the use of such methods provides greater
insight to causal relationships. This facilitates the management in marketing decision
making. Isolated sales recasts on the basis of time-series methods would not be helpful
in this regard.
Causal methods can be either (i) leading indicators or (ii) regression models. These are
briefly discussed here.
(i) Leading indicators
Sometimes one finds that changes in sales of a particular product or service are
preceded by changes one or more leading indicators. In such cases, it is necessary to
identify leading indicators and to closely observe changes in them. One example of a
leading indicator is the demand for various household appliances which follows the

construction of new houses. Likewise, the demand for many durables is preceded by an
increase in disposable income. Yet another example is of number of births. The demand
for baby food and other goods needed by infants can be ascertained by the number of
births in a territory. It may be possible to include leading indicators in regression
models.
(ii) Regression models
Linear regression analysis is perhaps the most frequently used and the most powerful
method among casual methods. As we have discussed regression analysis in detail in the
preceding chapters on Bivariate Analysis and Multivariate Analysis, we shall only dwell
on a few relevant points.
First, regression models indicate linear relationships within the range of observations
and at the les when they were made. For example, if a regression analysis of sales is
attempted on the basis of independent variables of population sizes of 15 million to 30
million and per capita income of Rs 1000 to Rs.2500, the regression model shows the
relationships that existed between these extremes in the two independent variables. If
the sales forecast is to be made on the basis of values of independent variables falling
outside the above ranges, then the relationships expressed by the regression model may
not hold good. Second, sometimes there may be a lagged relationship between the
dependent and independent variables. In such cases, the values of dependent variables
are to be related to those of independent variables for the preceding month or year as
the case may be. The search for factors with a lead-lag relationship to the sales of a
particular product is rather difficult. One should tryout several indicators before
selecting the one which is most satisfactory. Third, it may happen that the data required
to establish the ideal relationship, do not exist or are inaccessible or, if available, are not
useful. Therefore, the researcher has to be careful in using the data. He should be quite
familiar with the varied sources and types of data that can be used in forecasting. He
should also know about their strengths and limitations. Finally, regression model
reflects the association among variables. The causal interpretation is done by the
researcher on the basis of his understanding of such an association. As such, he should
be extremely careful in choosing the variables so that a real causative relationship can be
established among the variables chosen.
Input-Output Analysis
Another method that is widely used for forecasting is the input-output analysis. Here,
the researcher takes into consideration a large number of factors, which affect the
outputs he is trying to forecast. For this purpose, an input-output table is prepared,
where the inputs are shown horizontally as the column headings and the outputs
vertically as the stubs. It may be mentioned that by themselves input-output flows are of
little direct use to the researcher. It is the application of an assumption as to how the
output of an industry is related to its use of various inputs that makes an input-output
analysis a good method of forecasting. The assumption states that as the level of an
industry's output changes, the use of inputs will change proportionately, implying that

there is no substitution in production among the various inputs. This may or may not
hold good.
The use of input-output analysis in sales forecasting is appropriate for products sold to
governmental, institutional and industrial markets as they have distinct patterns of
usage. It is seldom used for consumer products and services. It would be most
appropriate when the levels and kinds of inputs required achieving certain levels of
outputs need to be known.
A major constraint in the use of this method is that it needs extensive data for a large
number of items which may not be easily available. Large business organisations may be
in a position to collect such data on a continuing basis so that they can use input-output
analysis for forecasting. However, this is not possible m case of small industrial
organisations on account of excessive costs involved in the collection of comprehensive
data. It is for this reason that input, output analysis is less widely used than most
analysts initially expected. A detailed discussion of input-output analysis is beyond the
scope of this book.
Econometric Model
Econometrics is concerned with the use of statistical and mathematical techniques to
verify hypotheses emerging in economic theory. "An econometric model incorporates
functional relationships between estimated techniques into an internally consistent and
logically self-contained framework". Econometrics use both exogenous and endogenous
variables. Exogenous variables are used as inputs into, but they themselves are
determined outside the model. These variables include policy variables controlled
events. In contrast, endogenous variables are those which are determined within the
system.
The use of econometric models is generally found at the macro level such as forecasting
national e and its components. Such models show how the economy or any of its specific
segments operates. As compared to an ordinary regression equation they bring out the
causalities involved more distinctly. This merit of econometric models enables them to
predict turning points more accurately. However, their use at the micro-level for
forecasting has so far been extremely limited.
Applications of research methodology in human resources management
Research methodology widely used in the domains of Human resources Management. It
is called as Human resources Metrics (HR Metrics).
To move to the center of the organization, HR must be able to talk in quantitative,
objective terms. Organizations are managed by data. Unquestionably, at times,
managers make decisions based on emotions as facts. Nevertheless, day-to-day
operations are discussed, planned and evaluated in hard data terms.

Perhaps the most crucial advantage of a sound HR metrics programme is that it enables
HR to converse with senior management in the language of business. Operational
decisions taken by HR are then based on cold, hard facts rather than gut feeling, the
figures being used to back up business cases and requests for resource. The HR function
is transformed from a bastion of 'soft' intangibles into something more 'scientific', better
able to punch its weight in the organisation. In addition, the value added by HR
becomes more visible. This will become increasingly important as more and more
functions attempt to justify their status as strategic business partners rather than merely
cost centres.
The five key practices of the Human Capital Index are as follows:
1. Recruiting excellence
2. Clear rewards and accountability,
3. Prudent use of resources,
4. Communications integrity
5. Collegial flexible workplace

They require the capture of metrics for their very definition. Metrics help
quantify and demonstrate the value of HR
Metrics can help guide workforce strategies and maximize return on HR
investments
Metrics provide measurement standards
Metrics help show what HR contributes to overall business results

SUMMARY
The above lesson has given brief introduction to various applications of various research
techniques in management. It has identified the appropriate tools for applications of
various domains of Management.
KEY TERMS

Marketing research
Brand positioning
Image profile analysis
Market Potential
Demand Measurement
Delphi Method
Time Series analysis
Moving average
HR Metrics

IMPORTANT QUESTIONS
1.
2.
3.
4.
5.
6.
7.
8.

What are all the various domains in which research tools can be used?
Explain the application of Image profile analysis with example.
Differentiate between Primary and secondary research.
What are the limitations of marketing research?
Describe the method for finding out the Market potential.
Explain the Various methods to estimate the Demand.
What do you mean by HR Metrics?
Note down the five key practices of the Human Capital Index.

- End of Chapter LESSON - 24


REPORT PREPARATIONS

OBJECTIVES

o
o
o
o
o
o
o

To learn the structure of Professional Research report


To understand the application of following diagrams
Area Chart
Line Graph
Bar chart
Pie chart
Radar diagram
Surface diagram
Scatter diagram

STRUCTURE

Research report
Report format
Data presentation
Pareto chart

RESEARCH REPORT

The final step in the research process is the preparation and presentation of the research
report. A research report can be defined as the presentation of the research findings
directed to a specific audience to accomplish a specific purpose.
Importance of report
The research report is important for the following reasons:
1. The results of research can be effectively communicated to management.
2. The report is the only aspect of the study, which executives are exposed to and their
consecutive evaluation of the project rests with the effectiveness of the written and oral
presentation.
3. The report presentations are typically the responsibility of the project worthiness. So,
the communication effectiveness and usefulness of the information provided plays a
crucial role in determining whether that project will be continued in future.
Steps in report preparation
Preparing a research report involves three steps
1. Understanding the research
2. Organizing the information
3. Writing with effectiveness
Guidelines
The general guidelines that should be followed for any report are as follows:
1. Consider the audience: The information resulting from research is ultimately
important to the management, who will use the results to make decisions. Decision
makers are interested in a clear, concise, accurate and interesting report, which directly
focuses on their information needs with a minimum of technological jargons. Thus, the
report has to be understood by them; the report should not be too technical and not too
much jargon should be used. This is a particular difficulty when reporting the results of
statistical analysis where there is a high probability that few of the target audience have
a grasp of statistical concepts. Hence, for example, there is a need to translate such
terms as standard deviation, significance level, confidence interval etc. into everyday
language.
2. Be concise, but precise: The real skill of the researcher is tested in fulfilling this
requirement. The report must be concise and must focus on the crucial elements of the
project. It should not include the unimportant issues. Researcher should know how
much emphasis has to be given to each area.

3. Be objective, yet effective: The research report must be an objective presentation


of the research findings. The researcher violates the standard of objectivity if the
findings are presented in a distorted or slanted manner. The findings can be presented
in a manner, which is objective, yet effective. The writing style of the report should be
interesting, with the sentence structure short and to the point.
4. Understand the results and draw conclusions: The managers who read the
report are expecting to see interpretive conclusions in the report. The researcher should
understand the results and be able to interpret it effectively to management. Simply
reiterating the facts will not do, implications has to be drawn using the "so what"
questions on the results.
REPORT FORMAT
Every person has a different style of writing. There is not really one right style of writing,
but the following outline is generally accepted as the basis format for most research
projects.
1. Title Page
2. Table of contents
3. Executive summary
4. Introduction
5. Problem statement
6. Research objective
7. Background
8. Methodology
9. Sampling design
10. Research design
11. Data collection
12. Data analysis
13. Limitation
14. Findings
15. Conclusions
16. Summary and conclusions
17. Recommendations
18. Appendices
19. Bibliography
1. Title page
The title page should contain a title which conveys the essence of the study, the date, the
name of the organization submitting the report, and the organization for whom there is
prepared. If the research report is confidential, the name of those individuals to receive
the report should be specified on the title page.
2. Table of contents

As a rough guide, any report of several sections that totals more than 6/to 10 pages
should have a table of contents. The table of contents lists the essence of topics covered
in the report, along with page references. Its purpose is to aid readers in finding a
particular section in the report. If there are many tables, charts, or other exhibits, they
should also be listed after the table of contents in a separate table of illustrations.
3. Executive summary
An executive summary can serve two purposes. It may be a report in miniature covering
all the aspects in the body of the report, but in abbreviated form, or it may be a concise
summary of major findings and conclusions including recommendations.
Two pages are generally sufficient for executive summaries. Write this section after the
report is finished. It must exclude the new information but may require graphics to
present a particular conclusion.
Expect the summary to contain a high density of significant terms since it is repeating
the highlights of report. A good summary should help the decision make and it is
designed to be action oriented.
4. Introduction
The introduction prepares the reader for the report by describing the parts of the
project: the problem statement, research objectives and background material.
The introduction must clearly explain the nature of decision problem. It should review
the previous research done on the problem.
5. Problem statement
The problem statement contains the need for the research project. The problem is
usually represented by a management questions. It is followed by a more detailed set of
objectives.
6. Research objectives
The research objectives address the purpose of the project. These may be research
question(s) and associated investigative questions.
7. Background
The Background material may be of two types. It may be the preliminary results of
exploration from an experience survey, focus group, or another source. Alternatively it
could be secondary data from the literature review. Background material may be placed
before the problem statement or after the research objectives. It contains information
pertinent to the management problem or the situation that led to the study.

8. Methodology
The purpose of the methodology section is to describe the nature of the research design,
the sampling plain, data collection and analysis procedure. Enough details must be
conveyed so that the reader can appreciate the nature the methodology used, yet the
presentation must not be boring overpowering. The use of technical jargon must be
avoided.
9. Research design
The coverage of the design must be adapted to the purpose. The type of research
adapted and reason for adapting that particular type should d be explained.
10. Sampling design
The research explicitly defines the target population being studied and the sampling
methods used. It has to explain the sampling frame, sampling method adapted and
sample size. Explanation of the sampling method, uniqueness of the chosen parameters
or other relevant points that need explanation should be covered with brevity.
Calculation of sample size can be placed either in this part or can be placed in an
appendix.
11. Data collection
This part of report describes the specifics of gathering the data. Its content depends on
the selected design. The data collection instruments (questionnaire or interview
schedule) field instructions can be placed in the appendix.
12. Data analysis
This section summarizes the methods used to analyze the data. Describe data handling,
preliminary analysis, statistical tests, computer programs and other technical
information. The rationale for the choice of analysis approaches should be clear. A brief
commentary on assumptions and appropriateness of use should be presented.
13. Limitations
Every project has weakness, which need to be communicated in a clear and concise
manner. In this process, the researcher should avoid belaboring minor study weakness.
The purpose of this section is not to disparage the quality of the research project, but
rather to enable the reader to judge the validity of the study results.
Generally the limitations will occur in sampling, no response inadequacies and
methodological weakness. It is the researcher's professional responsibility to clearly
inform the reader of these limitations.
14. Findings

The objective of this part is to explain the data rather than draw conclusions. When
quantitative data can be presented, this should be done as simply as possible with
charts, graphics and tables.
The findings can be presented in a small table or chart on the same page. While this
arrangement adds to the bulk of the report, it is convenient for the reader.
15. Conclusions
It can be further divide into two parts as summary and recommendations.
16. Summary and conclusions
The summary is brief statement of the essential findings. The conclusion should clearly
link the research findings with the information needs, and based on this linkage
recommendation for action can be formulated. In some research works the conclusions
were presented in a tabular form for easy reading and reference. The research questions
/objectives will be answered sharply in this part.
17. Recommendations
The researcher's recommendations may be weighed more heavily in favor of the
research findings. There are few ideas about corrective actions. The recommendations
are given for managerial actions rather than research action. The researcher may offer
several alternatives with justifications.
18. Appendices
The purpose of appendix is to provide a place for material, which is not absolutely
essential to the body of the report. The material is typically more specialized and
complex than material presented in the main report, and it is designed to serve the
needs of the technically oriented reader. The appendix will frequently contain copies of
the data collection forms, details of the sampling plan, estimates of statistical error,
interviewer instructions and detailed statistical tables associated with the data analysis
process. The reader who wishes to learn the technical aspects of the study and to look at
statistical breakdowns will want a complete appendix.
19. Bibliography
The use of secondary data requires a Bibliography. Proper citation, style and formats are
unique to the purpose of the report. The instructor, program, institution, or client often
specifies style requirements. It will be given as footnote or endnote format. The author
name, title, publication, year, pager number are the important elements of bibliography.
DATA PRESENTATION
The research data can be presented in Tabular & Graphic form.

Tables
The tabular form consists of the numerical presentation of the data. Tables should
contain the following elements:
1. Table number, this permits easy location in the report
2. Title: the title should clearly indicate the contents of the table or figure.
3. Box head and sub head: the box head contains the captions or labels to the column in
a table, while the sub head contains the labels for the rows.
4. Footnote: footnote explains the particular section or item in the table or figure.

Graphics
The graphical form involves the presentation of data in terms of visually interpreted
sizes. Graphs should contain the following elements:
1. Graph or figure number
2. Title
3. Footnote
4. Sub heads in the axis
Bar chart
A bar chart depicts magnitudes of the data by the length of various bars which have been
laid out with reference to a horizontal or vertical scale. The bar chart is easy to construct
and can be readily interpreted.

Column chart

These graphs compare the sizes and amounts of categories usually for the same time.
Mostly places the categories on X-axis and values on Y-axis.

Pie chart
The pie chart is a circle divided into sections such that the size of each section
corresponds to a portion of the total. It shows the relationship of parts to the whole.
Wedges are row values of data. It is one form of area chart. This type is often used with
business data.

Line Graph
Line graphs are used chiefly for time series and frequency distribution. There are several
guidelines for designing the line graph.

Put the independent variable in the horizontal axis


When showing more than one line, use different line types
Try not to put more than four lines on one chart
Use a solid line for the primary data.

I. Radar diagram
In this the radiating lines are categories; values are distances from the center. It can be
applied where multiple variables used.

II. Area (surface) diagram


An area chart is also used for a time series. Like line charts it compares changing values
but emphasis relative values of each series

III. Scatter diagram


This shows the values if there is relationship between variables follows a pattern; may be
used with one variable at different times.

Purpose of a Histogram
A histogram is used to graphically summarize and display the distribution of a process
data set. The purpose of a histogram is to graphically summarize the distribution of a
uni-variate data set.
The histogram graphically shows the following:
1. center (i.e. the location) of the data;
2. spread (i.e. the scale) of the data;
3. skewness of the data;
4. presence of outliers; and
5. presence of multiple modes in the data.
These features provide strong indications of the proper distributional model for the
data. The probability plot or a goodness-of-fit test can be used to verify the
distributional model.
Sample Bar Chart Depiction

How to construct a Histogram


A histogram can be constructed by segmenting the range of the data into equal sized
bins (also called segments, groups or classes). For example, if your data ranges from 1.1
to 1.8, you could have equal bins of 0.1 consisting of 1 to 1.1, 1.2 to 1.3, 1.3 to 1.4, and so
on.
The vertical axis of the histogram is labeled Frequency (the number of counts for each
bin), and the horizontal axis of the histogram is labeled with the range of your response
variable.
The most common form of the histogram is obtained by splitting the range of the data
into equal-sized bins (called classes). Then for each bin, the numbers of points from the
data set that fall into each bin are counted. That is,
- Vertical axis: Frequency (i.e., counts for each bin)
- Horizontal axis: Response variable
The classes can either be defined arbitrarily by the user or via some systematic rule. A
number of theoretically derived rules have been proposed by Scott (Scott 1992).
The cumulative histogram is a variation of the histogram in which the vertical axis gives
not just the counts for a single bin, but rather gives the counts for that bin plus all bins
for smaller values of the response variable.
Both the histogram and cumulative histogram have an additional variant whereby the
counts are replaced by the normalized counts. The names for these variants are the
relative histogram and the relative cumulative histogram.

There are two common ways to normalize the counts.


1. The normalized count is the count in a class divided by the total number of
observations. In this case the relative counts are normalized to sum to one (or 100 if a
percentage scale is used). This is the intuitive case where the height of the histogram bar
represents the proportion of the data in each class.
2. The normalized count is the count in the class divided by the number of observations
times the class width. For this normalization, the area (or integral) under the histogram
is equal to one. From a probabilistic point of view, this normalization results in a relative
histogram that is most akin to the probability density function and a relative cumulative
histogram that is most akin to the cumulative distribution function. If you want to
overlay a probability density or cumulative distribution function on top of thehistogram, use this normalization. Although this normalization is less intuitive (relative
frequencies greater than 1 are quite permissible), it is the appropriate normalization if
you are using the histogram to model a probability density function.
What questions does Histogram Answer?

What is the most common system response?


What distribution (center, variation and shape) does the data have?
Does the data look symmetric or is it skewed to the left or right?
Does the data contain outliers?

PARETO CHART
Purpose of a Pareto Chart
A Pareto Chart is used to graphically summarize and display the relative importance of
the differences between groups of data. A bar graph used to arrange information in such
a way that priorities for process improvement can be established.
Sample Pareto Chart Depiction

Purposes
- To display the relative importance of data.
- To direct efforts to the biggest improvement opportunity by highlighting the vital few
in contrast to the useful many.
Pareto diagrams are named after Vilfredo Pareto, an Italian sociologist and economist,
who invented this method of information presentation toward the end of the 19th
century. The chart is similar to the histogram or bar chart, except that the bars are
arranged in decreasing order from left to right along the abscissa. The fundamental idea
behind the use of Pareto diagrams for quality improvement is that the first few (as
presented on the diagram) contributing causes to a problem usually account for the
majority of the result. Thus, targeting these "major causes" for elimination results in the
most cost-effective improvement scheme.
How to Construct

Determine the categories and the units for comparison of the data, such as
frequency, cost, or time.
Total the raw data in each category, then determine the grand total by adding the
totals of each category. Re-order the categories from largest to smallest.
Determine the cumulative percent of each category (i.e., the sum of each category
plus all categories that precede it in the rank order, divided by the grand total and
multiplied by 100).
Draw and label the left-hand vertical axis with the unit of comparison, such as
frequency, cost or time.
Draw and label the horizontal axis with the categories. List from left to right in
rank order.

Draw and label the right-hand vertical axis from 0 to 100 percent. The 100
percent should line up with the grand total on the left-hand vertical axis.
Beginning with the largest category, draw in bars for each category representing
the total for that category.
Draw a line graph beginning at the right-hand corner of the first bar to represent
the cumulative percent for each category as measured on the right-hand axis.
Analyze the chart. Usually the top 20% of the categories will comprise roughly
80% of the cumulative total.

Guidelines for Effective applications of Pareto analysis:

Create before and after comparisons of Pareto charts to show impact of


improvement efforts.
Construct Pareto charts using different measurement scales, frequency, cost or
time.
Pareto charts are useful displays of data for presentations.
Use objective data to perform Pareto analysis rather than team members'
opinions.
If there is no clear distinction between the categories - if all bars are roughly the
same height or half of the categories are required to account for 60 percent of the
effect - consider organizing the data in a different manner and repeating Pareto
analysis.

What questions does the Pareto chart answer?

What are the largest issues facing our team or business?


What 20% of sources are causing 80% of the problems (80/20 Rule)?
Where should we focus our efforts to achieve the greatest improvements?

Example of constructing the Pareto Chart


The following table shows the reasons for failure of patients in a Hospital:

Pareto chart for above details can be drawn as follows:

When to Use a Pareto Chart


Pareto charts are typically used to prioritize competing or conflicting "problems," so that
resources are allocated to the most significant areas. In general, though, they can be
used to determine which of several classifications have the most "count" or cost
associated with them. For instance, the number of people using the various ATM's vs.
each of the indoor teller locations, or the profit generated from each of twenty product
lines. The important limitations are that the data must be in terms of either counts or
costs. The data cannot be in terms that can't be added, such as percent yields or error
rates.
PICTOGRAPH
A pictograph is used to present statistics in a popular yet less statistical way to those
who are not familiar with charts that contain numerical scales. This type of chart
presents data in the form of pictures drawn to represent comparative sizes, scales or
areas.
Again as with every chart, the pictograph needs a title to describe what is being
presented and how the data are classified as well as the time period and the source of
the data.
Examples of these types of charts appear below:

A pictograph uses picture symbols to convey the meaning of statistical information.


Pictographs should be used carefully because the graphs may, either accidentally or
deliberately, misrepresent the data. This is why a graph should be visually accurate. If
not drawn carefully, pictographs can be inaccurate
Stem Plots
In statistics, a Stemplot (or stem-and-leaf plot) is a graphical display of quantitative data
that is similar to a histogram and is useful in visualizing the shape of a distribution.
They are generally associated with the Exploratory Data Analysis (EDA) ideas of John
Tukey and the course Statistics in Society (NDST242) of the Open University, although
in fact Arthur Bowley did something very similar in the early 1900s.
Unlike histograms, Stemplots:
- retain the original data (at least the most important digits)
- put the data in order - thereby easing the move order-based inference and nonparametric statistics.
A basic Stemplot contains two columns separated by a vertical line. The left column
contains the stems and the right column contains the leaves. The ease with which

histograms can now be generated on computers has meant that Stemplots are less used
today than in the 1980's, when they first became widely used.
To construct a Stemplot, the observations must first be sorted in ascending order. Here
is the sorted set of data values that will be used in the example:
54 56 57 59 63 64 66 68 68 72 72 75 76 81 84 88 106
Next, it must be determined what the stems will represent and what the leaves will
represent. Typically, the leaf contains the last digit of the number and the stem contains
all of the other digits. In the case of very large or very small numbers, the data values
may be rounded to a particular place value (such as the hundreds place) that will be
used for the leaves. The remaining digits to the left of the rounded place value are used
as the stems.
In this example, the leaf represents the ones place and the stem will represent the rest of
the number (tens place and higher).
The Stemplot is drawn with two columns separated by a vertical line. The stems are
listed to the left of the vertical line. It is important that each stem is listed only once and
that no numbers are skipped, even if it means that some stems have no leaves. The
leaves are listed in increasing order in a row to the right of each stem.

Double Stem Plots (Stem-and-leaf Plot)


Splitting stems and the back-to-back stem plot are two distinct types of double stem
plots, which are a variation of the basic stem plot.
Splitting Stems
On the data set, splitting each of the stems into two or five stems may better illustrate
the shape of the distribution. When splitting stems, it is important to split all stems and
to split the stems equally. When splitting each stem into two stems, one stem contains
leaves from 0-4 and leaves from 5-9 are contained in the other stem. When splitting
each stem into five stems, one stem contains leaves 0-1, the next 2-3, the next 4-5, the

next 6-7, and the last leave 8-9. Here is an example of a split stem plot (using the same
data set from the example above) in which each stem is split into two:

SUMMARY
A research report can be defined as the presentation of the research findings directed to
a specific audience to accomplish a specific purpose.
General guidelines followed to write the report are
1.
2.
3.
4.

Consider the audience,


Be concise, but precise
Be objective, yet effective &
Understand the results and draw conclusions.

The main elements of report are title page, table of contents, executive summary,
introduction, methodology, findings, conclusions, appendices, and bibliography.
Tables and graphs are used for the presentation of data. Different types of graphs are
available like bar chart, pie chart, line chart, area diagram, radar diagram, scatter
diagram, etc. According to the nature of data and requirement the type of graph can be
selected and used effectively.
KEY TERMS

Executive summary
Sampling
Bibliography
Appendix
Interview schedule
Area chart
Line graph
Bar chart
Pie chart
Scatter diagram
Radar diagram
Surface diagram
Pareto Chart
Pictograph
Stem-graph

IMPORTANT QUESTIONS
1.
2.
3.
4.
5.
6.
7.
8.
9.

What do you mean by research report?


Why is the research report important?
Explain the general guidelines exist for writing a report?
What are the preparations required for writing the report?
What components are typically included in a research report?
What are the alternative means of displaying data graphically?
Explain the application of Pareto Chart with example
What are the applications of Pictograph?
What are the procedures to draw Stem Graph or Stem Plot?

REFERENCE BOOKS
1. Ramanuj Majumdar, Marketing research, Wiley Estern Ltd., New Delhi, 1991.
2. Harper W Boyd, Jr etal, Marketing Research, Richard D. Irevin Inc. USA 1990.
3. Paul E. Green et al, Research for Marketing Decisions, Prentice Hall of India Pvt.
Ltd., New Delhi, 2004.

- End of Chapter -

You might also like