Professional Documents
Culture Documents
Research Methods
The research process consists of the steps that need to be followed
in order to carry out research effectively. The most salient aspects
of the research process are listed below:
Develop a research problem
Carry out an extensive literature review
Develop a hypothesis or research question
Compose an appropriate research and sample design
Collect data and conduct analyses
Test the hypothesis
Interpret and Discuss
Make conclusions based on the data
Research Methodology
The above concept is defined as the science behind the methods
used to gain knowledge. In other words, methodology is the study
of the methods used and the reasoning behind why those
particular methods were used. It is a way in which to
systematically solve the research problem (i.e., analyse the logic
behind the steps taken by a researcher to answer said research
question). The methodology section in any research serves the
purpose of explaining the ways in which results were obtained
(i.e., the research methods that were employed and the way in
which results were analysed in order to allow the reader to
critically evaluate the research methods). A research methodology
provides an all encompassing theoretical and philosophical
framework that is used at the beginning of a project to explain the
functioning and reasoning behind chosen research methods, as
well as to guide the research process. More importantly, the
methodology for any research project is pivotal to maintaining
reliable research methods and results, which adds to the value of
the findings and interpretations[v]. Methodology seeks
consideration of the following aspects:
finding the suitable research method for the chosen problem,
discovering the accuracy of the results of a chosen method, and
ensuring the efficiency of the research method.
Methods
Are defined as the methods or techniques that are used to gather evidence
and conduct research.
Conclusion
A methodology is required in order to systematically solve the
research problem by analysing the logic behind different research
methods implemented. The use of a clear methodology renders
that are reliable, reproducible, and correct. In order to produce
research that is systematic, logical, and replicable, researchers
require an in-depth knowledge of research methodology.
---------------------------------------------------------------------------------------------------------------
What is Reliability?
The idea behind reliability is that any significant results must be more than a one-off finding and be
inherently repeatable.
Other researchers must be able to perform exactly the same experiment, under the same conditions
and generate the same results. This will reinforce the findings and ensure that the wider scientific
community will accept the hypothesis.
Without this replication of statistically significant results, the experiment and research have not
fulfilled all of the requirements of testability.
This prerequisite is essential to a hypothesis establishing itself as an accepted scientific truth.
For example, if you are performing a time critical experiment, you will be using some type of
stopwatch. Generally, it is reasonable to assume that the instruments are reliable and will keep true
and accurate time. However, diligent scientists take measurements many times, to minimize the
chances of malfunction and maintain validity and reliability.
At the other extreme, any experiment that uses human judgment is always going to come under
question.
For example, if observers rate certain aspects, like in Bandura’s Bobo Doll Experiment, then the
reliability of the test is compromised. Human judgment can vary wildly between observers, and the
same individual may rate things differently depending upon time of day and current mood.
This means that such experiments are more difficult to repeat and are inherently less reliable.
Reliability is a necessary ingredient for determining the overall validity of a scientific experiment and
enhancing the strength of the results.
Debate between social and pure scientists, concerning reliability, is robust and ongoing.
What is Validity?
Validity encompasses the entire experimental concept and establishes whether the results obtained
meet all of the requirements of the scientific research method.
For example, there must have been randomization of the sample groups and appropriate care and
diligence shown in the allocation of controls.
Internal validity dictates how an experimental design is structured and encompasses all of the steps
of the scientific research method.
Even if your results are great, sloppy and inconsistent design will compromise your integrity in the
eyes of the scientific community. Internal validity and reliability are at the core of any experimental
design.
External validity is the process of examining the results and questioning whether there are any other
possible causal relationships.
Control groups and randomization will lessen external validity problems but no method can be
completely successful. This is why the statistical proofs of a hypothesis called significant, not
absolute truth.
Any scientific research design only puts forward a possible cause for the studied effect.
There is always the chance that another unknown factor contributed to the results and findings. This
extraneous causal relationship may become more apparent, as techniques are refined and honed.
Conclusion
If you have constructed your experiment to contain validity and reliability then the scientific
community is more likely to accept your findings.
Eliminating other potential causal relationships, by using controls and duplicate samples, is the best
way to ensure that your results stand up to rigorous questioning.
What is an Experiment?
From the Latin word, “experior” which means “to attempt” or “to
experience”, experiment is defined as testing a hypothesis by
carrying out a procedure under highly controlled conditions. This
makes the method ideal in studying primary data. By
manipulating a certain independent variable, its effect on a
dependent variable can be measured. A cause and effect
relationship is verified by exposing participants to certain
treatments. For instance, researchers can measure how water
intake can affect people’s metabolism by letting the experimental
group drink 8 glasses of water each day while the control group
will only have 4 glasses. Their metabolism rates will then be
compared after a week and statistical treatments like T-test will be
employed to validate the results.
What is a Survey?
From the medieval Latin word, “supervidere” which means “to
see”, survey is defined as having a comprehensive view of certain
topics. Survey studies are largely conducted to look into people’s
opinions, feelings, and thoughts. It is best suited for descriptive
research which seeks to answer “what” questions regarding the
respondents. Questionnaires are ideal in collecting information
from a big population as they can be simultaneously administered
to different groups and individuals. Survey questions can be sent
to numerous respondents in both online and offline settings. For
instance, researchers who are studying happiness levels among
millennials floated questionnaires, made phone calls, and sent e-
mails regarding the participants’ perceived emotional states. The
data were then collated and statistical treatment such as getting
the weighted mean was utilized to analyze the responses.
Difference between
Experiment and Survey
1. Etymology of Experiment and Survey
Experiment came from the Latin word “experior” which means “to
attempt” or “to experience” while survey came from the Latin
word “supervidere” which means “to see”.
3. Data
6. Fields
10. Challenge
12. Manipulation
13. Relationships
15. Randomization
----------------------------------------------------------------------------------------
Multistage sampling divides large populations into stages to make the sampling process more practical. A
combination of stratified sampling or cluster sampling and simple random sampling is usually used.
Let’s say you wanted to find out which subjects U.S. school children preferred. A population list — a list of
all U.S. schoolchildren– would be near-impossible to come by, so you cannot take a sample of the
population. Instead, you divide the population into states and take a simple random sample of states. For
the next stage, you might take a simple random sample of schools from within those states. Finally you
could perform simple random sampling on the students within the schools to get your sample.
In order to classify multistage sampling as probability sampling, each stage must involve a probability
sampling method.
A pilot survey is a strategy used to test the questionnaire using a smaller sample compared to the
planned sample size. In this phase of conducting a survey, the questionnaire is administered to a
percentage of the total sample population, or in more informal cases just to a convenience sample.
Conducting a pilot survey prior to the actual, large-scale survey presents many benefits and
advantages for the researcher. One of these is the exploration of the particular issues that may
potentially have an antagonistic impact on the survey results. These issues include the
appropriateness of questions to the target population.
A pilot survey also tests the correctness of the instructions to be measured by whether all the
respondents in the pilot sample are able to follow the directions as indicated. It also provides better
information on whether the type of survey is effective in fulfilling the purpose of the study. Practically
speaking, pilot surveys save financial resources because if errors are found in the questionnaire or
interview early on, there would be a lesser chance of unreliable results or worse, that you would
need to start over again after conducting the survey.
All in all, the main objective of a pilot study is to determine whether conducting a large-scale survey
is worth the effort.
Read also: The Pilot Study.
Types of Pilot Survey
A. According to Organization
There are two types of pilot survey according to organization – external and internal. An external
pilot survey intends to administer the questionnaire to a small group of target participants who will
not be included in the main survey. On the other hand, an internal pilot survey will consider the
respondents in the pilot as the first participants in the main survey.
There are two types of pilot survey according to the participation of the respondents – undeclared
and participatory. In an undeclared pilot survey, you administer the survey to a certain number of
respondents as if it is the real and full scale survey, not a pretest one. On the other hand,
participatory pilot surveys involve informing the respondents that they are in the pre-test phase. The
respondents are to be asked what they can say about the questionnaire, specifically their reactions,
comments and suggestions. For instance, you may ask them about how clear the instructions are or
which questions are hard to answer. Converse and Presser (1986) recommend using the
participatory pilot survey first, and then conducting the undeclared pilot.
After obtaining and analyzing the results of the pilot survey, logistical, technical and other issues or
problems can be addressed. The questionnaire or interview format can be revised, or the type of
survey may be altered into a more suitable one. After the revision of the survey, the researcher may
opt to conduct a second pilot survey to determine whether the errors and issues are effectively
solved. If the problems were minor, then the large-scale survey can be executed.
Both Statistical as well Mathematical models involve mathematical formulas and equations
but this this not mean that both are the same thing. Mathematical models are kind of static
model that represent a natural/real phenomenon in mathematical form; the models once
formulated does not necessitate to change the form after they are formed. However
statistical or data based models are enough flexible to change as per arrival of new data as
they can incorporate new and emerging patterns and trends; this is where pattern
recognition/machine learning come into focus.
Mathematical models are generally exact and given by equations, where statistical models
are generally given by data and the probability of the dataset matching the statistical
model. There are usually blurs between the two types of models.
Mathematical models determine how the system changes from one state to the next and
describes the interdependence of the variables (or factors) involved. Statistical models
characterize the numerical data and attempts at estimating probabilistic future behavior of a
system based on its past behavior.
Analysis refers to breaking a whole into its separate components for individual examination. Data
analysis is a process for obtaining raw data and converting it into information useful for decision-
making by users. Data is collected and analyzed to answer questions, test hypotheses or disprove
theories.[3]
Statistician John Tukey defined data analysis in 1961 as: "Procedures for analyzing data, techniques
for interpreting the results of such procedures, ways of planning the gathering of data to make its
analysis easier, more precise or more accurate, and all the machinery and results of (mathematical)
statistics which apply to analyzing data."[4]
There are several phases that can be distinguished, described below. The phases are iterative, in
that feedback from later phases may result in additional work in earlier phases.[5] The CRISP
framework used in data mining has similar steps.
Data requirements[edit]
The data are necessary as inputs to the analysis, which is specified based upon the requirements of
those directing the analysis or customers (who will use the finished product of the analysis). The
general type of entity upon which the data will be collected is referred to as an experimental unit
(e.g., a person or population of people). Specific variables regarding a population (e.g., age and
income) may be specified and obtained. Data may be numerical or categorical (i.e., a text label for
numbers).[5]
Data collection[edit]
Data are collected from a variety of sources. The requirements may be communicated by analysts to
custodians of the data, such as information technology personnel within an organization. The data
may also be collected from sensors in the environment, such as traffic cameras, satellites, recording
devices, etc. It may also be obtained through interviews, downloads from online sources, or reading
documentation.[5]
Data processing[edit]
The phases of the intelligence cycle used to convert raw information into actionable intelligence or knowledge
are conceptually similar to the phases in data analysis.
Data initially obtained must be processed or organised for analysis. For instance, these may involve
placing data into rows and columns in a table format (i.e., structured data) for further analysis, such
as within a spreadsheet or statistical software.[5]
Data cleaning[edit]
Once processed and organised, the data may be incomplete, contain duplicates, or contain errors.
The need for data cleaning will arise from problems in the way that data are entered and stored.
Data cleaning is the process of preventing and correcting these errors. Common tasks include
record matching, identifying inaccuracy of data, overall quality of existing data,[6] deduplication, and
column segmentation.[7] Such data problems can also be identified through a variety of analytical
techniques. For example, with financial information, the totals for particular variables may be
compared against separately published numbers believed to be reliable.[8] Unusual amounts above
or below pre-determined thresholds may also be reviewed. There are several types of data cleaning
that depend on the type of data such as phone numbers, email addresses, employers etc.
Quantitative data methods for outlier detection can be used to get rid of likely incorrectly entered
data. Textual data spell checkers can be used to lessen the amount of mistyped words, but it is
harder to tell if the words themselves are correct.[9]
Data product[edit]
A data product is a computer application that takes data inputs and generates outputs, feeding them
back into the environment. It may be based on a model or algorithm. An example is an application
that analyzes data about customer purchasing history and recommends other purchases the
customer might enjoy.[5]
Communication[edit]