You are on page 1of 5

Role of computers in research

Computers in Research
The computers are indispensable throughout the research process. The role of computer becomes more important
when the research is on a large sample. Data can be stored in computers for immediate use or can be stored in
auxiliary memories like floppy discs, compact discs, universal serial buses (pen drives) or memory cards, so that
the same can be retrieved later. The computers assist the researcher throughout different phases of research
process.
Role of Computers in the phases of research process
There are five major phases of the research process where computer plays different vital roles. They are:
1) Role of Computer in Conceptual phase
2) Role of Computer in Design and planning phase
3) Role of Computer in Empirical phase
4) Role of Computer in Analytic phase and
5) Role of Computer in Dissemination phase
1) Role of Computer in Conceptual Phase
The conceptual phase consists of formulation of research problem, review of literature, theoretical frame work and
formulation of hypothesis.
Role of Computers in Literature Review: Computers help for searching the literatures (for review of literature)
and bibliographic references stored in the electronic databases of the world wide webs. It can thus be used for
storing relevant published articles to be retrieved whenever needed. This has the advantage over searching the
literatures in the form of books, journals and other newsletters at the libraries which consume considerable amount
of time and effort.
2) Role of Computers in Design and planning phase Design and planning phase consist of research design,
population, research variables, sampling plan, reviewing research plan and pilot study.
Role of Computers for Sample Size Calculation: Several software’s are available to calculate the sample size
required for a proposed study. NCSS-PASS-GESS is such software. The standard deviation of the data from the
pilot study is required for the sample size calculation.
3) Role of Computers in Empirical phase
Empirical phase consist of collecting and preparing the data for analysis.
Data Storage: The data obtained from the subjects are stored in computers as word files or excel spread sheets.
This has the advantage of making necessary corrections or editing the whole layout of the tables if needed, which
is impossible or time-consuming in case of writing in papers. Thus, computers help in data entry, data editing, data
management including follow up actions etc. Computers also allow for greater flexibility in recording the data
while they are collected as well as greater ease during the analysis of these data.
In research studies, the preparation and inputting data is the most labour-intensive and time consuming aspect of
the work. Typically the data will be initially recorded on a questionnaire or record form suitable for its acceptance
by the computer. To do this the researcher in conjunction with the statistician and the programmer, will convert the
data into Microsoft word file or excel spread sheet. These spread sheets can be directly opened with statistical
software’s for analysis.
4) Role of Computers in Data Analysis
This phase consist of statistical analysis of the data and interpretation of results.
Data Analysis: Much software is now available to perform the ‘mathematical part ‘of the research process i.e. the
calculations using various statistical methods. Software’s like SPSS, NCSS-PASS, STATA and Sysat are some of
the widely used. They can be like calculating the sample size for a proposed study, hypothesis testing and
calculating the power of the study. Familiarity with any one package will suffice to carry out the most intricate
statistical analyses.
Computers are useful not only for statistical analyses, but also to monitor the accuracy and completeness of the
data as they are collected.
5) Role of Computers in Research Dissemination
This phase is the publication of the research study.
Research publishing: The research article is typed in word format and converted to portable data format (PDF)
and stored and/or published in the World Wide Web.
Use of Computer in Data Processing and Tabulation
Research involves large amounts of data, which can be handled manually or by computers. Computers provide the
best alternative for more than one reason. Besides its capacity to process large amounts of data, it also analyses
data with the help of a number of statistical procedures. Computers carry out processing and analysis of data
flawlessly and with a very high speed. The statistical analysis that took months earlier takes now a few seconds or
few minutes. Today, availability of statistical software and access to computers has increased substantially over
the last few years all over the world. While there are many specialised software application packages for different
types of data analysis, Statistical Package for Social Sciences (SPSS) is one such package that is often used by
researchers for data processing and analysis. It is preferred choice for social work research analysis due to its easy
to use interface and comprehensive range of data manipulation and analytical tools.
Basic Steps in Data Processing and Analysis
There are four basic steps involved in data processingand analysis using SPSS. They are:
1) Entering of data into SPSS,
2) Selection of a procedure from the Menus,
3) Selection of variables for analysis, and
4) Examination of the outputs.
You can enter your data directly into SPSS DataEditor. Before data analysis, it is advised that youshould have a
detailed plan of analysis so that youare clear as to what analysis is to be performed.Select the procedure to work on
the data. All the variables are listed each time a dialog box is opened.Select variables on which you wish to apply a
statisticalprocedure. After completing the selection, executethe SPSS command. Most of the commands are
directly executed by clicking ‘O.K’. on the dialog box. The processor in the computer will execute the procedures
and display the results on the monitor as ‘output file’.
Conclusion
To conclude, computers are useful tools that make the research process easier and faster with accuracy and greater
reliability and fewer errors. The programmer or the computer operator should have a thorough knowledge about
the abilities and disabilities of the software’s used for better use of computers.
The data, after collection has to be processed and analysed in accordance with the outline laid down for the
purpose at the time of developing the research plan. This is essential for a scientific study and for ensuring that we
have all relevant data making contemplated comparision and analysis Technically speaking processing implises
editing, coding, classification, and tabulation of collected data so that they are amenable to analysis The term
analysis refers to the computation of certain measures along with searching for patterns of relationship that exit
among data group.
Data processing
Data processing is, broadly, "the collection and manipulation of items of data to produce meaningful
informations.”In this sense it can be considered a subset of information processing, "the change (processing) of
information in any manner detectable by an observer." 
Processing operations
We can now proceed with the explanation of all the processing operations.
Editing
Editing of data is a process of examining the collected raw material (especially in surveys) to detect errors and
omissions and to correct these when possible. As a matter of fact, editing involves a careful scruitiny of the
completed questionnaires and or schedules.
Types of editing
Editing may be in the form of field editing or central editing
a) Field editing: - field editing consists in the review of the reporting forms by the investigator for
completing what the letter has written in abbreviated and or in illegible form at the time of recording the
respondent’s responses.
b) Central editing: - central editing should take place when all forms or schedules have been completed and
returned to the office. This type of editing implies that all form should get a through editing by a single
editor in a small study and by a team of editors in case of large of inquiry.
Some points of editing
Editors must keep in view several points while performing their editing work which are as follows
a) They should be familiar with instructions given to the interviewers and coders as well as with the editing
instructions supplied to them.
b) They must make entries on the in some distinctive colour and that too in a standardised form.
Coding
Coding refers to the process of assigning numerals or other symbols to answers so that responses can be put into a
limited number of categories or classes. Coding is necessary for efficient analysis and through it the several replies
may be reduced to a small number of classes which contain the critical information required for analysis.
Coding is translating answers into numerical values or assigning numbers to the various categories of a variable to
be used in data analysis. Coding is done by using a code book, code sheet, and a computer card. Coding is done on
the basis of the instructions given in the codebook. The code book gives a numerical code for each variable.
Now-a-days, codes are assigned before going to the field while constructing the questionnaire/schedule. Pose data
collection; pre-coded items are fed to the computer for processing and analysis. For open-ended questions,
however, post-coding is necessary. In such cases, all answers to open-ended questions are placed in categories and
each category is assigned with a code.
Manual processing is employed when qualitative methods are used or when in quantitative studies, a small sample
is used, or when the questionnaire/schedule has a large number of open-ended questions, or when accessibility to
computers is difficult or inappropriate. However, coding is done in manual processing also.
Classification
Most research studies result in a large volume of raw data which must be reduced into homogeneous group if we
are to get meaning full relationships. The classification arranged on the basis of common characteristics.
Types of classification
Classification of two types
a) Classification according to attributes
b) Classification according to class intervals
Tabulation
When a mass of data has been assembled, it becomes necessary for the researcher to arrange the same in some
kind of concise and logical order .This procedure is referred to as tabulation. Thus tabulation is the process of
summarising raw data and displaying the same in compact form (i.e. in the form of statistical tables) in a broader
sense, tabulation is an orderly arrangement of data in columns and row.
After editing, which ensures that the information on the schedule is accurate and categorized in a suitable form, the
data are put together in some kinds of tables and may also undergo some other forms of statistical analysis.

Table can be prepared manually and/or by computers. For a small study of 100 to 200 persons, there may be little
point in tabulating by computer since this necessitates putting the data on punched cards. But for a survey analysis
involving a large number of respondents and requiring cross tabulation involving more than two variables, hand
tabulation will be inappropriate and time consuming.
Importance of tabulation
Tabulation is essential because of the following reasons:-
a) It facilitates the process of comparison.
b) It provides a basis for various statistical computation
c) The present an overall view of findings in a simpler way.
d) They identify trends.
e) They display relationships in a comparable way between parts of the findings.
f) By convention, the dependent variable is presented in the rows and the independent variable in the
columns.
Principles of tabulation
a) Every table should be given a distinct number to facilitate easy reference.
b) The column headings and the row headings of the table should be clear and brief.
c) The columns may be numbered to facilitate reference.
d) Total of row should normally be placed in the extreme right column and that of columns should be placed
at the bottom.
Some problems in processing
We can take up the following two problems of processing the data for analysis purpose.
a) The problem concerning “don’t know” (or DK) responses: - while processing the data the researcher
often comes across some responses that are difficult to handle. One category of such responses may be
„Don’t know response or simply DK response. When the DK response group is small it is of little
significance but when it is relatively big it becomes a matter of major concern in which the case the
question arise.
b) Use of percentage – percentages are often used in data presentation for reducing all of them to a 0 to 100
range.
Rules of percentage
a) Two or more percentages must not be averaged unless each is weighted by the group size from which it has
been derived.
b) Use of too large percentage should be avoided since a large percentage is difficult to understand and tends
to confuse, defeating the very purpose for which percentage are used.
c) Percentage hide the base from which they have been computed. If this is not kept in view the real
difference may not be correctly read.
Data analysis
Analysis of data is a process of inspecting, cleaning, transforming, and modeling data with the goal of discovering
useful information, suggesting conclusions, and supporting decision making. Data analysis has multiple facets and
approaches, encompassing diverse techniques under a variety of names, in different business, science, and social
science domains. As stated earlier, by analysis we mean the computation of certain indices or measures along with
searching for patterns of relationship that exist among the data groups. Data analysis is an ongoing activity, which not
only answers your question but also gives you the directions for future data collection. Data analysis procedures (DAP) help
you to arrive at the data analysis. The uses of such procedures put your research project in perspective and assist you in
testing the hypotheses with which you have started your research. Hence with the use of DAP, you can

a) convert data into information and knowledge, and


b) explore the relationship between variables.
Types of analysis
Analysis may, therefore be categorized as descriptive analysis and inferential analysis.
a) Descriptive analysis – descriptive analysis is largely the study of distribution of one variable .this study
provides us with profiles of companies, work groups, persons and other subjects on any of a multiple of
characteristics such as size , composition , efficiency , preferences.
b) Inferential analysis- is concerned with the various tests of significance for testing hypothesis in order to
determine with what validity data can be said to indicate some conclusion or conclusions. It is also
concerned with the estimation of population values.

You might also like