3 5 PDF
3 5 PDF
3 5 PDF
Research Methodology
Lesson 1 to 7
Research Methodology
By: Prof. O.P.
Lesion 1-7Verma
1
CONTENTS
2
Syllabus
MC 3.5 RESEARCH METHODOLOGY
Max Marks 80 Internal Assessment 20 Note: There will be nine (9) questions in
all. The first question is compulsory and consists of ten (10) short-questions having two (2)
marks each. The candidate will be required to attempt one question from each unit and each
question carries fifteen (15) marks
Course Contents:
Unit I:
Unit II:
Data collection and processing: Meaning of data, types of data; secondary data--advantages,
disadvantages, criteria for evaluating secondary data, secondary in Indian context; Primary
data—meaning, methods of collection of primary data- survey, observation, case study and
experiment; Questionnaire—types, guidelines for preparing questionnaire, steps in designing
the questionnaire, essential of good questionnaire and difference between schedule and
questionnaire; Processing of data—editing, coding, classification, tabulation, and
presentation by way of diagrams and graphs.
Unit III:
Unit IV:
Hypothesis, Data analysis and Report writing: Hypothesis- meaning, characteristic, sources,
hypothesis testing procedure, Type-I and Type- II errors, Application of mean, dispersion,
skewness, kurtosis, Correlation and regression in research; Application of Z-test, t-test, F-test,
Chi-square test and ANOVA; Introduction to SPSS-- data entry and descriptive statistics,
Report writing; Significance of Report-Writing; Steps in Report Writing, Layout of the
3
Research Report; Types of Report, Mechanics of Writing a Research Report; Precautions for
writing Research Report.
Books Recommended:
Research Methodology for Management with SPSS, Majhi & Khatua, HPH
4
LESSON-I
Structure:
Objectives
Introduction
Objectives of Research
Types of Research
Approaches to Research
Research Methods
Research Process
Self Assessment
Summary
Glossary
Terminal Questions
Suggested Readings
OBJECTIVES:
5
• know about the objectives, need and scope of research,
INTRODUCTION:
Research is an intellectual activity which promotes advancement of knowledge which
in turn helps to modernize the society. Research comprises "creative work undertaken on a
systematic basis in order to increase the stock of knowledge, including knowledge of man,
culture and society, and the use of this stock of knowledge to devise new applications."
Research is any such activity which helps to gain fresh insight into something. It is used to
establish or confirm facts, reaffirm the results of previous work, solve new or existing
problems. Research is not merely an enquiry, it is actually a scientific enquiry. It is carried
out scientifically or systematically or critically. It aims at seeking knowledge and endeavours
to discover new facts or principles. It is continuous search for truth.
The term research has different connotations in different contexts. It is, therefore
essential to study and examine various definitions of research given by different authorities.
The Webster’s International Dictionary defines research as “a careful critical enquiry or
examination in seeking facts for principles, diligent investigation in order to ascertain
something.” Research is an academic activity and as such the term should be used in a
technical sense. According to Clifford Woody research comprises defining and redefining
problems, formulating hypothesis or suggested solutions; collecting, organising and
evaluating data; making deductions and reaching conclusions; and at last carefully testing the
conclusions to determine whether they fit the formulating hypothesis. Encyclopaedia of
Social Sciences define research as “the manipulation of things, concepts or symbols for the
purpose of generalising to extend, correct or verify knowledge, whether that knowledge aids
in construction of theory or in the practice of an art.” P.M cook outlines the term “Research
“in a popular way. To him, “research” is an honest, exhaustive intelligent searching for facts
and their meanings or implications, with reference to a problem. In an easy to remember
method, he lists the attributes of research” taking the clue from each initial alphabet of the
word:
6
E - Expert and exhaustive treatment
E - Exactness
A - Analysis
R - Relationship of facts
Research is, thus, an original addition to the available knowledge which contributes to
the further advancement. It is an attempt to pursue truth through the methods of study,
observation, comparison and experiment.
OBJECTIVES OF RESEARCH:
The objective of research is to find out answers to questions through the application of
systematic and scientific way.
A good research is objective in the sense that it must answer the research questions.
This necessitates the formulation of a proper hypothesis, otherwise there may be a lack of
congruence between the research questions and the hypothesis.
• Control
7
A good research must be able to control all the variables. This requires randomization
at all stages, e.g., in selecting the subjects, the sample size and the experimental treatments.
This shall ensure an adequate control over the independent variables.
• Generaliability
We should be able to have almost the same results by using an identical methodology
so that we can apply the results to similar situations.
A good research should be free from the researcher’s personal biases and must be
based on objectivity and not subjectivity.
• Systematic
A good research study must have various well planned steps, i.e., all steps must be
interrelated and one step should lead to another step.
Business research assists decision maker’s shift from intuitive information gathering
to organized and objective study. Even though researchers in different functional fields may
examine different phenomena, they are comparable to each other simply because they make
use of similar research techniques. Research is the fountain of knowledge for the sake of
knowledge and it is a crucial source of providing guidelines for solving various business
issues. Thus, we can say that the scope of research is enormous.
• Business research is almost always applied research. This means that it is problem-
oriented with the objective of obtaining information to help solve a specific business
problem or make a decision.
• Business research tends to have a time limit. It makes little sense to gather
information about a decision that was made three weeks ago or to get information
next week that we need today. Information is highly perishable. It gets old and useless
very quickly. Like fish, if you don’t consume it when fresh, it loses a lot of its flavour—
and starts to smell bad!
8
• Business research is done in an environment in which conditions change rapidly. The
answers to research questions asked today may very well differ from the answers to
the same questions obtained yesterday. In those areas where the manager needs to
“stay in tune” with the changing landscape, the same research questions might need to
be asked repeatedly. Successful organizations develop a culture where constant
research is an integral part of their operation.
• Business is a for-profit enterprise, so research must have some cost limits applied.
Research can be expensive and time consuming (although many economical
approaches can be used). If faced with a $10 problem, don’t waste $1,000 on research.
Likewise, a million-dollar decision must have a bigger budget for research than a
$100 decision.
Production Management
Personnel Management
Marketing Management
Research performs an important part in choice and size of target market, the consumer
behaviour with regards to attitudes, life style, and influences of the target market. It is the
primary tool in determining price policy, selection of channel of distribution and
development of sales strategies, product mix, promotional strategies, etc.
Financial Management
Materials Management
9
It is utilized in choosing the supplier, making the decisions relevant to make or buy as
well as in selecting negotiation strategies.
General Management
TYPES OF RESEARCH:
Basic types of research are as follows:
Basic or pure research is conducted solely for the purpose of gathering information. In
basic research, general theories, ideas, and questions are explored and tested. It is a type of
research that involves investigating theoretical issues to add to the scientific knowledge base.
While this type of research contributes to our understanding of the human mind and
behaviour, it does not necessarily help to solve immediate practical problems
Applied Research
Applied research refers to scientific study and research that seeks to solve practical
problems. Applied research is used to find solutions to everyday problems, cure illness, and
develop innovative technologies. Psychologists working in human factors or industrial/
organizational fields often do this type of research.
Descriptive Research
Historical Research
Historical research is the type of research that examines past events or combinations
of events to arrive at an account of what has happened in the past. It is the research on past
social forces which have shaped the present. For example to study the present state of Indian
labour we may research on past historical forces.
10
Formulative or Exploratory Research
Exploratory research is an initial research which analyzes the data and explores the
possibility of obtain ing as many relationships as possible between different variables without
knowing their end-applications. This means that a general study will be conducted without
having any specific end-objective except to establish as many relationships as possible
between the variables of the study. Exploratory research helps us to investigate any problem
with suitable hypothesis. This research on social science is particularly important for
clarification of any concept and throwing new light for further research on principles of
developing hypothesis and it’s testing with statistical tools.
Experimental Research
An experiment research is used to study the effect of a set of factors on the response
variable of a system of study. This research is conducted in a controlled environment.
Although experimental research is primarily possible in areas of physical science, with the
help of hypothesis, may also be carried out in social science if such research enables us to
quantify the findings, to apply the statistical and mathematical tools and to measure the
results thus quantified.
Ex-Post-Facto Research
APPROACHES TO RESEARCH:
In the world of research, there are two general approaches to gathering and reporting
information: qualitative and quantitative approaches. The qualitative approach to research is
focused on understanding a phenomenon from a closer perspective. The quantitative
approach tends to approximate phenomena from a larger number of individuals using survey
methods. Each approach has its benefits and demerits, and is more suitable to answering
certain kinds of questions.
Qualitative Approach
11
this type of research, because to carry out such a research endeavor requires many resources
and much time. Interviews can vary from being highly structured and guided by open-ended
questions, or be less structured and take the form of a conversational interview. Because of
the investment in this type of research and the relatively few number of participants, findings
from qualitative research cannot be generalized to the whole population. However, such
research serves as a spring board for larger studies and deeper understanding that can inform
theory, practice, and specific situations.
• May reveal information that would not be identified through pre-determined survey
questions
Limitations
Quantitative Approach
12
example, the same survey can be used with a group of mentors that is receiving training
(often called the intervention or experimental groups) and a group of mentors who does not
receive such a training (a control group). It is then possible to compare these two groups on
outcomes of interest, and determine what influence the training had. It is also relatively easy
to survey people a number of times, thereby allowing the conclusion that a certain features
(like matching) influence specific outcomes (well-being or achievement later in life).
Using survey methods across a large group of individuals enables generalization. For
example, if policy makers wanted to instantiate a policy about mentor training, they would
likely require some evidence that this training actually works. Interviewing a few individuals,
or conducting a focus group with forty matches, might be reflective of specific cases in which
the mentoring training worked, however, it would not provide strong evidence that such
training is beneficial overall. Stronger support for successful training would be evident if
using quantitative methods.
Limitations
13
RESEARCH METHODS AND RESEARCH
METHODOLOGY:
• Research methods are the methods or techniques employed by researchers in
conducting research operations.
14
change occurring in the country. Research assumes a significant role in solving various
operational and planning problems associated with business and industry. In several ways,
operations research, market research, and motivational research are vital and their results
assist in taking business decisions.
Research is equally important to the social scientist for analyzing social relationships
and seeking explanation to various social problems. It gives intellectual satisfaction of
knowing things for the sake of knowledge. It also possesses practical utility for the social
scientist to gain knowledge so as to be able to do something better or in a more efficient
manner.
RESEARCH PROCESS:
Research process comprises a series of steps or actions required for effectively
conducting research and for the sequencing of these steps. The following are the various steps
that provide useful procedural guideline regarding the conduct of research.
(7) Hypothesis-testing
Problem Definition
The first and foremost step in the research process consists of problem or opportunity
identification. The research problem must be identified and defined without any ambiguity.
There are two types of research problems, viz., those which relate to state of nature and those
which relate to relationships between variables. At the very outset the researcher must single
15
out the problem he wants to study, i.e., he/she must decide the general area of interest or
aspect of a subject- matter that he/she would like to inquire into.
There may be situation in which the researcher is fully aware of the symptoms
relating to certain deficiency in achieving an organizational goal. But, he/she may not be in a
position to clearly spell out the problem which is causing such deficiency. Unless it is clearly
identified, it will not be possible to proceed further to carry out the project. If a researcher
proceeds with ill-defined problems, he/she may end up with misleading conclusions or
aborting the research project in the middle due to poor pay-off identified through interim
evaluation of the research. Hence, the research problem should be clearly defined.
Now that the problem has been identified, the researcher must learn more about the
topic under investigation. To do this, the researcher must review the literature related to the
research problem. This step provides foundational knowledge about the problem area. The
review of literature also educates the researcher about what studies have been conducted in
the past, how these studies were conducted, and the conclusions in the problem area. For this
purpose, the abstracting and indexing journals and published or unpublished bibliographies
are the first place to go to. In this process, it should be remembered that one source will lead
to another. The earlier studies, if any, which are similar to the study in hand, should be
carefully studied.
After extensive literature survey, researcher should state in clear terms the working
hypothesis or hypotheses. A hypothesis is an assumption or suggested explanation about how
two or more variables are related. It is a crucial step in the scientific method and, therefore, a
vital aspect of all scientific research. Working hypothesis is tentative assumption made in
order to draw out and test its logical or empirical consequences. As such the manner in which
research hypotheses are developed is particularly important since they provide the focal point
for research. Hypothesis should be very specific and limited to the piece of research in hand
because it has to be tested. The role of the hypothesis is to guide the researcher by delimiting
the area of research and to keep him on the right track.
Research Design
The next step in research process is to design research. The research design provides a
complete guideline for data collection. It is the plan, structure and strategy of investigation
16
conceived so as to obtain answers to research questions and to control variance. Following
are the essence of research design:
• Design of experiment
• Design of questionnaire
A sampling plan is mechanism by which the sampling units of a study are selected
from the sampling frame of the population. The selection of the sampling plan in a study in
turn affects the cost and time to conduct the study, and the reliability of interference of the
study. Hence, it should be selected with utmost care. The sampling plan can be classified into
probability sampling plans and non-probability sampling plans. Different sampling plans in
each of these categories are listed as follows:
• Systematic sampling
• Cluster sampling
• Multi-stage sampling
• Convenience sampling
• Judgment sampling
• Quota sampling
Design of experiment
17
• Identify the response variables of the study.
• Decide on the type of each of the factors (a factor may be either fixed factor or
random factor).
Design of questionnaire
The data can be classified into primary and secondary data. The data which is
collected for the first time by direct observation is primary data. The data which is obtained
from existing records, publications, etc., is known as secondary one. The different methods of
primary data collection are observation method, personal interview, telephone interview and
mail survey. The success of survey methods depends on the strength of the questionnaire
used. A questionnaire consists of a set of well-formulated questions to probe and obtain
responses from respondents. The questionnaire must contain provisions to collect all the data
items which are required for testing different hypotheses of the experiment as well as for
testing the hypotheses of other tests relating to various research issues.
Data Collection
Data is the basic input to any decision making process. The collection of data is a
critical step in providing the information needed to answer the research question. Every study
includes the collection of some type of data—whether it is from the literature or from
subjects—to answer the research question. Data can be collected in the form of words on a
survey, with a questionnaire, through observations, or from the literature.
The different methods which are used for primary data collection are observation
method, personal interview, telephone interview and mail survey. In an observation method,
the investigator will collect data through personal observations. Personal interview can be
classified into door-to-door interview, executive interview, mail intercept surveys, self-
administered questionnaires and purchase intercept technique. Telephonic interview is
considered to be a cost effective and dominant data collection method. Mail survey is a data
collection method in which questionnaires are mailed to potential respondents who in turn fill
and return them at their convenience.
18
The secondary data can be obtained from internal sources and external sources. The
internal sources of secondary data for marketing applications are sales records, cost
information, distributor reports and customer feedback. The different external sources of
secondary data are government publications, journals, books, magazines, newspapers, annual
reports, etc.
Data Analysis
After the data has been collected, the researcher turns to the task of analyzing them.
Proper tools and techniques should be used for classification and analysis of data.
Hypothesis-Testing
After analysing the data as stated above, the researcher is in a position to test the
hypotheses, if any, he had formulated earlier. Do this facts support the hypotheses or they
happen to be contrary? This is the usual question which should be answered while testing
hypotheses. Various tests, such as Chi square test, t-test, F-test, have been developed by
statisticians for the purpose. Hypothesis-testing will result in either accepting the hypothesis
or in rejecting it. If the researcher had no hypotheses to start with, generalisations established
on the basis of data may be stated as hypotheses to be tested by subsequent researches in time
to come.
Data Interpretation
The researcher must infer the results of the original research issues from the results
obtained through data analysis. if a hypothesis is tested and upheld several times, it may be
possible for the researcher to arrive at generalisation. The real value of research lies in its
ability to arrive at certain generalizations. If the researcher had no hypothesis to start with, he
might seek to explain his findings on the basis of some theory. It is known as interpretation.
The process of interpretation may quite often trigger off new questions which in turn may
lead to further researches
19
Research task remains incomplete till the report has been presented and/ or written.
Preparation of report is the last step in a research process and requires a set of skills
somewhat different from those called for in respect of the earlier stages of research. This task
should be accomplished by utmost care, researcher may seek the assistance and guidance of
experts for the purpose.
SELF ASSESSMENT:
Fill in the blanks:
SUMMARY:
GLOSSARY
Observation – is a systematic data collection approach. Researchers use all of their sense to
examine people in natural settings or naturally occurring situations.
20
Parameters – is a descriptive summary of a characteristics of set of data or summary
statistics of the relationship between two sets of data.
Variables – defined as anything that has a quantity or quality that varies or a variable is any
factor, trait, or condition that can exist in different amount or types.
Jon design – is the process of work arrangement aimed at reducing or overcoming job
dissatisfaction and employees alienation arising from repetitive and mechanistic tasks.
TERMINAL QUESTIONS:
SUGGESTED READINGS:
• Bhandarkar, P.L. and Wilkinson, T.S. Methodology and Techniques of Social Research.
• Dubin, Robert, Theory Building New York; MacMillan Publishing Co., Quoted in W Emory,
Business Research Methods.
• Gupta, S.P. Statistical Method, Sultan Chand and Sons, New Delhi.
21
LESSON-II
Structure:
Objectives
Introduction
Research Problem.
Research Design
Self Assessment
Summary
Glossary
Terminal Questions:
Suggested Readings:
OBJECTIVES:
22
• explain the sources and criteria of a good research problem,
• understand the different types of research design used for conducting different types of
researches.
INTRODUCTION :
The identification of research problem is the first and foremost step that every
researcher has to undertake. At times, it becomes rather difficult for an inexperienced
researcher or a novice/beginner in research to conceptualize a research problem. In general, a
research problem should be understood as some difficulty, unclear situation which a
researcher experiences in practical or theoretical context and wants to obtain a tangible
explanation, clarification or offer solution to it. For students, this problem may be as a result
of theoretical encounter in the area of specialization. As such, before embarking on any
research, you should identify the major research area of your interest, mostly the area of your
specialization. For instance from: Education, Social sciences, Humanities, Business
administration among others.
The problem defines the goal of the researcher in clear terms. Thus, without a problem,
research cannot proceed because there is nothing to processed from and proceed toward.
RESEARCH PROBLEM :
A research problem, or phenomenon as it might be called in many forms of qualitative
research, is the topic you would like to address, investigate, or study, whether descriptively or
experimentally. It is the focus or reason for engaging in your research. It is typically a topic,
phenomenon, or challenge that you are interested in and with which you are at least
somewhat familiar.
2. There must be some objectives pursuing the problem, otherwise it would be repugnant
to reason and common understanding to undertake the research.
23
3. There must be at least two lines of action to be taken to attain the objective. This
means that there must be at least two mean variable to a researcher for if he has no
choice of means, he cannot have a problem.
4. There must remain some doubt in the mind of a researcher with regard to the selection
of alternatives.
• daily problems
• technological changes
• un explored areas
A research may select a problem for investigation from a given theory in which he has
considerable interest. In such situations the researcher must have thorough knowledge of that
theory and should be sufficiently inquisitive to explore some unexplained aspects or
assumptions of that theory.
Technological changes in a fast changing society are constantly brought forth new
problems and new opportunities for research. What is the impact of a changed technology on
the existing socio economic set up, always interests the researcher and tempts him to
undertake such studies as are revealing regarding the impact of new technology on the
existing system.
Research problems can be both abstract and of applied interest. These may also be
selected from those areas which have not been explored so far. Such area may be theoretical
or empirical in nature.
24
Sometimes the researcher while discussing the interest with some other people may
come across a problem that can be researched by the investigator. The problem may relate to
any source as discussed above. In the same way reading assignments in text books, special
assignments, research reports and term papers may also suggest some additional areas of
needed research. Many research articles suggest some additional areas of needed research.
Many research articles suggest problem for further investigation that may prove fruitful.
Novelty
Interesting
The problem should be interesting for the investigator himself. If he is not interested
in to, he will be able to face and overcome the obstacles which come at every step in
research. His interest should be purely intellectual and should not be there only for a reward,
material benefit, advancement in position, increased authority, etc.
Importance
Immediate Application
The investigator should ask himself question, will my research help in solving an
urgent problem
Feasibility or Amenability
The suitability of the problem for a particular research worker is the matter of its
feasibility. The investigator should be able to carry it to a successful conclusion. He should
25
possess the required competence, knowledge and understanding. He should be skilful enough
to develop, administer, and interpret the necessary data gathering devices and procedures etc.
• Availability of data
• Availability of guidance
First of all the problem should be stated in a broad general way, keeping in view
either some practical concern or some scientific or intellectual interest. For this purpose, the
researcher must immerse himself thoroughly in the subject matter concerning which he
wishes to pose a problem. In case of social research, it is considered advisable to do some
field observation and as such the researcher may undertake some sort of preliminary survey
or what is often called pilot survey. Then the researcher can himself state the problem or he
can seek the guidance of the guide or the subject expert in accomplishing this task.
26
Understanding the nature of the problem
The next step in defining the problem is to understand its origin and nature clearly.
The best way of understanding the problem is to discuss’ it with those who first raised it in
order to find out how the problem originally came about and with what objectives in view. If
the researcher has stated the problem himself, he should consider once again all those points
that induced him to make a general statement concerning the problem. For a better
understanding of the nature of the problem involved, he can enter into discussion with those
who have a good knowledge of the problem concerned or similar other problems.
All available literature concerning the problem at hand must necessarily be surveyed
and examined before a definition of the research problem is given. This means that the
researcher must be well-conversant with relevant theories in the field, reports and records as
also all other relevant literature.
Finally, the researcher must sit to rephrase the research problem into a working
proposition. Once the nature of the problem has been clearly understood, the environment
(within which the problem has got to be studied) has been defined, discussions over the
problem have taken place and the available literature has been surveyed and examined,
rephrasing the problem into analytical or operational terms is not a difficult task. Through
rephrasing, the researcher puts the research problem in as specific terms as possible so that it
may become operationally viable and may help in the development of working hypotheses.
RESEARCH DESIGN:
The most important problem after defining the research problem is preparing the
design of the research, which is popularly known as the research design. A research design
27
helps to decide upon issues like what, when, where, how much, by what mean etc. with
regard to an enquiry or a research study. A research design is the framework or plan for a
study used as a guide in collecting and analyzing data. It is a blueprint for conducting a study
with maximum control over factors that may interfere with the validity of the findings. Thus,
research design provide an outline of what the researcher is going to do in terms of framing
the hypothesis, its operational implications, and the final data analysis
• It constitute a plan that defines the type and sources of information required for the
research problem;
• It constitutes a strategy that specifies the methods of data collection and analysis
which would be adopted;
• It also specifies the time period of research and monetary budget involved in
conducting the study, which comprise the two major constraints of undertaking any
research.
A magnitude that varies is known as a variable. The concept may assume different
quantitative values like height, weight, income etc. Quantitative variables are not quantifiable
in the strictest sense or objectively. However, the qualitative phenomena may also be
quantified in terms of the presence or absence of the attributes considered. Phenomena that
assumes different values quantitatively even in decimal points are known as ‘continuous
variable’. But all variables need not be continuous. Values that can be expressed only in
integer values are called ‘non- continuous variable’. When changes in one variable depends
upon the changes in one or more other variables, it is known as dependent or endogenous
variable, and the variables that cause the changes in the dependent variable are independent
or explanatory variables
• Extraneous variable
28
The independent variables which are not directly related to the purpose of the study
but affect the dependent variable are known as extraneous variables. For instance, assume
that a researcher wants to test the hypothesis that there is a relationship between children’s
school performance and their self-concepts, in which case the latter is an independent
variable and the former the dependent variable. In this context, intelligence may also
influence the school performance. However, since it is not directly related to the purpose of
the study undertaken by the researcher, it would be known as extraneous variable.
• Control
The term ‘control’ is used when a researcher designs the study in such a manner that it
minimizes the effects of extraneous independent variables.
• Confounded relationship
• Research hypothesis
When a prediction or a hypothesized relationship is tested by adopting scientific
methods, it is known as research hypothesis. Generally, hypothesis must consist of at least
one dependent variable and one independent variable.
• Experimental and non-experimental hypothesis
A research in which the independent variable is manipulated is known as
‘experimental hypothesis-testing research’. Whereas a research in which the independent
variable is not manipulated is termed as ‘non-experimental hypothesis-testing research’.
• Experimental and control group
When a group is exposed to usual conditions in an experimental hypothesis-testing
research, it is known as control group. On the other hand, when the group is exposed to
certain new or special condition, it is known as experimental group.
• Treatments
Treatments are referred to as the different conditions to which the experimental and
control groups are subject to. For example, if we want to determine through an experiment
the comparative impact of three varieties of fertilizers on the yield of wheat, in that case the
three varieties of fertilizers will be treated as three treatments.
• Experiment
29
An experiment refers to the process of verifying the truth of a statistical hypothesis
relating to a given research problem. Experiments may be categorized into two types, namely
absolute experiment and comparative experiment. If a researcher wishes to determine the
impact of a chemical fertilizer on the yield of a particular variety of rice crop, then it is
known as absolute experiment. Meanwhile, if the researcher wishes to determine the impact
of chemical fertilizer as compare to the impact of bio-fertilizer, then the experiment is known
as a comparative experiment.
(10) Experiment unit(s)
Experiment units refer to the pre-determined plots characteristics or the blocks, to
which the different treatments are applied. It is worth mentioned here that such experimental
units must be selected with great caution.
30
Exploratory research usually involves only a relatively small group of people, and (2) these
people are almost never randomly selected to participate.
Some of the more popular methods of exploratory research include literature searches,
depth interviews, analysis of ‘insight-stimulating’ examples.
Literature search
One of the quickest and least costly ways to discover hypotheses is to conduct a
literature search. Almost all research projects should start here. There is an incredible amount
of information available in libraries, through online sources, in commercial data bases, and so
on. The literature search may involve popular press (newspapers, magazines, etc.), trade
literature, academic literature, or published statistics from research firms or governmental
agencies.
Depth interviews
Depth interviews are used to tap the knowledge and experience of those with
information relevant to the problem or opportunity at hand. Anyone with relevant information
is a potential candidate for a depth interview, including current customers, members of the
target market, executives and managers of the client organization, sales representatives,
wholesalers, retailers, and so on. The object of such an interview is to obtain insight into the
relationship between variables and new ideas relating to the research problem. For such an
interview people who are competent and can contribute new ideas may be carefully selected
as respondents to ensure a representation of different type experience. It may enable the
researcher to find the problem more concisely and help in the formulation of the research
hypothesis.
31
Descriptive research design is a scientific method which involves observing and
describing the behaviour of a subject without influencing it in any way. The descriptive
research attempts to describe, explain and interpret conditions of the present i.e. “what is’.
The purpose of a descriptive research is to examine a phenomenon that is occurring at a
specific place(s) and time. A descriptive research is concerned with conditions, practices,
structures, differences or relationships that exist, opinions held, processes that are going on or
trends that are evident. Whereas, diagnostic research studies determine the frequency with
which something occurs or its association with something else. As a research design both
descriptive and diagnostic studies share common requirements, hence they may grouped
together. The research design must make appropriate provision for protection against bias and
thus maximize reliability, with due regard to the completion of research study in as
economical manner as possible. The research design in such studies should be rigid and not
flexible.
32
• Principle of Local Control.
33
variability it causes can be measured and hence eliminated from the experimental error. This
means that we should plan the experiment in a manner that we can perform a two-way
analysis of variance, in which the total variability of the data is divided into three components
attributed to treatments (varieties of rice in our case), the extraneous factor (soil fertility in
our case) and experimental error.* In other words, according to the principle of local control,
we first divide the field into several homogeneous parts, known as blocks, and then each such
block is divided into parts equal to the number of treatments. Then the treatments arc
randomly assigned to these parts of a block. This activity is known as ‘blocking’. In general,
blocks are the levels at which we hold an extraneous factor fixed, so that we can measure its
contribution to the total variability of the data by means of a two-way analysis of variance. In
brief, through the principle of local control we can eliminate the variability due to extraneous
factor(s) from the experimental error.
We may briefly deal with each of the above stated informal as well as formal
experimental designs.
34
• Before and after without control design: In such a design a single test group or area
is selected and the dependent variable is measured before the introduction of the
treatment. The treatment is then introduced and the dependent variable is measured
again after the treatment has been introduced. The effect of the treatment would be
equal to the level of the phenomenon after the treatment minus the level of the
phenomenon before the treatment. The main difficulty of such a design is that with
the passage of time considerable extraneous variations may be there in its treatment
effect.
• After-only with control design: In this design two groups or areas (test area and
control area) are selected and the treatment is introduced into the test area only. The
dependent variable is then measured in both the areas at the same time. Treatment
impact is assessed by subtracting the value of the dependent variable in the control
area from its value in the test area. The basic assumption in such a design is that the
two areas are identical with respect to their behaviour towards the phenomenon
considered. If this assumption is not true, there is the possibility of extraneous
variation entering into the treatment effect. However, data can be collected in such a
design without the introduction of problems with the passage of time. In this respect
this design is superior to before-and-after without control design.
• Before and-after with control design: In this design two areas are selected and the
dependent variable is measured in both the areas for an identical time-period before
the treatment. The treatment is then introduced into the test area only, and the
dependent variable is measured in both for an identical time-period after the
introduction of the treatment. The treatment effect is determined by subtracting the
change in the dependent variable in the control area from the change in the dependent
variable in test area. This design is superior to the above two designs for the simple
reason that it avoids extraneous variation resulting both from the passage of time and
from non comparable test and control areas. But at times, due to lack of historical
data, time or a comparable control area, we should prefer to select one of the first two
informal designs stated above.
• Completely Randomized Design (C. R. design): involves only two principles viz.,
the principle of replication and the principle of randomization of experimental
designs. The essential characteristic of this designs that subjects are randomly
assigned to experiment treatments (or vice-versa). For instance, if we have 10 subjects
35
and if we wish to test 5 under treatment A and 5 under treatment B, the randomization
process gives every possible group of 5 subjects selected from a set of 10 an equal
opportunity of being assigned to treatment A and treatment B. One-way analysis of
variance (or one-way ANOVA) is used to analyze such a design. Even unequal
replications can also work in this design. It provides maximum number of degrees of
freedom to the error. Such a design is generally used when experimental areas happen
to be homogenous. Technically, when all the variations due to the uncontrolled
extraneous factors are included under the heading of chance variation, we refer to the
design of experiment as C. R. design.
SELF ASSESSMENT:
36
1. Research project can involve both identification and problem-solving research.
2. Depth interviews are used to tap the knowledge and experience of those with
information relevant to the problem or opportunity at hand.
2.12 SUMMARY:
Research problem is a topic, phenomenon or challenge that one is interested in and which one is at
least somewhat familiar. The research problem may be selected from the theory of one’s own interest,
daily problems, technical changes, unexplored area and discussion with other people. Novelty,
interest, importance, immediate application and feasibility should be considered at the time of
selecting a research problem. Research problem should be defined in a systematic manner, giving due
weightage to all related points. The following steps should be followed one after the other: 1)
statement of the problem in a general way; II) understand the nature of the problem, II) surveying the
available literature IV) developing the idea through discussions, and v) rephrasing the research
problem in a working position.
The most important task after defining the research problem is preparing the design of the research
which is known as the research design. A research design is the framework or plan for a study used as
a guide in collecting and analyzing data. It is a blueprint for conducting a study with maximum
control over factors that may interfere with the validity of the findings. There are different types of
research designs for different types of researches which are explained in detail in the lesson.
2.13. GLOSSARY
Conceptualisation of research problems – is the process of not only selecting a topic, but
formulating a defensible and researchable research problem.
Preliminary survey – consists of fieldwork and gathering data from a variety of sources to
identify all issues in the project development process early enough to allow time to address.
37
Insight stimulating – the method consist of intensive study of selected instances of the
phenomenon in which one is interested.
TERMINAL QUESTIONS:
SUGGESTED READINGS:
• Bhandarkar, P.L. and Wilkinson, T.S. Methodology and Techniques of Social Research.
• Dubin, Robert, Theory Building New York; MacMillan Publishing Co., Quoted in W
Emory, Business Research Methods.
• Gupta, S.P. Statistical Method, Sultan Chand and Sons, New Delhi.
38
LESSON-III
Objectives
Meaning of Measurement
Levels of Measurement
Scaling
Scaling Techniques
Self Assessment
Summary
Glossary
Terminal Questions:
Suggested Readings:
OBJECTIVES:
• define Measurement,
39
• explore the source of error in measurement and know about the test of sound measurement,
MEANING OF MEASUREMENT:
LEVELS OF MEASUREMENT:
A level of measurement is the precision by which a variable is measured. The four
levels of measurement, from lowest to highest, are:
• Nominal
• Ordinal
• Interval
• Ratio
The nominal level of measurement describes variables that are categorical in nature.
The characteristics of the data researcher collecting fall into distinct categories. If there are a
limited number of distinct categories (usually only two), then you're dealing with a discrete
40
variable. If there are an unlimited or infinite number of distinct categories, then you're
dealing with a continuous variable. Nominal variables include demographic characteristics
like sex, race, and religion.
The ordinal level of measurement describes variables that can be ordered or ranked in
some order of importance. It describes most judgments about things, such as big or little,
strong or weak. Most opinion and attitude scales or indexes in the social sciences are ordinal
in nature.
The interval level of measurement describes variables that have more or less equal
intervals, or meaningful distances between their ranks. For example, if you were to ask
somebody if they were first, second, or third generation immigrant, the assumption is that the
distance or number of years, between each generation is the same. All crime rates in criminal
justice are interval level measures, as is any kind of rate.
The ratio level of measurement describes variables that have equal intervals and a
fixed zero (or reference) point. It is possible to have zero income, zero education, and no
involvement in crime, but rarely do we see ratio level variables in social science since it's
almost impossible to have zero attitudes on things, although "not at all", "often", and "twice
as often" might qualify as ratio level measurement.
a) Respondent
41
b) Situation
Situational factors may also come in the way of correct measurement. Any condition
which places a strain on interview can have serious effects on the interviewer-respondent
rapport e.g, if someone else is present, he can distort responses by joining in or merely by
being present. If the respondent feels that anonymity is not assured, he may be reluctant to
express certain feelings.
c) Measurer
d) Instrument
Error may arise because of the defective measuring instrument. The use of complex
words, beyond the comprehension of the respondent, ambiguous meanings, poor printing,
inadequate space for replies, response choice omissions, etc. are few things that make the
measuring instrument defective and may result in measurement errors.
Test of Validity
42
Validity is the extent to which differences found with a measuring instrument reflect
true differences among those being tested. But the question arises: how can one determine
validity without direct confirming knowledge? The answer may be that we seek other
relevant evidence that confirms the answers we have found with our measurement tool. What
is relevant, evidence often depends upon the nature of the research problem and the judgment
of the researcher.
• face
• content
• criterion
• construct
Face validity is the least statistical estimate (validity overall is not as easily quantified
as reliability) as it's simply an assertion on the researcher's part claiming that they've
reasonably measured what they intended to measure. It's essentially a "take my word for it"
kind of validity. Usually, a researcher asks a colleague or expert in the field to vouch for the
items measuring what they were intended to measure.
43
find at least one mass murderer, and investigate if the predictive factors on your scale,
retrospectively, affected them earlier in life. With criterion validity, you're concerned with
how well your items are determining your dependent variable.
Construct validity is the extent to which your items are tapping into the underlying
theory or model of behaviour. It's how well the items hang together (convergent validty) or
distinguish different people on certain traits or behaviours (discriminant validity). It's the
most difficult validity to achieve. You have to either do years and years of research or find a
group of people to test that have the exact opposite traits or behaviours you're interested in
measuring.
Test of Reliability
• test-retest
• multiple forms
• inter-rater
• split-half
44
The multiple forms technique has other names, such as parallel forms and disguised
test-retest, but it's simply the scrambling or mixing up of questions on your survey, for
example, and giving it to the same group twice. The idea is that it's a more rigorous test of
reliability.
Split-half reliability is estimated by taking half of your test, instrument, or survey, and
analyzing that half as if it were the whole thing. Then, you compare the results of this
analysis with your overall analysis. There are different variations of this technique, one of the
most common being called Cronbach's alpha (a frequently reported reliability statistic)
which correlates performance on each item with overall score. Another technique, closer to
the split-half method, is the Kuder-Richardson coefficient, or KR-20. Statistical packages on
most computers will calculate these for you, although in graduate school, you'll have to do
them by hand and understand that all test statistics are derived from the formula that all
observed scores consist of a true score and error score.
Test of Practicality
45
a) Concept development
This is the first step. In this case, the researcher should have a complete understanding
of all the important concepts relevant to his study. This step is more applicable to theoretical
studies compared to practical studies where the basic concepts are already established
beforehand.
Here, the researcher is required to specify the dimensions of the concepts, which were
developed in the first stage. This is achieved either by adopting an intuitive approach or by an
empirical correlation of the individual dimensions with that concept and/or other concepts.
c) Indicator selection
In this step, the researcher has to develop the indicators that help in measuring the
elements of the concept. These indicators include questionnaires, scales, and other devices,
which help to measure the respondents opinion, mindset, knowledge, etc. Using more than
one indicator lands stability and improves the validity of the scores.
d) Index formation
Here, the researcher combines the different indicators into an index. In case, there are
several dimensions of a concept the researcher needs to combine them.
SCALING
During research especially when the concepts we want to measure are complex and
abstract and there are no standardized measurement tools available, we face problems of
measurement. Alternatively, when we are measuring something which can lead to subject
bias like attitudes and opinions, there is a problem of their valid measurement. A similar
problem may be faced in a lesser degree while measuring physical or institutional concepts.
Therefore, knowledge of some such procedures which may enable accurate measurement of
abstract concepts is extremely essential. This brings us to the study of scaling techniques.
46
Scaling can be done in two ways: (i) making a judgement about an individuals
characteristics and then placing him on a scale which is defined in terms of that characteristic,
and (ii) constructing questionnaires where individual’s responses score assign them a place
on a scale. A scale is a continuum, consisting of the highest point and the lowest point along
with several intermediate points between these two extremities. These scale-point positions
are hierarchically related to each other. Numbers for measuring the degree of differences in
the attitudes or opinions are assigned to individuals corresponding to their positions in a
scale. Therefore, the term ‘scaling’ implies procedures for determination of quantitative
measures of subjective abstract concepts.
• Subject orientation
• Response form
• Degree of subjectivity
In this, the scale data is based on whether we measure subjective personal preferences
or just make non-preference judgements. In the former case, the respondent is asked to select
which person or solution he favors to be employed, whereas in the latter case he is simply
asked to judge which person or solution will be more effective without reflecting any
personal preference.
• Scale properties
In this, the scales can be classified as nominal, ordinal, interval and ratio scales.
Nominal scales merely classify without indicating order, distance or unique origin. Ordinal
scales indicate magnitude relationships of ‘more than’ or ‘less than’, but indicate no distance
47
or unique origin. Interval scales have both order and distance values, but no unique origin.
Whereas, ratio scales possess all these features.
• Number of dimensions
• Arbitrary approach: In this, the scales are developed on ad hoc basis. It is the
most widely used approach.
• Consensus approach: In this, a panel of judges evaluates the items chosen for
inclusion in the instrument regarding whether they are relevant to the topic area
and unambiguous in implication.
• Item analysis approach: In this, a number of individual items are developed into
a test that is given to a group of respondents. Post administering the test, total
scores are evaluated, and the individual items are analyzed to determine which
items discriminate between persons or objects with high and low total scores.
• Cumulative scales: These are chosen on the basis of their conforming to some
ranking of items with ascending and descending discriminating power.
SCALING RECHNIQUES:
We know take up of the important scaling techniques often used in the context of
research specially in context of social or business research. Scaling techniques are of two
types i.e. rating scale and ranking scale.
Rating Scales
48
against some specified criteria i.e. we judge properties of objects without reference to the
similar objects. These rating may be in such forms as “like-dislike, “above average, average,
below average”, or other classifications with more categories such as “likely very much-like
somewhat-neutral-dislike somewhat- dislike very much” and so on. There is no specific rule
whether to use a two point scale, three point scale or scale with still more points. Rating scale
may be either a graphic rating scale or an itemized rating scale.
• The graphic rating scale: is quite simple and is commonly used in practice. Under it
the various points are usually put along the line to form a continuum and the rater
indicates his rating by simply making a mark at the appropriate point on a line that
runs from one extreme to the other. Scale points with brief descriptions may be
indicated along the line, their function being to assist the rater in performing his job.
• The itemized rating scale: presents a series of statements from which a respondent
selects one as best reflecting his evaluation. These statements are ordered
progressively in terms of more or less of some property.
Ranking Scales: Under ranking scales we make relative judgements against other
similar objects. The respondents under this method directly compare two or more objects and
make choices among them. There are two generally used approaches of ranking scale viz.
• Method of paired comparisons: Under it the respondents can express his attitude by
making a choice between a new flavor of soft drink and an established brand of drink.
But when there is more than two stimuli to judge, the number of judgements required
in a paired comparison is given by the formula:
N= n(n-1)/2
• Method of Rank order: Under this method of comparative scaling, the respondents
are asked to rank their choice. This method is easier and faster than the method of
paired comparisons stated above. The problem of transitivity is also not there in case
we adopt method of rank order.
49
A “Likert scale” is actually the sum of responses to several Likert items. These items
are usually displayed with a visual aid, such as a series of radio buttons or a horizontal bar
representing a simple scale. It is a five (or seven) point scale which is used to allow the
individual to express how much they agree or disagree with a particular statement. Likert-
type or frequency scales use fixed choice response formats and are designed to measure
attitudes or opinions. These ordinal scales measure levels of agreement/disagreement. A Likert-
type scale assumes that the strength/intensity of experience is linear, i.e. on a continuum from
strongly agree to strongly disagree, and makes the assumption that attitudes can be measured.
Respondents may be offered a choice of five to seven or even nine pre- coded responses with
the neutral point being neither agree nor disagree.
• As a first step, the researcher collects a large number of statements which are relevant
to the attitude being studied and each of the statements expresses definite
favourableness or unfavourableness to a particular point of view.
• After the statements have been gathered, a trial test should be administered to a
number of subjects.
• The response to various statements are scored in such a way that a response indicative
of the most favourable attitude is given the highest score of 5 and that with the most
unfavourable attitude is given the lowest score, say, of 1.
• Then the total score of each respondent is obtained by adding his scores that he
received for separate statements.
• The next step is to array these total scores and find out those statements which have a
high discriminatory power.
• Only those statements that correlate with the total test should be retained in the final
instrument and all others must be discarded from it.
50
which one or more respondents’ rate one or more concepts on each scale item. Various steps
involved in developing semantic differential scale are as follows:
• First of all the concept to be studied are selected. The concepts are usually chosen by
personal judgement, keeping in view the nature of the problem.
• The next step is to select the scales bearing in mind the criterion of factor composition
and the criterion of scale’s relevance to the concepts being judged. One more criterion
to be kept in view is that scales should be stable across subjects and concepts.
• Then a panel of judges are used to rate the various stimuli on the various selected
scales and the responses of all judges would then be combined to determine the
composite scaling
It is the simplest among all the data methods. In this method, each respondent is given
a number of points, say 10,100 or 200. Then he or she allocates these points to different
stimuli according to some attitudinal characteristic. The sum of all the points allocated to all
the stimuli should be equal to the total number of points given to the respondents. Before
allocating points to different stimuli, the respondent should compare the relative importance
of the stimuli with respect to the assumed attitudinal characteristics.
Differential Scales
Under such an approach, the selection of items is made by a panel of judge who
evaluate the items in terms of whether they are relevant to the topic area and unambiguous in
implication. The detailed procedure is as follows:
• The researcher gathers a large number of statements, usually twenty or more, that
express various points of view toward a group, institution, idea, or practice.
• These statements are then submitted to a panel of judges, each of whom arranges
them in eleven groups or piles ranging from one extreme to another in position. Each
of the judges is requested to place generally in the first pile the statement which he
thinks are most unfavourable to the issue, in the second pile to place those statements
which he thinks are next most unfavourable and he goes on doing so in this manner
till in the eleventh pile he puts the statements which he considers to be the most
favourable.
51
• This sorting by each judge yields a composite position for each of the items. In case
of disagreement between the judges in assigning a position to an item, that item is
discarded.
• For items that are retained, each is given its median scale value between one and
eleven as established by the panel.
• A final selection of statements is then made. For this purpose a sample of statement,
whose median scores are spread evenly from one extreme to the other is taken.
The cumulative scale or Guttman scale measures to what degree a person has a
positive or negative attitude to something. It makes use of a series of statements that are
growing or descending in how positive or negative a person is towards the subject. If for
instance on a scale with seven statements the respondent agrees with the fifth statement, it
implies that he or she also agrees with the first four statements, but not with statement
number six and seven. To create a Guttman scale, you need to:
• Determine the focus: what concept are you going to measure (see what people's
attitudes are toward it)?
• Ask a group of people (or a person) to write down different statements about this
concept, reflecting different opinions or attitudes about the subject. Make sure you
have a large number of statements, making sure that people can either degree or
disagree with them (no - open - questions for instance).
• Rating the scale items: the next step is to have your judges rate each statement,
indicating whether the statement expresses a positive (favourable) or negative
(unfavourable) attitude towards the concept. The members of the group must not
express their own opinion about the concept; they must only indicate whether the
statement is favourable or unfavourable.
• Developing the scale: construct a matrix or table that shows the responses of all the
respondents on all of the items. Then sort this matrix so that respondents who agree
with more statements are listed at the top and those agreeing with fewer are at the
bottom. For respondents with the same number of agreements, sort the statements
from left to right from those that most agreed to those that fewest agreed to.
52
• If there are lots of items, you need to use a data analysis called scalogram analysis to
determine the subsets of items from our pool that best approximate the cumulative
property. Then review these items and select the final scale elements. There are
several statistical techniques for examining the table to find a cumulative scale.
• Because there is seldom a perfectly cumulative scale you usually have to test how
good it is. These statistics also estimate a scale score value for each item. This scale
score is used in the final calculation of a respondent's score.
Two approaches, viz., the metric approach and the non-metric approach, are usually
talked about in the context of MDS. The metric approach to MDS treats the input data as
interval scale data and solves applying statistical methods for the additive constant which
minimize the dimensionality of the solution space. The non-metric approach first gathers the
non-metric similarities by asking respondents to rank order all possible pairs that can be
obtained from a set of objects. Such non-metric data is then transformed into some arbitrary
metric space and then the solution is obtained by reducing the dimensionality.
SELF ASSESSMENT
53
5. The most sophisticated statistical analysis can be made if data is collected
by…………
3.11 SUMMARY:
Measurement can be defined as the assignment of a number to an object which reflects the degree of
possession of a characteristic by that object. There are four levels of measurement namely nominal,
ordinal, interval and ratio. The nominal level of measurement describes variables that are categorical
in nature.
The ordinal level of measurement describes variables that can be ordered or ranked in some order of
importance. The interval level of measurement describes variables that have more or less equal
intervals, or meaningful distance between their ranks. The ratio level of measurement describes
variables that have equal interval and a fixed zero. Test of validity, test of reliability and test of
practicality are the test of sound measurement.
Scaling is a process of assigning numbers to various degrees of attitudes, preferences, opinion and
other concepts. Scaling techniques are of two types namely rating scale and ranking scale. The most
commonly used scale for measuring attitudes of people are Likert type scale, semantic differential
scale, constant sum scale, differential scales, cumulative scale and multidimensional scale.
3.12. GLOSSARY
Respondents – are those persons who have been invited to participate in a particulat study
and have actually taken part in the study.
TERMINAL QUESTIONS:
54
• Describes the test for sound measurement.
• How would you proceed to construct a Likert scale ? Give its advantages and
disadvantages.
SUGGESTED READINGS:
• Bhandarkar, P.L. and Wilkinson, T.S. Methodology and Techniques of Social Research.
• Dubin, Robert, Theory Building New York; MacMillan Publishing Co., Quoted in W
Emory, Business Research Methods.
• Gupta, S.P. Statistical Method, Sultan Chand and Sons, New Delhi.
55
Lesson –IV
DATA COLLECTION
STRUCTURE
Objectives
Introduction
Primary Data
Secondary Data
Self Assessment
Summary
Glossary
Terminal Question:
Suggested Readings:
OBJECTIVES:
56
• know about the various methods of primary data collection,
INTRODUCTION:
Once the researcher has decided the research design, the next job is of data collection.
Data collection starts with determining what kind of data required followed by the selection
of a sample from a certain population. The reliability of managerial decisions depends on the
quality of data. The quality of data can be expressed in terms of its representative feature of
the reality which can be ensured by the usage of a fitting data collection methods. Depending
upon the source utilized, whether the data has come from actual observations or from records
that are kept for normal purposes, statistical data can be classified into two categories i.e
primary and secondary:
PRIMARY DATA:
Primary data is a type of information that is obtained directly from first-hand sources
by means of surveys, observation or experimentation. It is data that has not been previously
published and is derived from a new or original research study and collected at the source
such as in marketing.
• The primary data are original and relevant to the topic of the research study so the
degree of accuracy is very high.
• Primary data is that it can be collected from a number of ways like interviews,
telephone surveys, focus groups etc. It can be also collected across the national
borders through emails and posts. It can include a large population and wide
geographical coverage.
• Moreover, primary data is current and it can better give a realistic view to the
researcher about the topic under consideration.
57
• Reliability of primary data is very high because these are collected by the concerned
and reliable party.
• A lot of time and efforts are required for data collection. By the time the data
collected, analysed and report is ready the problem of the research becomes very
serious or out dated. So the purpose of the research may be defeated.
• It has design problems like how to design the surveys. The questions must be simple
to understand and respond.
• Some respondents do not give timely responses. Sometimes, the respondents may
give fake, socially acceptable and sweet answers and try to cover up the realities.
With more people, time and efforts involvement the cost of the data collection goes
high. The importance of the research may go down.
• In some primary data collection methods there is no control over the data collection.
Incomplete questionnaire always give a negative impact on research.
In the observation method, the investigator will collect data through personal
observations. In this method the investigator will observe the behaviour of the respondents in
disguise. Under observation method information is sought by way of investigators own direct
observation without asking from the respondents.
Advantages of Observation
1. Very direct method for collecting data or information – best for the study of human
behaviour.
58
5. Helps in understanding the verbal response more efficiently.
6. Observation is less demanding in nature, which makes it less bias in working abilities.
Disadvantages of Observation
Questionnaire Method
59
Advantages of Questionnaires
• It is free from the bias of the interviewer; answers are in respondents own words.
• There is low cost even when the universe is large and is widely spread geographically.
• Respondents, who are not easily approachable, can also be reachable conveniently.
• Large samples can be made use of and thus the results can be made more dependable
and reliable.
Disadvantage of Questionnaire
• There is inbuilt inflexibility because of the difficulty of amending the approach once
questionnaire have been dispatched.
Interview Method
In this method the interviewer personally meets the informants and asks necessary
questions to them regarding the subject of enquiry. Usually a set of questions or a
questionnaire is carried by him and questions are also asked according to that. The
interviewer efficiently collects the data from the informants by cross examining them. The
interviewer must be very efficient and tactful to get the accurate and relevant data from the
informants. Interviews like personal interview/depth interview or telephone interview can be
conducted as per the need of the study.
This method acts as a very vital tool for the collection of the data in the social
research as it is all about the direct systematic conversation between an interviewer and the
respondent. By this the interviewer is able to get relevant information for a particular research
problem.
Personal interviews
Under this method there is face to face contact between the interviewer and the
interviewee. This sort of interview may be in the form of direct personal investigation or it
60
may be indirect oral investigation. In the case of direct personal investigation the interviewer
has to collect the information personally from the sources concerned. He has to be on the spot
and has to meet people from whom data have to be collected. This method is particularly
suitable for intensive investigations.
Telephone interview
• Very good technique for getting the information about the complex, emotionally laden
subjects.
• Data collected by this method is likely to be more correct compared to the other
methods that are used for the data collection.
Mail Survey
61
straightforward. A major disadvantage of a mail survey is that it usually has lower response
rates than other data collection methods. This may lead to problems with data quality.
This method of data collection is very much like the collection of data through
questionnaire, with little difference which lies in the fact that schedules are being filled by the
enumerators who are specially appointed for the purpose. These enumerators along with
schedules go to respondents, put to them the questions from the proforma in the order the
questions are listed and record the replies in the space meant for the same in the proforma.
Enumerators explain the aims and objects of the investigation and also remove the difficulties
which any respondent may feel in understanding the implications of a particular question.
This method of data collection is very useful in extensive enquiries and can lead to
fairly reliable results. It is, however, very expensive and is usually adopted in investigations
conducted by governmental agencies or by some big organizations.
S. Questionnaire Schedule
No
62
5. The questionnaire method is likely Information is collected well in time
to be very slow since many respondents as they are filled by enumerators.
do not return the questionnaire.
7. This method can be used only when The information can be gathered
respondents are literate and even when the respondents happen to
cooperative. be illiterate.
10. The success of questionnaire methods It depends upon the honesty and
lies more on the quality of the competence of enumerators
questionnaire itself.
63
SECONDARY DATA:
Secondary data is the data that have been already collected by and readily available
from other sources. Such data are cheaper and more quickly obtainable than the primary data
and also may be available when primary data cannot be obtained at all.
Researcher must be very careful in using secondary data. He must make a minute
scrutiny because it is just possible that the secondary data may be unsuitable or may be
inadequate in the context of the problem which the researcher wants to study. By way of
caution, the researcher, before using secondary data, must see that they posses following
characteristics:
• Reliability of data
The reliability can be tested by finding out such things about the said data: (a) Who
collected the data (b) What were the sources of data? (c) Were data collected by using proper
methods? (d) At what time were they collected? (e) Was there any bias of the compiler? (f)
What level of accuracy was desired? Was it achieved?
• Suitability of data
The data that are suitable for one enquiry may not be necessarily be found suitable in
another enquiry. Hence, if the available data are found to be unsuitable, they should not be
used by the researcher. In this context, researcher must very carefully scrutinize the definition
of various terms and units of collection used at the time of collecting the data from the
primary source originally. Similarly, the object, scope and nature of the original enquiry must
also be studied. If the researcher finds differences in these, the data will remain unsuitable for
the present enquiry and should not be used.
• Adequacy of data
If the level of accuracy achieved in data is found inadequate for the purpose of the
present enquiry, they will be considered as inadequate and should not be used by the
researcher. The data will be considered inadequate if they are related to an area which may be
either narrow or wider than the area of the present enquiry.
64
• It saves time and finance.
• It is very difficult to find secondary data which exactly fulfils the need of present
investigation.
The nature, the scope as well as the object of the enquiry are very important as they
affect the choice of the method. The method selected should be such that it suits the type of
enquiry that is to be conducted by the researcher. This factor is also important in deciding
whether the data already available (secondary data) are to be used or the data not yet
available (primary data) are to be collected.
• Availability of funds
When a method is chosen, it's important to check whether there is adequate amount of
funds to make it work. If the method is too expensive, it will be very hard to do the
experiment.
• Time factor
Time is an important factor as decided when the experiment has to end. Some
methods take relatively more time, whereas with others the data can be collected in a
65
comparatively shorter duration. The time at the disposal of the researcher, thus, affects the
selection of the method by which the data are to be collected.
• Precision required
If available, internal secondary data may be obtained with less time, effort and money
than the external secondary data. In addition, they may also be more pertinent to the situation
at hand since they are from within the organization. The internal sources include
• Accounting resources
This gives so much information which can be used by the marketing researcher.
They give information about internal factors.
• Internal Experts
These are people who are heading the various departments. They can give an idea of
how a particular thing is working
• Miscellaneous Reports
These are what information you are getting from operational reports.
External Sources are sources which are outside the company in a larger environment.
Collection of external data is more difficult because the data have much greater variety and
the sources are much more numerous. External data can be divided into following classes:
66
• Govt. Publications
Government sources provide an extremely rich pool of data for the researchers. In
addition, many of these data are available free of cost on internet websites. The
Central Statistical Organization (CSO) and various state govt. collect, compile and
publish data on regular basis.
• International Bodies
Semi govt. organizations municipalities, District Boards and others also publish
reports in respect of birth, death and education, sanitation and many other related
fields.
• Private Publications
Some commercial and research institutes publish reports regularly. They are like
Institutes of Economic Growth, Stock Exchanges, National Council of Education
Research and Training (NCERT), National Council of Applied Economic Research
(NCAER) etc.
• Research Scholars
Individual research scholars collect data to complete their research work which
further is published with their research papers.
67
SELF ASSESSMENT
3. The data which is primary in the hands of one become/remain .............................. in the
hands of another.
4. The requirement of primary and secondary data depends upon…………………..
5. Data is reliable when it is collected from……………………
SUMMARY:
The reliability of managerial decision depends on the quality of data. The quality of data can be
expressed in terms of its representative feature of the reality which can be ensured by the usage of
fitting data collection methods. Statistical data can be classified in two categories viz primary and
secondary.
Primary data is a type of information that is obtained directly from first hand information by means of
surveys, observations of experimentation. Primary data can be collected by observation method,
questionnaire methods, interview method and interview schedule method.
On the other hand, secondary data is the data that have been already collected by the others and
readily available from other sources, however, researcher must be very careful in using secondary
data. Thus, before using secondary data, researcher must check the reliability , suitability and
adequacy of such data. As there are many methods for collection of data and it is important to
consider the following factors while selecting a particular method
• Availability of funds
• Time factor
• Decision required
68
Secondary data can be collected from internal sources as well as external sources. The internal
sources include accounting resources, sales reports, internal experts and miscellaneous reports.
External sources include government publications, international bodies, semi-government
publications, newspapers and magazines and research scholars.
GLOSSARY
Sample – is a group of people, objects, or items that are taken from a larger population for
measurement.
Sales report – describes the record of all the calls which have been made and products that
have been sold during a particular time frame by a salesperson or the management.
TERMINAL QUESTIONS:
• What is primary data? Discuss the various methods of collecting primary data?
• What are the merits and limitations of secondary data? Explain various sources of secondary
data.
• What do you mean by questionnaire? What is the difference between questionnaire and
schedule?
SUGGESTED READINGS:
• Bhandarkar, P.L. and Wilkinson, T.S. Methodology and Techniques of Social Research.
• Dubin, Robert, Theory Building New York; MacMillan Publishing Co., Quoted in W
Emory, Business Research Methods.
• Gupta, S.P. Statistical Method, Sultan Chand and Sons, New Delhi.
69
• Kothari, C.R. Research Methodology-Methods and Techniques, Whishwa
Prakashan, New Delhi.
70
LESSON-V
SAMPLING DESIGN
Structure
Objectives
Introduction
Need of Sampling
Methods of Sampling
Limitations of Sampling
Self Assessment
Summary
Glossary
Terminal questions
Suggestive reading
OBJECTIVES:
• frame a sample,
71
• understand the various methods of sampling and selection of a suitable method under the
given situation,
INTRODUCTION:
Successful statistical practice is based on focused problem definition. In sampling,
this includes defining the population from which the sample is drawn. Population is also
known as a well-defined collection of individuals or objects known to have similar
characteristics. It is the aggregate of all elements, usually defined prior to the selection of the
sample. The population is said to be completely defined if at least the element, sampling
units, extent and time are specified.
Sample: The sample is a part or a small section of the population selected for study.
Sampling Frame: A set of information used to identify a sample population for statistical
treatment. A sampling frame includes a numerical identifier for each individual, plus other
identifying information about characteristics of the individuals, to aid in analysis and allow
for division into further frames for more in-depth analysis.
72
ESSENTIALS OF GOOD SAMPLING:
In order to reach at right conclusions, a sample must possess the following essential
characteristics.
1. Representative
The sample should truly represent the characteristics of the universe. For this
investigator should be free from bias and the method of collection should be appropriate.
2. Adequacy
The size of the sample should be adequate i.e., neither too large nor small but
commensurate with the size of the population.
3. Homogeneity
There should be homogeneity in the nature of all the units selected for the sample. If
the units of the sample are of heterogeneous character it will be impossible to make a
comparative study with them.
4. Independentability
The method of selection of the sample should be such that the items of the sample are
selected in an independent manner. This means that selection of one item should not
influence the selection of another item in any manner and that each item should be selected
on the basis of its own merit.
NEED OF SAMPLING:
Sampling is necessary because of the following reasons
• If data collection takes a long time, then value of some characteristics may change
over the period of time thus, defeating the very purpose of data analysis. Thus, due to
importance of time element sample is required.
• Sample, if representative, may yield more accurate results than the total census
because sample can be more accurately supervised.
73
• Quality of some product is tested by destroying the products.
Defining the population of interest, for business research, is the first step in sampling
process. In general, target population is defined in terms of element, sampling unit, extent,
and time frame. The definition should be in line with the objectives of the research study. For
example, if a kitchen appliances firm wants to conduct a survey to ascertain the demand for
its micro ovens, it may define the population as ‘all women above the age of 20 who cook
(assuming that very few men cook)’. However this definition is too broad and will include
every household in the country, in the population that is to be covered by the survey.
Therefore the definition can be further refined and defined at the sampling unit level, that, all
women above the age 20, who cook and whose monthly household income exceeds
Rs.20,000. This reduces the target population size and makes the research more focused. The
population definition can be refined further by specifying the area from where the researcher
has to draw his sample, that is, households located in Hyderabad.
74
Once the definition of the population is clear a researcher should decide on the
sampling frame. A sampling frame is the list of elements from which the sample may be
drawn. Continuing with the micro oven example, an ideal sampling frame would be a
database that contains all the households that have a monthly income above Rs.20,000.
However, in practice it is difficult to get an exhaustive sampling frame that exactly fits the
requirements of a particular research. In general, researchers use easily available sampling
frames like telephone directories and lists of credit card and mobile phone users. Various
private players provide databases developed along various demographic and economic
variables. Sometimes, maps and aerial pictures are also used as sampling frames. Whatever
may be the case, an ideal sampling frame is one that entire population and lists the names of
its elements only once.
A sampling frame errorcpopsup when the sampling frame does not accurately
represent the total population or when some elements of the population are missing another
drawback in the sampling frame is over –representation. A telephone directory can be over
represented by names/household that have two or more connections.
A sampling unit is a basic unit that contains a single element or a group of elements of
the population to be sampled. In this case, a household becomes a sampling unit and all
women above the age of 20 years living in that particular house become the sampling
elements. If it is possible to identify the exact target audience of the business research, every
individual element would be a sampling unit. This would present a case of primary sampling
unit. However, a convenient and better means of sampling would be to select households as
the sampling unit and interview all females above 20 years, who cook. This would present a
case of secondary sampling unit.
The sampling method outlines the way in which the sample units are to be selected.
The choice of the sampling method is influenced by the objectives of the business research,
availability of financial resources, time constraints, and the nature of the problem to be
investigated. All sampling methods can be grouped under two distinct heads, that is,
probability and non-probability sampling.
75
The sample size plays a crucial role in the sampling process. There are various ways
of classifying the techniques used in determining the sample size. A couple of those hold
primary importance and are worth mentioning are whether the technique deals with fixed or
sequential sampling and whether its logic is based on traditional or Bayesian methods. In non-
probability sampling procedures, the allocation of budget, thumb rules and number of sub-
groups to be analyzed, importance of the decision, number of variables, nature of analysis,
incidence rates, and completion rates play a major role in sample size determination. In the case
of probability sampling, however, formulas are used to calculate the sample size after the levels
of acceptable error and level of confidence are specified. The details of the various techniques
used to determine the sample size will be explained at the end of the chapter.
In this step, the specifications and decisions regarding the implementation of the
research process are outlined. Suppose, blocks in a city are the sampling units and the
households are the sampling elements. This step outlines the modus operandi of the sampling
plan in identifying houses based on specified characteristics. It includes issues like how is the
interviewer going to take a systematic sample of the houses. What should the interviewer do
when a house is vacant? What is the recontact procedure for respondents who were
unavailable? All these and many other questions need to be answered for the smooth
functioning of the research process. These are guidelines that would help the researcher in
every step of the process. As the interviewers and their co-workers will be on field duty of
most of the time, a proper specification of the sampling plans would make their work easy
and they would not have to revert to their seniors when faced with operational problems.
This is the final step in the sampling process, where the actual selection of the sample
elements is carried out. At this stage, it is necessary that the interviewers stick to the rules
outlined for the smooth implementation of the business research. This step involves
implementing the sampling plan to select the sampling plan to select a sample required for
the survey.
METHODS OF SAMPLING:
76
When a sample is required to be reflected from a population, it is necessary to decide
which method should be applied. Basically, there are two different types of sample designs,
namely, probability sampling and non-probability sampling. Each of the two is described
below.
This type of sampling is also known as random sampling or chance sampling. This
sampling procedure gives each element in the population an equal chance of getting selected
for the sample; besides, all choices are independent of one another. The obtained results of
probability sampling can be assured in terms of probability. In other words, we can measure
the errors of estimation or the significance of obtained results from a random sample. In fact,
due to this very reason probability sampling design is superior to the deliberate sampling
design. Probability sampling ensures the law of Statistical Regularity, which states that if the
sample chosen is a random one, the sample will have the same composition and
characteristics as the universe. Hence, probability sampling is more or less the best technique
to select a representative sample
We shall now discuss some of the various sampling methods under two separate
headings as follows:
77
• Stratified Sampling
• Systematic Sampling
• Cluster Sampling
• Judgment Sampling
• Quota Sampling
• Convenience Sampling
• Snowball Sampling
A simple random sample is one in which every member of the target population has
the same probability of being included in the sample. The selection is thus free from personal
bias because the investigator does not exercise his discretion of preference in the choice of
items. Since selection of items in the sample depends entirely on chance, this method is also
known as the method of chance selection. There are a variety of ways of selecting a random
sample from a population list. One way is to use a random number generator to assign a
random number to each member of the target population, order the members on the list
according to the random number and choose the first n members on the list, where n is the
required sample size.
Stratified Sampling
78
component parts; we get a better estimate of the whole. In brief, stratified sampling results is
more reliable and detailed information.
Systematic Sampling
Systematic sampling involves selecting every nth member from a population list. If
the list is random, then selecting every nth member is another method of obtaining a simple
random sample. However, if the list is not random, this procedure can introduce bias. Non-
random order would include alphabetical order or date of birth order.
Cluster Sampling
Multistage Sampling
As the name implies, this method refers to a sampling procedure which is carried out
in several stages. The material is regarded as made up of a number of first stage sampling
units, each of which is made of a number of second stage units, etc. At first, the first stage
units are sampled by some suitable method, such as random sampling. Then, a sample of
second stage unit is selected from each of the selected first stage units again by some suitable
method which may be the same or different from the method employed for the first stage
units. Further stages may be added as required.
Judgment Sampling
In judgment sampling, the selection of the sample is based on the judgment of the
investigator, who is studying a situation. In other word, the investigator exercises his
judgment in the choice of sample items and includes those items in the sample which he
thinks are most typical of the population with regard to the characteristics under
investigation. This method is also known as ‘purpose sampling’ or ‘deliberate sampling’.
This sampling method should be carried out by an expert in the field as his judgment will
influence the final outcome of the study.
79
Quota Sampling
Convenience Sampling
This method is based on the convenience of the researcher. The researcher uses the
sources available to him to come to the conclusion.
Snowball Sampling
This involves asking people who have participated in a survey to nominate other
people they believe would be willing to take part. Sampling continues until the required
number of responses is obtained. This technique is often used when the population is difficult
for the researchers to identify. For example, we might expect software hackers to be known
to one another, so if we found one to take part in our survey, we could ask him/her to identify
other possible participants.
• Less time: since the sample is a study of part of the population, considerable time and
labour are saved when a sample survey is carried out. Time is saved not only in
collecting data but also in processing it. For these reasons, a sample provides more
timely data in practice than a census.
• More reliable results: Although the sampling technique involves certain inaccuracies
owing to sampling errors, the result obtained is generally more reliable than that
obtained from a complete count. There are several reasons for it. First, it is always
possible to determine the extent of sampling errors. Secondly, other types of errors to
which a survey is subjected, such as inaccuracy of information, incompleteness of
returns, etc. are likely to be more serious in a complete census than in a sample
survey.
80
• More detailed information: since the sampling technique save time and money, it is
possible to collect more detailed information in a sample survey.
• Less cost: The amount of effort and expenses involved in collecting information is
always greater per unit of the sample than a complete census. The total financial
burden of a sample survey is generally less than that of a complete census.
LIMITATIONS OF SAMPLING:
• A sample survey must be carefully planned and executed, otherwise, the results
obtained may be inaccurate and misleading.
• Sampling generally requires the services of experts, if only for consultation purposes.
In the absence of qualified and experienced persons, the information obtained from
sample survey cannot be relied upon.
• At times, the sampling plan may be so complicated that it requires more time, labour
and money than a complete count. This is so if the size of the sample is a large
proportion of the total population and if complicated weighted procedures are used.
• If the information is required for each and every unit in the domain of study, a
complete enumeration survey is necessary.
Sampling Errors
The sampling errors arise due to drawing faulty interference about the population
based upon the results of the samples. The sampling error would be less, if the sample size is
large relative to the population and vice-versa. The reason is that the estimate is based on a
part and not on the whole.
Even if utmost care has been taken in selecting a sample, the results derived from the
sample may not be representative of the population from which it is drawn, because samples
are seldom, if ever, perfect miniatures of the population. This gives rise to sampling errors.
Sampling error are of two types- biased and unbiased.
81
1. Biased Errors
These errors arise from any bias in selection, estimation etc. For example if in place
of simple random sampling, deliberate sampling has been used in particular case some bias is
introduced in the result and hence, such errors are called biased sampling errors.
These errors arise due to chance differences between the members of population
included in the sample and those not included. An error in statistics is the difference between
the value of a statistic and that of the corresponding parameter.
• Diversity of population.
Non-Sampling Errors
82
Causes of Non- Sampling Errors
Non - sampling errors may arise from one or more of the following factors:
• Data specification being inadequate and inconsistent with respect to the objectives of
the census or survey.
SELF ASSESSMENT
Fill in the blanks:
1. A part of population is known as ………..
2. The list of sampling units from which a sample is taken is called…………….
3. Quota sampling is a.............................. sampling.
4. Select cluster and then select elements from each selected cluster is an
example of……………..
5 .............................is known as accidential sampling.
SUMMARY:
A complete enumeration of all items in the population is known as census inquiry. A sample is a
subset of a population that is used to represent the entire group as a whole. Thus, sample is the
selection of part of the universe for the purpose of drawing conclusion or inference about the entire
universe from the study of these parts. In order to reach at right conclusions, a sample must possess
the following essential characteristics.
• Representativeness
• Adequacy
83
• Homogeneity
• Independentability
Further, various steps are to be followed in sampling. These include defining the target population,
specifying the sampling frame, specifying the sampling and selection of a sampling method,
determination of sample size, specifying the sampling plan and selecting the sample.
Sampling can be done either by probability sampling method or by non probability sampling method.
Probability sampling means where each and every unit of the population has equal chance to be
selected in the sample. Non probability is that when the organisations of the inquiry deliberately
choose the particular units of the universe to compose a sample on the basis that the small mass
selected out of a large would represent the whole. Probability sampling methods include simple
random sampling, stratified sampling, systematic sampling, cluster sampling and multistage sampling.
On the other hand, non –probability sampling methods include judgment sampling, convenience
sampling, quota sampling. In the data collection process, some scope for errors still remains Inspite of
taking all precautions. These errors can be sampling and non sampling errors. Sampling error is the
deviation of the selected sample from the true characteristics, traits, behaviours and qualities of the
entire population. Non sampling errors occur at the time of observation approximation and processing
of data.
5.11. GLOSSARY
Coding – is a process of identifying a passage in the text or other data items,
searching and identifying concept and finding relationship between them.
Credit card – is a payment card issued to users to enable the cardholder to pay a
merchant for goods and services based on the cardholder’s promise to the card
issuer to pay them for the amounts plus the other agreed charges.
Tabulation – is the systematical arrangement of the statistical data in columns or
rows.
TERMINAL QUESTIONS:
• Define the term sampling? What should be the qualities of good sampling?
84
• Discuss in brief probability and non probability methods of sampling?
SUGGESTED READINGS:
• Bhandarkar, P.L. and Wilkinson, T.S. Methodology and Techniques of Social Research.
• Dubin, Robert, Theory Building New York; MacMillan Publishing Co., Quoted in W
Emory, Business Research Methods.
85
LESSON –VI
Structure:
Objectives.
Introduction
Types of Analysis
dispersion meaning
Skewness meaning
Measures of skewness
Kurtosis
Correlation meaning
Types of correlation
Method of correlation
Regression analysis
Regression equations
Regression coefficients
Hypothesis Testing
Characteristics of Hypothesis
86
Student t-Test
ANOVA Analysis
Z-Test
Chi-Square Test
Self Assessment
Summary
Glossary
Terminal Questions:
Suggested Readings:
OBJECTIVES:
• acquire knowledge regarding the consideration to be taken into account while analyzing data,
• know about the two tailed and one tailed test of hypothesis ,and
INTRODUCTION:
The data, after collection, has to be analysed in accordance with the outline laid down
for the purpose at the time of developing the research plan. This is essential for a scientific
study and for ensuring that we have all relevant data for making contemplated comparisons
and analysis. The term analysis refers to the computation of certain measures along with
searching for patterns of relationship that exists among data-groups. Thus, in the process of
analysis, relationships or differences supporting or conflicting with original or new
87
hypotheses should be subjected to statistical tests of significance to determine with what
validity data can be said to indicate any conclusions.
There are many benefits of data analysis however; the most important ones are as
follows:
Data analysis helps in structuring the findings from different sources of data
collection like survey research. It is again very helpful in breaking a macro problem into
micro parts. Data analysis acts like a filter when it comes to acquiring meaningful insights out
of huge data-set. Every researcher has sort out huge pile of data that he/she has collected,
before reaching to a conclusion of the research question. Mere data collection is of no use to
the researcher. Data analysis proves to be crucial in this process. It provides a meaningful
base to critical decisions.
One of the most important uses of data analysis is that it helps in keeping human bias
away from research conclusion with the help of proper statistical treatment. With the help of
data analysis a researcher can filter both qualitative and quantitative data for an assignment
writing projects. Thus, it can be said that data analysis is of utmost importance for both the
research and the researcher. Or to put in another words data analysis is as important to a
research as it is important for a doctor to diagnose the problem of the patient before giving
him any treatment.
88
investigators should have substantially more than a basic understanding of the rationale for
selecting one method of analysis over another. This can allow investigators to better
supervise staff who conduct the data analyses process and make informed decisions.
While methods of analysis may differ by scientific discipline, the optimal stage for
determining appropriate analytic procedures occurs early in the research process and should
not be an afterthought. Statistical advice should be obtained at the stage of initial planning of
an investigation so that, for example, the method of sampling and design of questionnaire are
appropriate.
Environmental/Contextual issues
89
how they react to each other’s responses. Since the data collection process could be
influenced by the environment/context, researchers should take this into account when
conducting data analysis.
TYPES OF ANALYSIS:
As stated earlier, by analysis we mean the computation of certain indices or measures
along with searching for patterns of relationship that exist among the data groups. Analysis,
particularly in case of survey or experimental data, involves estimating the values of
unknown parameters of the population and testing of hypotheses for drawing inferences.
Analysis may, therefore be categorized as descriptive analysis and inferential analysis.
“Descriptive analysis is largely the study of distributions of one variable. This study provides
us with profiles of companies, work groups, persons and other subjects on any of a multiple
of characteristics such as size, composition, efficiency, preferences etc.” this sort of analysis
may be in respect of one variables (described as univariate analysis), or in respect of two
variables (described as bivariate analysis) or in respect of more than two variables (described
as multivariate analysis).
Univariate analysis explores each variable in a data set. It looks at the range of values,
as well as the central tendency of the value. It describes the pattern of response to the
variable. it describes each variable on its own.
1) Raw Data
Obtain a printout of the raw data for all the variables. Raw data resembles a matrix,
with the variable name heading the columns, and the information for each case or record
displayed across the rows.
90
Example: Raw data for a study of injuries among county workers (first 10 cases).
1 County A Fall 3
2 County B Auto 4
3 County C Fall 6
4 County C Fall 4
5 County B Fall 5
6 County A Violence 9
7 County A Auto 3
8 County A Violence 2
9 County A Violence 9
10 County B Auto 3
It is difficult to tell what is going on with each variable in this data set. Raw data is
difficult to grasp, especially with large number of cases or records. Univariate descriptive
statistics can summarize large quantities of numerical data and reveal patterns in the raw data.
In order to present the information in a more organized format, start with univariate
descriptive statistics for each variable. For example, the variable severity of injury.
Severity of Injury
91
3
• Frequency Distribution
Obtain a frequency distribution of the data for the variable. This is done by
identifying the lowest and highest values of the variables, and then putting all the values of
the variable in order from lowest to highest. Next, count the number of appearance of each
value of the variable. This is a count of the frequency with which each value occurs in the
data set. For example, for the variable “Severity of Injury”, the values range from 2 to 9.
Severity of Number of
Injury Injuries with this
Severity
2 1
3 3
4 2
5 1
6 1
9 2
Total 10
• Grouped Data
The severity of injury ratings can be collapsed into just a few categories or groups.
Grouped data usually has from 3 to 7 groups. There should be no groups with a frequency of
zero (for example, there are no injuries with a severity rating of 7 or 8).
One way to construct groups is to have equal class intervals (e.g., 1-3, 4-6, 7-9).
Another way to construct groups is to have about equal numbers of observations in each
group. Remember that class intervals must be both mutually exclusive and exhaustive.
Cumulative frequency distributions include a third column in the table (this can be
done with either simple frequency distribution or with grouped data).
92
Severity of Injury Number of Injuries with Cumulative frequency
this Severity
2 1 1
3 3 4
4 2 6
5 1 7
6 1 8
9 2 10
A cumulative frequency distribution can answer questions such as, how many of the
injuries were at level 5 or lower? Answer= 7.
• Percentage Distributions
93
Bar Graphs
Bar graphs are used to display the frequency distributions for variables measured at
the nominal and ordinal levels. But graphs use the same width for all the bars on the graph,
and there is space between the bars. Label the parts of the graph, including the title, the left
(Y) or vertical axis, the right (x) or horizontal axis, and the bar labels.
HISTOGRAM
A histogram is a chart that is similar to a bar chart, but it is used for interval and ratio
level variables. With a histogram, the width of the bar is important, since it is the total area
under the bar that represents the proportion of the phenomenon accounted for by each
category. The bars convey the relationship of one group or class of the variable to the
other(s).
For example, in the case of the counties and employee injuries, we might have
information on the rate of injury according to the number of workers in each county in State
X.
94
County I 1.6
County J 1.0
County K 0.9
County L 0.4
If we group the injury rates into three groups, then a low rate of injury would be 0.0-
1.9 injuries per 1,000 workers; moderate would be 2.0-3.9; and high would be 4.0 and above
(in this case, up to 5.9). This could be graphed as follows:
FREQUENCY POLYGON
For example, the following table shows the average injury rate per 1,000 employees
for counties in State X for the years 1980 to 1990.
Year 1980 1981 1982 1983 1984 1985 1986 1987 1988 1988 1990
Rate 3.6 4.2 3.4 5.5 3.8 3.1 1.7 1.8 1.0 1.6 0.9
95
A cumulative frequency polygon is used to display the cumulative distribution of
values for a variable
PIE CHART
96
Rates and Ratios
Other way to look at the sub-groups or classes within one variable is by the relation of
each sub-group or class to the whole. This can be calculated with a proportion. A proportion
is obtained by dividing the frequency of observations counted for one group or class (written
as f ) by the total number of observations counted for the variable (written as N).
A rate is the relationship between two different numbers, for example, the number of
injuries among county workers and the population of the county. This can be calculated as the
first number (N1, or injuries) divided by the second number (N2, or population).
Many health statistics are expressed as rates, for example, the birth rate is the number
of births per some population, such as number of births per 1,000 women.
DISPERSION
An average can represent a series only as best as a single figure can, but it certainly cannot
reveal the entire story of any phenomenon under study. Specially, the averages cannot alone
describe adequately a set of observations, unless all the observations are the same. It is
necessary to describe the variability or dispersion of the observation. In two or more
distributions the central value may be the same but still there can be wide disparities in the
formation of the distribution. Measures of the dispersion help us in studying this important
characteristics’ of a distribution.
MEASURES OF DISPERSION
97
Range = L-S
Where, L = Largest item
S = Smallest item
The relative measure corresponding to range, called the co-efficient of range, is obtained by
applying the following formula:
∑|D|
M. D. =
N
Where, M.D = Mean Deviation
|D| is the absolute value of the deviation ignoring plus and minus signs.
N is the total number of items.
The relative measure corresponding to the mean deviation, called the co-efficient of
mean deviation, is obtained by dividing mean deviation by particular average used in
computing mean deviation.
M. D
Co − efficient M. D. =
Mean
98
• Standard deviation – it is the most widely used measure of dispersion of a series and
is denoted by the ′σ′ symbol pronounced as sigma. Its significant lies in the fact that it
is free from those defects from which the earlier method suffer and satisfies most of
the properties of a good measure of dispersion. Standard deviation for the values of
individual items in a series is obtained from the arithmetic average.
SKEWNESS
skewness tells the direction and extent of skewness/ asymmetry in a series. Mean, median and
mode are identical in a symmetrical distribution. Asymmetry or skewness is larger if mean
moves away from the mode. In the present study, skewness will be calculated by means of
Karl Pearson’s Coefficient of Skewness as it is an important characteristic for defining the
precise pattern of a distribution. Skewness is defined as a measure of asymmetry. In a
symmetrical distribution, the values of mean, median and mode coincide. If skewness is
positive than mode< median < mean and in case of negative skewness mean < median <
mode. It has been used to study the direction (either towards the lower side or mean score or
towards the higher side of the mean score).
MEASURES OF SKEWNESS
The followings are the important measures of skewness .
99
1. Karl pearson’s co-efficient of skewness – this is method of measuring skewness is based
upon the difference between mean and mode. The difference is divided by standard deviation
to give a relative measure. Symbolically,
SKB = Q3 + Q1 – 2 Median / Q3 – Q1
Or
Absolute SK = D1 + D9 – 2 Median
SKk = P1 + P 9 – 2 Median / P9 – P1
Co-efficient of Skewness
The co- efficient of skewness, as a statistical tool, helps in the study of the degree and
direction of variation from the centre value. It also shows that a particular distribution is
positively or negatively skewed. This method is useful in studying the concentration of
responses of the respondents either on the higher side or lower side of mean score with
100
respect to their opinion on different statements. In the case of normal distribution, the value
of skewness will be zero. The positive skewness is denoted by Mode<Median<Mean and in
case of the negative skewness we find Mean<Median<Mode it has been calculated with the
help of following formula:
X̄ − Z
SKp =
σ
KURTOSIS
Kurtosis is also a characteristic of a frequency distribution which tells about the shape
of the frequency distribution. In statistics Kurtosis refers, whether a frequency distribution is
more flat-topped or more peaked than the normal distribution. Kurtosis is a measure of
weather the data are heavy-tailed or light-tailed relative to a normal distribution. The value of
the coefficient 𝛽2 is the most important measure of kurtosis. It is calculated as:
𝛽2 = 𝜇 4
𝜇22
For a normal curve, the value of 𝛽2=3. The curve is more peaked than the normal curve
when the value of 𝛽2 is greater than 3, i.e., leptokurtic. When the value of 𝛽2 is less than 3 then
the curve is less pleaded than the normal curve i.e., platykutic. And if 𝛽3=3, then it called
mesokurtic.
CORREALTION
Correlation is measure of association between two variables. The variables are not designed
as dependents or independent. The most popular correlation coefficients are: Spearman’s
correlation coefficient and Pearson’s product moment correlation coefficient. When
calculating a correlation coefficient for ordinal data, select Spearman’s techniques. For
interval or ratio type data, use Pearson’s techniques. The value of a correlation coefficient can
vary from minus one to plus one. A minus one indicates a perfect negative correlation, while
a plus one indicates a perfect positive correlation. A correlation of zero means there is no
relationship between two variables. When there is negative correlation between two
variables, as value of one variable increase, the value of the other variable decrease, and vice
versa.
11
TYPES OF CORRELATION
1. Positive correlation: if two variables X & Y moves in the same direction, i.e. if one
rises, other rises too & vice versa it is called positive correlation. Examples;
relationship between price and supply.
2. Negative correlation: if two variable X and Y move in opposite directions, i.e. if one
rise other falls & if one falls, other rises, then it is called as negative correlation.
Examples: relationship between demand & price.
3. Linear correlation: the ratio of change of two variables X& Y remain constant
throughout, then they are said to be linearly correlated, like as when every time supply
of commodity rises by 20% as often as its price by 10%, then such two variables have
linear relationship.
4. Non linear correlation: if the ratio between the two variables is not constant but
changing, correlation is said to be curvi-linear, like as when every time price of a
commodity rises by 10%, then sometimes supply rises by 20%, sometimes by 10% &
sometimes by 40%.
5. Simple correlation: when we study the relationship between the two variables only,
then it is called simple correlation. Example, relationship between price and demand.
6. Partial correlation: when three or more variables are taken but relationship between
any two of the variables is studied, assuming other variables as constant, then it is
called partial relationship. Suppose, under constant temperature, we study the
relationship between amount of rainfall and wheat yield, then this will be called as
partial correlation.
7. Multiple correlation: when we study the relationship among three or more variables,
then it is called multiple correlation. For example, if we study the relationship
between rainfall, temperature and yield of wheat, then it is called multiple correlation.
METHODS OF CORREALTION
12
Scatter Diagram – Scatter diagram is a graphic method of finding out correlation between
two variables. By this method, direction of correlation can be ascertained. For constructing a
scatter diagram, X-variable is represented on X-axis and the Y-variable on Y-axis. Each pair
of values of X and Y series is plotted in two dimensional space of X-Y. Thus we get a scatter
diagram by plotting all the pair of values.
Graphic Method – this method is very simple. The data pertaining to two series X and Y are
plotted on a graph sheet. A visual observation of graph will give a broad idea of the direction
of movement and closeness of two curves. If the curves move in the same direction, then
correction is positive. If they move in the opposite direction, then the correlation is negative.
Rank Correlation Coefficient – the Karl Pearson method discussed above is based on the
assumption that the population being studied is normally distributed. When it is known that
the population is not normal or when the shape of the distribution is not kwon, there is need
for a measure of correlation that involves no assumptions about the parameter of the
populations. It is possible to avoid making any assumption about the population being the
studied by ranking the observations according to size and basing the calculation or the ranks
rather than upon the original observations. This method is finding out co variability or the
lack of it between two variables was developed by the British psychologist Charles Edward
Spearman in 1904. The main objective of this coefficient is to determine the extent to which
the two sets of ranking are similar r dissimilar.
REGRESSION ANALYSIS
The study of regression has special importance in statistical analyses. We know that the
mutual relationship between two series is measured with the help of correlation. Under
13
correlation, the direction and magnitude of the relationship between two variables is
measured. But it is not possible to make the best estimate of the value of a dependents
variables is measured. But it is not possible to make the best estimate of the value of
dependent variables on the basis of the given values of the independent variable by
correlation analysis. Therefore, to make the best estimates and future estimation, the study of
regression analysis is very useful and important. According to Oxford English Dictionary, the
word ‘Regression’ analysis means “Stepping back” or “Returning to average value”. The
term was first of all used by a famous Biological Scientist in 19th century, Sir Francis Galton
relating to a study of hereditary characteristics’. In statistical analyses, the term ‘Regression’
is taken in wider sense. Regression is the study of nature of relationship between the variables
so that one may be able to predict the unknown value of one variable for a known value of
another variable. In regression, one variable is considered as an independent variable and
other is independent variable.
In simple regression analysis, we study only two variables at a time, in which one variable is
dependent and other is independent. The functional relationship between income and
expenditure is an example of simple regression. On the contrary, we study more than two
variables at a time in multiple regression analysis (i.e. at least three variables) in which one is
dependent variable and other are independent variable. The study of fact of rain and irrigation
on yield of wheat is an example of multiple regression.
REGRESSION EQUATIONS
Simple linear regression is a statistical device with the help of which we are in a position to
estimate the unknown values of one variable from known values of another variable. The
variable which is used to predict the variable of interest is called the independent variable or
explanatory variable and the variable we want to predict is called the dependent variable or
explained variables. The independent variable is denoted by X and dependent by Y. The
analysis used is called the simple linear regression analysis-simple because there is only one
independent variable, and linear regression analysis simple because there is only one
independent variable, and linear because of the assumed linear relationship between the
dependent and the independent variables. The basic relationship between X and Y variables is
given by
14
Yc = a + bX
Regression equations are the algebraic formulation of regression lines. Regression equation
represents regression lines. Just as there are two regression lines, similarly there are two
regression equations, which are as follows:
REGRESSION COEFFICIENTS
There are two regression equations; similarly there are two regression coefficients.
Regression coefficient measures the average change in the value of one variable for a unit
change in the value of another variable. For two variables X and Y, there are two regression
coefficients, which are given as follows:
σy
byx = r.
σx
The value of byx can also be found out by other formulae.
15
σx
bxy = r.
σy
HYPOTHESIS TESTING:
A hypothesis is a tentative statement about the relationship between two or more
variables. A hypothesis is a specific, testable prediction about what researcher expects to
happen in his study. Hypothesis requires as convenient mathematical approach for
simplifying cumbersome calculation. Setting-up and testing hypothesis is an integral art of
statistical inference.
CHARACTERISTICS OF HYPOTHESIS:
i) Hypothesis is an attempt explanation of certain puzzled facts. It is the attempt to find
out the possibility of the occurrence of such a fact. Sometimes it is framed to explain
certain scientific truth or to explain a law.
iv) Through hypothesis facts are organized in a systematic manner. The aim of
hypothesis is to reach at the real explanation and to remove the puzzlement
concerning the event.
16
Null Hypothesis
A hypothesis stated in the hope of being rejected is called a null hypothesis and is
denoted by H0.
Alternative Hypothesis
• The next step is to select a significance level. Typically the 0.05 or the 0.01 level is
used.
• The third step is to calculate a statistic analogous to the parameter specified by the
null hypothesis.
• The fourth step is to calculate the probability value (often called the p value). The p
value is the probability of obtaining a statistic as different or more different from the
parameter specified in the null hypothesis as the statistic computed from the data. The
calculations are made assuming that the null hypothesis is true.
• The probability value computed in Step 4 is compared with the significance level
chosen in Step 2. If the probability is less than or equal to the significance level, then
the null hypothesis is rejected; if the probability is greater than the significance level
then the null hypothesis is not rejected. When the null hypothesis is rejected, the
outcome is said to be "statistically significant" when the null hypothesis is not
rejected then the outcome is said be "not statistically significant."
17
• If the outcome is statistically significant, then the null hypothesis is rejected in favor
of the alternative hypothesis. If the rejected null hypothesis were that μ1- μ2 = 0, then
the alternative hypothesis would be that μ1≠ μ2. If M1 were greater than M2 then the
researcher would naturally conclude that μ1 ≥ μ2.
• The final step is to describe the result and the statistical conclusion in an
understandable way.
When the population mean is either lower or higher than some hypothesized value,
one tailed test is considered to be appropriate.
Suppose that you suspect that a particular class's performance on a proficiency test is
not representative of those people who have taken the test. The national mean score on the
test is 74.
Or in notation: H a : μ ≠ 74
In notation: H 0: μ = 74
As you decide 5 percent probability level for the test. Both tests have a region of
rejection, then, of 5 percent, or 0.05. In this example, however, the rejection region must be
split between both tails of the distribution—0.025 in the upper tail and 0.025 in the lower tail—
because your hypothesis specifies only a difference, not a direction, as shown in Figure 1(a).
You will reject the null hypotheses of no difference if the class sample mean is either much
higher or much lower than the population mean of 74.
18
Figure 1.Comparison of (a) a two‐tailed test and (b) a one‐tailed test, at the same
probability level (95 percent).
The decision of whether to use a one or a two‐tailed test is important because a test
statistic that falls in the region of rejection in a one‐tailed test may not do so in a two‐tailed
test, even though both tests use the same probability level.
STUDENT T-TEST:
The Student t-test is probably the most widely used parametric test. It was developed
by a statistician working at the Guinness brewery and is called the Student t-test because of
proprietary rights. A single sample t-test is used to determine whether the mean of a sample is
different from a known average. A two-sample t-test is used to establish whether a difference
occurs between the means of two similar data sets. The t-test uses the mean, standard
deviation, and number of samples to calculate the test statistic. .
As an example, given 1000 men measured for height in China and Japan, are the
mean heights different? China’s mean is 169.1 cm with a standard deviation of 6.21 cm, and
Japan’s mean height is 168.6 cm with a standard deviation of 5.7 6cm. The t-value is 1.88;
therefore, the mean heights are not statistically different.
ANOVA
An analysis of variance partitions the overall variation between the observation Y into
variation which has been accounted for by the regression on X and residual or unexplained
variation. Analysis of variance (ANOVA) is a procedure for testing the differences among the
means of populations by examining the amount of variation within each of these samples,
19
relative to the amount of variation between the samples. The essence of ANOVA is that the
total amount of variation in a set of data is broken down into two types, that amount which
can be attributed to chance and that amount which can be attributed to specified causes. There
may be variation between samples and also within sample items. ANOVA consists in
splitting the variance for analytical purposes. Hence, it is a method of analyzing the variance
to which a response is subject into its various components corresponding to various sources
of variation. In its simplest form, ANOVA provides a statistical test of whether or not the
means of several groups are equal
Analysis of variance (ANOVA) splits the variance of the variable into two components. One
component is the variability among group means. It is computed by summing square of the
differences between every group mean and overall mean of the distribution. This value is
divided by the degree of freedom k-1 where k number of groups to obtain Mean Sum of
squares between groups (MSB) The other component is the variability within the groups (also
called residual variation). It is quantified as the sum of squares of the differences between
each observation and its respective group mean. This value is divided by degrees of freedom n-
k where n is total number of observations and k is number of groups, to find Mean Sum of
squares within groups (MSW). The ratio of the MSB and MSW is called the F ratio. The
calculated F ratio values are compared to the standardized table value of F from the F-
distribution. If the calculated F-ratio value is greater than the table value at an acceptable
level of significance, we will reject the null hypothesis of equality of means and conclude that
the means of the groups are significantly different. In other words, large F ratios signify that
the variation among group means is more than it would we if this variation were simply
outcome of chance.
Z -TEST:
The next test, which is very similar to the Student t-test, is the z-test. However, with
the z-test, the variance of the standard population, rather than the standard deviation of the
study groups, is used to obtain the z-test statistic. Using the z-chart, like the t-table, we see
what percentage of the standard population is outside the mean of the sample population. If,
like the t-test, greater than 95% of the standard population is on one side of the mean, the p-
value is less than 0.05 and statistical significance is achieved. As some assumption of sample
size exists in the calculation of the z-test, it should not be used if sample size is less than 30.
If both the n and the standard deviation of both groups are known, a two sample t-test is best.
110
CHI-SQUARE TEST:
The chi-square test is used to determine whether there is a significant difference
between the expected frequencies and the observed frequencies in one or more categories.
SELF ASSESSEMENT
Fill in the blanks:
SUMMARY:
Analysis refers to the computation of certain measures along with searching for pattern of relationship
that exist among data groups. Data analysis helps in structuring the findings from different sources of
data collection like survey research. It is also helpful in breaking a macro problem into micro parts.
Data analysis acts like filter when it comes to acquiring meaning insight out of huge data set. It
provides a meaningful base to critical decisions. Data analysis can be done by way of frequency
111
distribution, diagrams and graphs. The averages cannot alone describe adequately a set of
observations, unless all the observations are the same. It is necessary to describe the
variability or dispersion of the observation. In statistics Kurtosis refers, whether a frequency
distribution is more flat-topped or more peaked than the normal distribution.
Correlation is measure of association between two variables. The variables are not designed
as dependents or independent. Regression is a statistical device with the help of which we are
in a position to estimate the unknown values of one variable from known values of another
variable. The testing of hypothesis is a process of testing the significance of a parameter of the
population on the basis of sample. T-test, ANOVA, z-test, chi-square test etc. are used to test the
hypothesis.
GLOSSARY
Probability – probability is a measure quantifying the likelihood that events will occurs.
Probability quantifies as a number between 0 and 1, where, roughly speaking, 0 indicates
impossibility and 1 indicates certainty.
TERMINAL QUESTIONS:
• What do you mean by Data Analysis? What considerations are taken into account while
analyzing data?
112
• Explain the concept of correlation and regression? Discuss the difference between correlation
and regression.
• Define the term ‘Measure of dispersion’. What are various methods of measuring dispersion?
SUGGESTED READINGS:
• Dubin, Robert, Theory Building New York; MacMillan Publishing Co., Quoted in W
Emory, Business Research Methods.
• Gupta, S.P. Statistical Method, Sultan Chand and Sons, New Delhi.
113
LESSON-VII
INTERPRETATION AND REPORT WRITING
Structure
Objectives
Introduction
Interpretation
Need of Interpretation
Technique of Interpretation
Precautions in Interpretation
Report Writing
Types of Reports
Self Assessment
Summary
Glossary
Terminal Questions:
Suggested Readings:
OBJECTIVES:
114
• interpret the results of analysis,
• know the technique of interpretation and precaution to be taken at the time of interpretation,
• know about the guidelines for writing a report and layout of the research report,
INTRODUCTION:
Interpretation is the process of making sense of numerical data that has been
collected, analyzed, and presented. A common method of assessing numerical data is known
as statistical analysis, and the activity of analyzing and interpreting data in order to
make predictions is known as inferential statistics. After collecting and analyzing the data,
the researcher has to accomplish the task of drawing inferences followed by report writing.
Only through interpretation that the researcher can expose relations and processes that
underlie his findings. All this analytical information and consequential inference(s) may well
be communicated, preferably through research report, to the consult of research results who
may be either an individual or a group of individuals or some public private organization.
INTERPRETATION:
Interpretation refers to the task of drawing inferences from the collected facts after an
analytical and or experimental study. In fact, it is a search for broader meaning of research
findings. The task of interpretation has two major aspects viz., the effort to establish
continuity in research through linking the results of a given study with those of another, and
the establishment of some explanatory concepts. “In one sense, interpretation is concerned
with relationships within the collected data, partially overlapping analysis. Interpretation also
extends beyond the data of the study to inch the results of other research, theory and
hypotheses. ” Thus interpretation is the device through which the factors that seem to explain
what has been observed by researcher in the course of the study can be better understood and
it also provides a theoretical conception which can serve as a guide for further researches.
NEED OF INTERPRETATION:
115
Interpretation is essential for the simple reason that the usefulness and utility of
research findings lies in proper interpretation. It is being considered a basic component of
research process because of the followings reasons:
• Researcher can better appreciate only through interpretation why his findings are what
they are and can make others to understand the real significance of his research
findings.
• It is through interpretation that the researcher can well understand the abstract
principle that works beneath his findings. Through this ha can link up his findings
with those of other studies, having the same abstract principle, and they can predict
about the concrete world of events. Fresh inquiries can test these predictions later on.
This way the continuity in research can be maintained.
• The interpretation of the findings of exploratory research study often results into
hypotheses for experimental research and as such interpretation is involved in the
transition from exploratory to experimental research. Since an exploratory study does
not have a hypothesis to start with, the findings of such a study have to be interpreted
on a post-factum basis in which case the interpretation is technically described as ‘post-
factum’ interpretation.
TECHNIQUE OF INTERPRETATION:
Technique of Interpretation often involves the following steps:
(i) Researcher must give reasonable explanations of the relations which he has found and
he must interpret the lines of relationship in terms of the underlying processes and
must try to find out the thread of uniformity that lies under the surface layer of his
diversified research findings. In fact this is the technique of how generalization should
be done and concepts to be formulated. In fact this is the technique of how
generalization should be done and concepts to be formulated.
(ii) Extraneous information, if collected during the study, must be considered while
interpreting the final results of research study, for it may prove to be a key factor in
understanding the problem under consideration.
116
(iii) It is advisable, before embarking upon final interpretation, to consult someone having
insight into the study and who is frank and honest and will not hesitate to point out
omissions and errors in logical argumentation. Such a consultation will result in
correct interpretation and, thus, will enhance the utility of research results.
(iv) Researcher must accomplish the task of interpretation only after considering all
relevant factors affecting the problem to avoid false generalization. He must be in no
hurry while interpreting results, for quite often the conclusions, which appear to be all
right at the beginning, may not at all be accurate.
PRECAUTIONS IN INTERPRETATION:
One should always remember that even if the data are properly collected and
analyzed, wrong interpretation would lead to inaccurate conclusions. It is therefore;
absolutely essential that the task of interpretation be accomplished with patience in an
impartial manner and also in correct perspective. Researcher must pay attention to the
following points for correct interpretation:
• (a) the data are appropriate, trustworthy and adequate for drawing inferences; (b) the
data reflect good homogeneity; and that (c) proper analysis has been done through
statistical methods.
• The researcher must remain cautious about the errors that can possibly arise in the
process of interpreting results. He should be well equipped with and must know the
correct use of statistical measures for drawing inferences concerning his study.
• He must never lose sight of the fact that his task is not only to make sensitive
observations of relevant occurrences, but also to identify and disengage the factors
that are initially hidden to the eye. This will enable him to do his job of interpretation
on proper lines. Broad generalization should be avoided as most research is not amen-
able to it because the coverage may be restricted to a particular time, a particular area
and particular conditions. Such restrictions, if any, must invariably be specified and
the results must be framed within their limits.
• The researcher must remember that “ideally in the course of a research study, there
should be constant interaction between initial hypothesis, empirical observation and
theoretical conceptions. It is exactly in this area of interaction between theoretical
117
orientation and empirical observation that opportunities for originality and creativity
lie.” He must pay special attention to this aspect while engaged in the task of
interpretation.
REPORT WRITING:
A Report may be defined as a document in which a given problem is examined for the
purpose of conveying information, reporting findings, putting forward ideas and, sometimes,
making recommendations.
The purpose of conducting a research is to come out with inferences and suggestions.
Generally, a research involves utilization of different types of resources and results in
significant cost of conducting it. Hence, the outcome of research should be well documented
in the form of a research report for implementation and future use. The design and
presentation of the research report involves a pre-specified sequence of steps.
TYPES OF REPORTS:
Research reports vary greatly in length and type. In each individual case, both the
length and the form are largely dictated by the problems at hand. For instance, business firms
prefer reports in the letter form, just one or two pages in length. Banks, insurance
organisations and financial institutions are generally fond of the short balance-sheet type of
tabulation for their annual reports to their customers and shareholders. Mathematicians prefer
to write the results of their investigations in the form of algebraic notations. Students of
literature usually writs long reports presenting the critical analysis of some writer or period or
the like with a liberal use of quotations from the works of the author under discussion. A
research report can be classified into decision-oriented (technical) report and popular report
The technical report is generally intended for other researchers, or for research
managers. The report should enable another researcher to critique methodology, check
calculations and accuracy and to follow everything which is done on a step-by-step basis. In
the technical report the main emphasis is on (i) the methods employed (ii) assumptions made
in the course of study, (iii) the detailed presentations of the findings including their
limitations and supporting data.
118
• Identification of the problem
• Establishment objectives
Problem identification is the process of selecting a problem which will give very good
pay-off in terms of value addition to the organization. The objectives or criteria of the
decision problem should be listed and finally a meaningful combination of the objective is to
be selected by taking the ground realities of the decision environment into account. The
objective / criteria of the decision problem can be fulfilled/ achieved by many ways. The
researcher should list all the meaningful decision alternatives of achieving the set of criteria
for the decision problem. Then, the best combination of the decision alternatives is to be
selected such that the value/ pay-off of the decision is maximized. The next step is to draw an
action plan to implement the selected set of decision alternatives to achieve the objectives of
the decision problem. Sometimes, there is deviation from the action plan while implementing
the decision alternatives. So, the decision maker should be proved with contingency plan of
tackling such a situation.
The popular report is intended for a more general audience, one that is not that
conversant with the detail of research methods and terminology. Compared to the technical
report, the presentation will be a bit more lively with increased attention to headlines, flow
diagrams, charts, tables and occasional summaries for the purpose of stressing major points.
• Consider the Audience: Make the report clear, use only words familiar to the readers
and defined all technical terms. To make the comparison of figures easier, use
119
percentages, rounded off figures, ranks or ratios; put the exact data in a table within
the text or in the appendix.
• Be Concise, Yet Complete: Most managers will not want to read about the details of
a research report. Knowing what to include and what to leave out is a difficult task. It
is up to you, the researcher, to take into account the information needs of the decision
maker when writing your report.
• Be Objective: You will probably face at least one situation in which you know that
the client will not easily accept the results. The findings may be conflict with the
decision makers experience and judgment or they may reflect unfavorable on the
wisdom of previous decisions. In these circumstances, there is a strong temptation to
start the report by making the result more acceptable to the management. A
professional researcher, however, will present the research findings in an objective
manner and will present the research findings in an objective manner and will defend
their validity if they are challenged by the client.
• Style: The style of writing a research report is important because it shows a way of
presentation. Here are a few a tips to help you write a report that is easy to read.
• Be concise.
• Avoid clichés.
120
• Preliminary Page
• Main Text
• End Matter
These must be title of the research topic and data. There must be preface of foreword
to the research work. It should be followed by table of contents. The list of tables, maps
should be given.
It provides the complete outline of research report along with all details. The title
page is reported in the main text. Details of text are given continuously as divided in different
chapters.
(a) Introduction
Introduction
Its purpose is to introduce the research topic to readers. It must cover statement of the
research problem, hypotheses, objectives of study, review of literature, and the methodology
to cover primary and secondary data, limitations of study and chapter scheme. Some may
give in brief in the first chapter the introduction of the research project highlighting the
importance of study. This is followed by research methodology in separate chapter.
The methodology should point out the method of study, the research design and
method of data collection.
This is crux of his research. It highlights main theme of his study. It must be in
nontechnical language. It should be in simple manner so ordinary reader may follow it. The
social research must be made available to common man. The research in agricultural
problems must be easy for farmers to read it.
121
Analysis of data
Data so collected should be presented in systematic manner and with its help,
conclusions can be drawn. This helps to test the hypothesis. Data analysis must be made to
confirm the objectives of the study.
Implications of Data
The results based on the analysis of data must be valid. This is the main body of
research. It contains statistical summaries and analysis of data. There should be logical
sequence in the analysis of data. The primary data may lead to establish the results. He must
have separate chapter on conclusions and recommendations. The conclusions must be based
on data analysis. The conclusions must be such which may lead to generalization and its
applicability in similar circumstances. The conditions of research work limiting its scope for
generalization must be made clear by the researcher.
Summary
End Matter
122
It is the first step which is primarily concerned with the development of a subject.
There are two ways in which to develop a subject (a) logically and (b) chronologically. The
logical development is made on the basis of mental connections and associations between the
one thing and another by means of analysis. Chronological development is based on a
connection or sequence in time or occurrence. The directions for doing or making something
usually follow the chronological order.
It is next step in writing the research report “Outlines are the framework upon which
long written works are constructed. They are an aid to the logical organization of the material
and a reminder of the points to be stressed in the report.
This follows the logical analysis of the subject and the preparation of the final
outlines. Such a step is of utmost importance for the researchers now sit to write down what
he has done in the context of his research study. He will write down the procedure adopted by
him in collecting the material for his study along with various limitations faced by him, the
technique of analysis adopted by him, the broad findings and generalizations and the various
suggestions he wants to offer regarding the problem concerned.
This step happens to be most difficult part of all formal writing. Usually this step
require more time than the writing of the rough draft. The careful revision makes the
difference between a mediocre and a good piece of writing. While rewriting and polishing,
one should check the report for weaknesses in logical development or presentation. The
researcher should also “see whether or not the material, as it is presented, has unity and
cohesion; does the report stand upright and firm and exhibit a definite pattern, like a marble
arch? Or does it resemble an old wall of moldering cement and loose brick.” In addition the
researcher should give due attention to the fact that in his rough draft he has been consistent
or not. He should check the mechanics of writing – grammar, spelling and usage.
Next in order comes the task of the preparation of the final bibliography. The
bibliography, which is generally appended to the research report, is a list of books in some
way pertinent to the research which has been done. It should contain all those works which
123
the researcher has consulted. The bibliography should be arranged alphabetically and may be
divided into two parts; the first part may contain the names of books and pamphlets, and the
second part may contain the names of magazine and newspaper articles.
This constitutes the last step. The final draft should be written in a concise and
objective style and in simple language, avoiding vague expressions such as “it seems”, “there
may be” and like ones. While writing the final draft, the researcher must avoid abstract
terminology and technical jargon. Illustrations and examples based on common experiences
must be incorporated in the final draft as they happen to be most effective in communicating
the research findings to others. A research report should not be dull, but must enthuse people
and maintain interest and must show originality. It must be remembered that every report
should be an attempt to solve some intellectual problem and must contribute to the solution of
a problem and must add to the knowledge of both the researcher and the research.
124
be decided as soon as the materials for the research paper have been assembled. The
following points deserve mention so far as the mechanics of writing a report are concerned:
Treatment of Quotations
The Footnotes
The footnotes serve two purposes viz., the identification of materials used in
quotations in the report and the notice of materials not immediately necessary to the body of
the research text but still of supplemental value. Footnotes are placed at the bottom of the
page on which the reference or quotation which they identify or supplement ends. Footnotes
should be numbered consecutively, usually beginning with in each chapter separately
Statistics are usually presented in the form of tables, charts, bars and line-graphs and
pictograms.
Revising and rewriting the rough draft of the report should be done with great care
before writing the final draft. For the purpose, the researcher should put to himself questions
like:
Are the sentences written in the report clear? Are they grammatically correct?
Do the various points incorporated in the report fit together logically?
125
At the end of the report, an index should invariably be given, the value of which lies
in the fact that it acts as a good guide, to the reader. Index may be prepared both as subject
index and as author index. The index should always be arranged alphabetically.
SELF ASSESSMENT
State whether the following statements are true or false:
1. Interpretation is the process of making sense of numerical data that has been
collected, analyzed, and presented.
2. Logical analysis of the subject matter is the first step in report writing.
3. The layout of research report means as to what the research report should contain.
SUMMARY:
Interpretation refers to the task of drawing inferences from the collected facts after an analytical and
or experimental study. The task of interpretation has two major aspects viz. the effort. It establish
continuity in research through linking the results of given study with those of another and the
establishment of some explanatory concepts. At the time of interpretation, the researcher must take
certain precautions.
A report is prepared after interpretation of data. It is defined as a document in which a given problem
is examined for the purpose of conveying information, report findings, putting forward ideas and
sometimes making recommendations. A research report can be classified into decision oriented report
and popular report. In the popular report, the main emphasis is on the methods employed, assumptions
made in the course of study and the detailed presentation of the findings including their limitations
and supporting data. The popular report is intended detail of research methods and terminology.
The contests of the research report must include preliminary pages, main text and the end matter.
Different steps involved in report writing are:
126
• Rewriting and polishing
The following points deserve mention as for as the mechanics of writing a report are concerned.
• Treatment of quotations
• The footnotes
7.15. GLOSSARY
Post factum – means occurring after the fact, or ex post facto design is a quasi-experimental
study examining how an independent variable, present prior to the study in the participants,
affects a dependent variable.
Quotation- quoting means repeating the author’s exact words. In some disciplines, such as
literary studies and history, quoting is used frequently to support the argument.
Generalisation – which is an act of reasoning that involves drawing broad inferences from
particular observations, is widely-acknowledged as a quality standard in quantitative research
but more controversial in qualitative research.
TERMINAL QUESTIONS:
127
4. Explain the layout of the research report.
SUGGESTED READINGS:
• Dubin, Robert, Theory Building New York; MacMillan Publishing Co., Quoted in W
Emory, Business Research Methods.
128
ASSIGENMENTS –
Attempt any four-
2. Explain the Research Design? Discuss the different types of Research Designs?
3. How would you proceed to construct a Likert scale? Give its advantages and disadvantages?
4. What do you mean by questionnaire? What is the difference between questionnaire and
schedule?
5. Define the term sampling? Discuss in brief probability and non probability methods of
sampling? Also explain various sampling and non sampling errors?
7. Explain the concept of correlation and regression? Discuss the difference between correlation
and regression?
8. What is report writing? Discuss the guidelines to be taken in to account for writing a report?
129