You are on page 1of 25

BUSINESS RESEARCH MS 108

MBA II, JUNE 2013


Q1 STEPS IN RESEARCH PROCESS
Scientific research involves a systematic process that focuses on being objective and gathering a
multitude of information for analysis so that the researcher can come to a conclusion. This
process is used in all research and evaluation projects, regardless of the research method
(scientific method of inquiry, evaluation research, or action research). The process focuses on
testing hunches or ideas in a park and recreation setting through a systematic process. In this
process, the study is documented in such a way that another individual can conduct the same
study again. This is referred to as replicating the study. Any research done without documenting
the study so that others can review the process and results is not an investigation using the
scientific research process. The scientific research process is a multiple-step process where the
steps are interlinked with the other steps in the process. If changes are made in one step of the
process, the researcher must review all the other steps to ensure that the changes are reflected
throughout the process. Parks and recreation professionals are often involved in conducting
research or evaluation projects within the agency. These professionals need to understand the
eight steps of the research process as they apply to conducting a study.
Step 1: Identify the Problem
The first step in the process is to identify a problem or develop a research question. The research
problem may be something the agency identifies as a problem, some knowledge or information
that is needed by the agency, or the desire to identify a recreation trend nationally. Let us
consider a problem that the agency has identified is childhood obesity, which is a local problem
and concern within the community. This serves as the focus of the study.
Step 2: Review the Literature
Now that the problem has been identified, the researcher must learn more about the topic under
investigation. To do this, the researcher must review the literature related to the research
problem. This step provides foundational knowledge about the problem area. The review of
literature also educates the researcher about what studies have been conducted in the past, how
these studies were conducted, and the conclusions in the problem area. In the obesity study, the
review of literature enables the programmer to discover horrifying statistics related to the longterm effects of childhood obesity in terms of health issues, death rates, and projected medical
costs. In addition, the programmer finds several articles and information from the Centers for
Disease Control and Prevention that describe the benefits of walking 10,000 steps a day. The
information discovered during this step helps the programmer fully understand the magnitude of
the problem, recognize the future consequences of obesity, and identify a strategy to combat
obesity (i.e., walking).

Step 3: Clarify the Problem


Many times the initial problem identified in the first step of the process is too large or broad in
scope. In step 3 of the process, the researcher clarifies the problem and narrows the scope of the
study. This can only be done after the literature has been reviewed. The knowledge gained
through the review of literature guides the researcher in clarifying and narrowing the research
project. In the example, the programmer has identified childhood obesity as the problem and the
purpose of the study. This topic is very broad and could be studied based on genetics, family
environment, diet, exercise, self-confidence, leisure activities, or health issues. All of these areas
cannot be investigated in a single study; therefore, the problem and purpose of the study must be
more clearly defined. The programmer has decided that the purpose of the study is to determine
if walking 10,000 steps a day for three days a week will improve the individuals health. This
purpose is more narrowly focused and researchable than the original problem.
Step 4: Clearly Define Terms and Concepts
Terms and concepts are words or phrases used in the purpose statement of the study or the
description of the study. These items need to be specifically defined as they apply to the study.
Terms or concepts often have different definitions depending on who is reading the study. To
minimize confusion about what the terms and phrases mean, the researcher must specifically
define them for the study. In the obesity study, the concept of individuals health can be
defined in hundreds of ways, such as physical, mental, emotional, or spiritual health. For this
study, the individuals health is defined as physical health. The concept of physical health may
also be defined and measured in many ways. In this case, the programmer decides to more
narrowly define individual health to refer to the areas of weight, percentage of body fat, and
cholesterol. By defining the terms or concepts more narrowly, the scope of the study is more
manageable for the programmer, making it easier to collect the necessary data for the study. This
also makes the concepts more understandable to the reader.
Step 5: Define the Population
Research projects can focus on a specific group of people, facilities, park development,
employee evaluations, programs, financial status, marketing efforts, or the integration of
technology into the operations. For example, if a researcher wants to examine a specific group of
people in the community, the study could examine a specific age group, males or females, people
living in a specific geographic area, or a specific ethnic group. Literally thousands of options are
available to the researcher to specifically identify the group to study. The research problem and
the purpose of the study assist the researcher in identifying the group to involve in the study. In
research terms, the group to involve in the study is always called the population. Defining the
population assists the researcher in several ways. First, it narrows the scope of the study from a
very large population to one that is manageable. Second, the population identifies the group that
the researchers efforts will be focused on within the study. This helps ensure that the researcher

stays on the right path during the study. Finally, by defining the population, the researcher
identifies the group that the results will apply to at the conclusion of the study. Researcher has
identified the population of the study as children ages 10 to 12 years. This narrower population
makes the study more manageable in terms of time and resources.
Step 6: Develop the Instrumentation Plan
The plan for the study is referred to as the instrumentation plan. The instrumentation plan serves
as the road map for the entire study, specifying who will participate in the study; how, when, and
where data will be collected; and the content of the program. In the obesity study, the researcher
has decided to have the children participate in a walking program for six months. The group of
participants is called the sample, which is a smaller group selected from the population specified
for the study. The study cannot possibly include every 10 to 12-year-old child in the community,
so a smaller group is used to represent the population. The researcher develops the plan for the
walking program, indicating what data will be collected, when and how the data will be
collected, who will collect the data, and how the data will be analyzed. The instrumentation plan
specifies all the steps that must be completed for the study. This ensures that the programmer has
carefully thought through all these decisions and that she provides a step-by-step plan to be
followed in the study.
Step 7: Collect Data
Once the instrumentation plan is completed, the actual study begins with the collection of data.
The collection of data is a critical step in providing the information needed to answer the
research question. Every study includes the collection of some type of datawhether it is from
the literature or from subjectsto answer the research question. Data can be collected in the
form of words on a survey, with a questionnaire, through observations, or from the literature. In
the obesity study, the programmers will be collecting data on the defined variables: weight,
percentage of body fat, cholesterol levels, and the number of days the person walked a total of
10,000 steps during the class. The researcher collects these data at the first session and at the last
session of the program. These two sets of data are necessary to determine the effect of the
walking program on weight, body fat, and cholesterol level. Once the data are collected on the
variables, the researcher is ready to move to the final step of the process, which is the data
analysis.
Step 8: Analyze the Data
All the time, effort, and resources dedicated to steps 1 through 7 of the research process
culminate in this final step. The researcher finally has data to analyze so that the research
question can be answered. In the instrumentation plan, the researcher specified how the data will
be analyzed. The researcher now analyzes the data according to the plan. The results of this
analysis are then reviewed and summarized in a manner directly related to the research
questions. In the obesity study, the researcher compares the measurements of weight, percentage

of body fat, and cholesterol that were taken at the first meeting of the subjects to the
measurements of the same variables at the final program session. These two sets of data will be
analyzed to determine if there was a difference between the first measurement and the second
measurement for each individual in the program. Then, the data will be analyzed to determine if
the differences are statistically significant. If the differences are statistically significant, the study
validates the theory that was the focus of the study. The results of the study also provide valuable
information about one strategy to combat childhood obesity in the community.
As you have probably concluded, conducting studies using the eight steps of the scientific
research process requires you to dedicate time and effort to the planning process. Researcher
cannot conduct a study using the scientific research process when time is limited or the study is
done at the last minute. Researchers who do this conduct studies that result in either false
conclusions or conclusions that are not of any value to the organization.
Q2a) Questionnaire: A questionnaire is a research instrument consisting of a series
of questions and other prompts for the purpose of gathering information from respondents.
Although they are often designed for statistical analysis of the responses, this is not always the
case. Questionnaires have advantages over some other types of surveys in that they are cheap, do
not require as much effort from the questioner as verbal or telephone surveys, and often have
standardized answers that make it simple to compile data. A distinction can be made between
questionnaires with questions that measure separate variables, and questionnaires with questions
that are aggregated into either a scale or index. Questionnaires within the former category are
commonly part of surveys, whereas questionnaires in the latter category are commonly part of
tests.
Questionnaires with questions that measure separate variables could for instance include
questions on:

preferences (e.g. political party)

behaviors (e.g. food consumption)

facts (e.g. gender)

Questionnaires with questions that are aggregated into either a scale or index, include for
instance questions that measure:

latent traits (e.g. personality traits such as extroversion)

attitudes (e.g. towards immigration)

an index (e.g. Social Economic Status)

Q2b) Observation: Observation is way of gathering data by watching behavior, events, or noting
physical characteristics in their natural setting. Observations can be overt (everyone knows they
are being observed) or covert (no one knows they are being observed and the observer is
concealed). The benefit of covert observation is that people are more likely to behave naturally if
they do not know they are being observed. However, you will typically need to conduct overt
observations because of ethical problems related to concealing your observation. Observations
can also be either direct or indirect. Direct observation is when you watch interactions,
processes, or behaviors as they occur; for example, observing a teacher teaching a lesson from a
written curriculum to determine whether they are delivering it with fidelity. Indirect observations
are when you watch the results of interactions, processes, or behaviors; for example, measuring
the amount of plate waste left by students in a school cafeteria to determine whether a new food
is acceptable to them.
Observation method is used in the following condtions:
Trying to understand ongoing process or situation: Through observation a process can be
monitored or a situation can be evaluated as it occurs.
Gathering data on individual behaviors or interactions between people: Observation allows to
watch peoples behaviors and interactions directly, or watch for the results of behaviors or
interactions.
Physical setting: Seeing the place or environment where something takes place can help increase
understanding of the event, activity, or situation
Evaluating: For example, it can be used to observe whether a classroom or training facility is
conducive to learning.
When data collection from individuals is not a realistic option. If respondents are unwilling or
unable to provide data through questionnaires or interviews, observation is a method that
requires little from the individuals from where data need to be collected.
Q3 TYPES OF SAMPLING DESIGN
The sampling design can be broadly grouped on two basis viz., representation and element
selection. Representation refers to the selection of members on a probability or by other means.
Element selection refers to the manner in which the elements are selected individually and
directly from the population. If each element is drawn individually from the population at large,
it is an unrestricted sample. Restricted sampling is where additional controls are imposed, in
other words it covers all other forms of sampling. The classification of sampling design on the
basis of representation and element selection is shown below:
Element

Representation Basis

Selection
Unrestricted

Restricted

Probability
Simple random
Complex random
Systematic
Stratified
Cluster
Double

Non-probability
Convenience
Purposive
Judgment
Quota
Snowball

PROBABILITY SAMPLING
Probability sampling is where each sampling unit in the defined target population has a known
nonzero probability of being selected in the sample. The actual probability of selection for each
sampling unit may or may not be equal depending on the type of probability sampling design
used. Specific rules for selecting members from the operational population are made to ensure
unbiased selection of the sampling units and proper sample representation of the defined target
population. The results obtained by using probability-sampling designs can be generalized to the
target population within a specified margin of error. The different types of probability sampling
designs are discussed below;
Unrestricted or Simple Random Sampling
In the unrestricted probability sampling design every element in the population has a known,
equal nonzero chance of being selected as a subject. For example, if 10 employees (n = 10) are to
be selected from 30 employees (N = 30), the researcher can write the name of each employee in
a piece of paper and select them on a random basis. Each employee will have an equal known
probability of selection for a sample. The same is expressed in terms of the following formula;
Probability of selection = Size of sample
-------------------------Size of population
Each employee would have a 10/30 or .333 chance of being randomly selected in a drawn
sample. When the defined target population consists of a larger number of sampling units, a more
sophisticated method can be used to randomly draw the necessary sample. A table of random
numbers can be used for this purpose. The table of random numbers contains a list of randomly
generated numbers. The numbers can be randomly generated through the computer programs
also. Using the random numbers the sample can be selected.
Restricted or Complex Probability Sampling
As an alternative to the simple random sampling design, several complex probability sampling
design can be used which are more viable and effective. Efficiency is improved because more

information can be obtained for a give sample size using some of the complex probability
sampling procedures than the simple random sampling design. The five most common complex
probability sampling designs viz., systematic sampling, stratified random sampling, cluster
sampling, area sampling and double sampling are discussed below:
Systematic random sampling
The systematic random sampling design is similar to simple random sampling but requires that
the defined target population should be ordered in some way. It involves drawing every nth
element in the population starting with a randomly chosen element between 1 and n. In other
words individual sampling units are selected according their position using a skip interval. The
skip interval is determined by dividing the sample size into population size. For e.g. if the
researcher wants a sample of 100 to be drawn from a defined target population of 1000, the skip
interval would be 10(1000/100). Once the skip interval is calculated, the researcher would
randomly select a starting point and take every 10 th until the entire target population is proceeded
thorough. The steps to be followed in a systematic sampling method are enumerated below;
Total number of elements in the population should be identified
The sampling ratio is to be calculated ( n = total population size divided by size of the
desired sample)
The random start should be identified
A sample can be drawn by choosing every nth entry

Two important considerations in using the systematic random sampling are;


It is important that the natural order of the defined target population list be unrelated to
the characteristic being studied.
Skip interval should not correspond to the systematic change in the target population.

Stratified Random Sampling


Stratified random sampling requires the separation of defined target population into different
groups called strata and the selection of sample from each stratum. Stratified random sampling is
very useful when the divisions of target population are skewed or when extremes are present in
the probability distribution of the target population elements of interest. The goal in stratification
is to minimize the variability within each stratum and maximize the difference between strata.
The ideal stratification would be based on the primary variable under study. Researchers often
have several important variables about which they want to draw conclusion. A reasonable
approach is to identify some basis for stratification that correlates well with other major
variables. It might be a single variable like age, income etc or a compound variable like on the
basis of income and gender.

Stratification leads to segmenting the population into smaller, more homogeneous sets of
elements. In order to ensure that the sample maintains the required precision in terms of
representing the total population, representative samples must be drawn from each of the smaller
population groups.
There are three reasons as to why a researcher chooses a stratified random sample;
To increase the samples statistical efficiency
To provide adequate data for analyzing various sub population
To enable different research methods and procedures to be used in different strata.
Drawing a stratified random sampling involves the following steps;

Determine the variables to use for stratification

Select proportionate or disproportionate stratification

Divide the target population into homogeneous subgroups or strata

Select random samples from each stratum

Combine the samples from each stratum into a single sample of the target population.
There are two common methods for deriving samples from the strata viz., proportionate and
disproportionate. In proportionate stratified sampling, each stratum is properly represented so
the sample drawn from it is proportionate to the stratums share of the total population. The
larger strata are sampled more because they make up a larger percentage of the target population.
This approach is more popular than any other stratified sampling procedures due to the following
reasons;
It has higher statistical efficiency than the simple random sample
It is much easier to carry out than other stratifying methods
It provides a self-weighting sample ie the population mean or proportion can be
estimated simply by calculating the mean or proportion of all sample cases.
In disproportionate stratified sampling, the sample size selected from each stratum is
independent of that stratums proportion of the total defined target population. This approach is
used when stratification of the target population produces sample sizes that contradict their
relative importance to the study.
An alternative of disproportionate stratified method is optimal allocation. In this method,
consideration is given to the relative size of the stratum as well as the variability within the
stratum to determine the necessary sample size of each stratum. The logic underlying the optimal
allocation is that the greater the homogeneity of the prospective sampling units within a
particular stratum, the fewer the units that would have to be selected to estimate the true
population parameter accurately for that subgroup. This method is also opted for in situation

where it is easier, simpler and less expensive to collect data from one or more strata than from
others.
Cluster Sampling
Cluster sampling is a probability sampling method in which the sampling units are divided into
mutually exclusive and collectively exhaustive subpopulation called clusters. Each cluster is
assumed to be the representative of the heterogeneity of the target population. Groups of
elements that would have heterogeneity among the members within each group are chosen for
study in cluster sampling. Several groups with intergroup heterogeneity and intergroup
homogeneity are found. A random sampling of the clusters or groups is done and information is
gathered from each of the members in the randomly chosen clusters. Cluster sampling offers
more of heterogeneity within groups and more homogeneity among the groups.
Single Stage and Multistage Cluster Sampling
In single stage cluster sampling, the population is divided into convenient clusters and required
numbers of clusters are randomly chosen as sample subjects. Each element in each of the
randomly chosen cluster is investigated in the study. Cluster sampling can also be done in several
stages, which is known as multistage cluster sampling. For example to study the banking
behaviour of customers in a national survey, cluster sampling can be used to select the urban,
semiruban and rural geographical locations of the study. At the next stage, particular areas in
each of the location would be chosen. At the third stage, the banks within each area would be
chosen. Thus multi stage sampling involves a probability sampling of the primary sampling
units; from each of the primary units, a probability sampling of the secondary sampling units is
drawn; a third level of probability sampling is done from each of these secondary units, and so
on until the final stage of breakdown for the sample units are arrived at, where every member of
the unit will be a sample.
Area Sampling
Area sampling is a form of cluster sampling in which the clusters are formed by geographic
designations. For example, state, district, city, town etc., Area sampling is a form of cluster
sampling in which any geographic unit with identifiable boundaries can be used. Area sampling
is less expensive than most other probability designs and is not dependent on population frame.
A city map showing blocks of the city would be adequate information to allow a researcher to
take a sample of the blocks and obtain data from the residents therein.
Stratified random sampling Vs Cluster sampling
The cluster sampling differs from stratified sampling in the following manner;

In stratified sampling the population is divided into a few subgroups, each with many
elements in it and the subgroups are selected according to some criterion that is related to
the variables under the study. In cluster sampling the population is divided into many
subgroups each with a few elements in it. The subgroups are selected according to some
criterion of ease or availability in data collection.
Stratified sampling should secure homogeneity within the subgroups and heterogeneity
between subgroups. Cluster sampling tries to secure heterogeneity within subgroups and
homogeneity between subgroups.
The elements are chosen randomly within each subgroup in stratified sampling. In cluster
sampling the subgroups are randomly chosen and each and every element of the subgroup
is studied in-depth.

Double sampling
This is also called sequential or multiphase sampling. Double sampling is opted when further
information is needed from a subset of group from which some information has already been
collected for the same study. It is called as double sampling because initially a sample is used in
the study to collect some preliminary information of interest and later a subsample of this
primary sample is used to examine the matter in more detail The process includes collecting data
from a sample using a previously defined technique. Based on this information, a sub sample is
selected for further study. It is more convenient and economical to collect some information by
sampling and then use this information as the basis for selecting a sub sample for further study.
NON-PROBABILITY SAMPLING
In non-probability sampling method, the elements in the population do not have any probabilities
attached to being chosen as sample subjects. This means that the findings of the study cannot be
generalized to the population. However at times the researcher may be less concerned about
generalizability and the purpose may be just to obtain some preliminary information in a quick
and inexpensive way. Sometime when the population size is unknown, then non-probability
sampling would be the only way to obtain data. Some non-probability sampling technique may
be more dependable than others and could often lead to important information with regard to the
population. The non-probability sampling designs are discussed below;
Convenience sampling
Non-probability samples that are unrestricted are called convenient sampling. Convenience
sampling refers to the collection of information from members of population who are
conveniently available to provide it. Researchers or field workers have the freedom to choose as
samples whomever they find thus it is named as convenience. It is mostly used during the

exploratory phase of a research project and it is the best way of getting some basic information
quickly and efficiently. The assumptions is that the target population is homogeneous and the
individuals selected as samples are similar to the overall defined target population with regard to
the characteristics being studied. However in reality there is no way to accurately assess the
representativeness of the sample. Due to the self-selection and voluntary nature of participation
in data collection process the researcher should give due consideration to the non-response error.
Advantages and disadvantages
Convenient sampling allows a large number of respondents to be interviewed in a relatively short
time. This is one of the main reasons for using convenient sampling in the early stages of
research. However the major drawback is that the use of convenience samples in the
development phases of constructs and scale measurements can have a serious negative impact on
the overall reliability and validity of those measures and instruments used to collect raw data.
Another major drawback is that the raw data and results are not generalizable to the defined
target population with any measure of precision. It is not possible to measure the
representativeness of the sample, because sampling error estimates cannot be accurately
determined.
Purposive Sampling
A non-probability sample that conforms to certain criteria is called purposive sampling. There
are two major types of purposive sampling viz.., Judgment sampling and Quota sampling.
Judgment sampling
Judgment sampling is a non-probability sampling method in which participants are selected
according to an experienced individuals belief that they will meet the requirements of the study.
The researcher selects sample members who conform to some criterion. It is appropriate in the
early stages of an exploratory study and involves the choice of subjects who are most
advantageously placed or in the best position to provide the information required. This is used
when a limited number or category of people have the information that are being sought. The
underlying assumption is that the researchers belief that the opinions of a group of perceived
experts on the topic of interest are representative of the entire target population.
Quota sampling
The quota sampling method involves the selection of prospective participants according to
prespecified quotas regarding either the demographic characteristics (gender, age, education,
income, occupation etc.,) specific attitudes (satisfied, neutral, dissatisfied) or specific behaviours
(regular, occasional, rare user of product) .The purpose of quota sampling is to provide an

assurance that prespecified subgroups of the defined target population are represented on
pertinent sampling factors that are determined by the researcher. It ensures that certain groups are
adequately represented in the study though the assignment of the quota.
Snowball Sampling
Snowball sampling is a non-probability sampling method in which sets of respondents are
chosen who helps the researcher to identify additional respondents to be included in the study.
This method of sampling is also called as referral sampling because one respondent refers other
potential respondents. Snowball sampling is typically used in research situations where the
defined target population is very small and unique and compiling a complete list of sampling
units is a nearly impossible task. While the traditional probability and other non-probability
sampling methods would normally require an extreme search effort to qualify a sufficient
number of prospective respondents, the snowball method would yield better result at a much
lower cost. The researcher has to identify and interview one qualified respondent and then solicit
his help to identify other respondents with similar characteristics.
Q4a) Survey Errors: Errors may occur at any stage during the collection and processing of
survey data, whether it is a census or a sample survey. There are two main sources of survey
error: Sampling error (errors associated directly with the sample design and estimation methods
used) and non-sampling error (a blanket term used to cover all other errors). Non-sampling errors
are usually sub-divided as follows:

Coverage errors, which are mainly associated with the sampling frame, such as missing
units, inclusion of units not in the population of interest, and duplication.

Response errors, which are caused by problems related to the way questions were
phrased, the order in which the questions were asked, or respondents' reporting errors
(also referred to as measurement error if possible errors made by the interviewer are
included in this category).

Non-response errors, which are due to respondents either not providing information or
providing incorrect information. Non-response increases the likelihood of bias in the
survey estimates. It also reduces the effective sample size, thereby increasing the
observed sampling error. However, the risk of bias when non-response rates are high is
generally more dangerous than the reduction in sample size per se.

Data capture errors, which are due to coding or data entry problems.

Edit and imputation ("E&I") errors, which can be introduced during attempts to find and
correct all the other non-sampling errors.

All of these sources may contribute to either, or both, of the two types of survey error. These are
bias, or systematic error, and variance, or random error.
Sampling error is not an error in the sense of a mistake having been made in conducting the
survey. Rather it indicates the degree of uncertainty about the 'true' value based on information
obtained from the number of people that were surveyed.
It is reasonably straightforward for knowledgeable, experienced survey-taking organizations to
control sampling error through the use of suitable sampling methods and to estimate its impact
using information from the sample design and the achieved sample. Any statement about
sampling errors, namely variance, standard error, margin of sampling error or coefficient of
variation, can only be made if the survey data come from a probability sample.
The non-sampling errors, especially potential biases, are the most difficult to detect, to control
and to measure, and require careful planning, training and testing.
Q4b) Reliability: Internal consistency reliability is used to assess the reliability of a summated
scale where several items are summed to form a total score. In internal consistency reliability
estimation a single measurement instrument is administered to a group of people on one occasion
to estimate reliability. In effect the reliability of the instrument is judged by estimating how well
the items that reflect the same construct yield similar results. We are looking at how consistent
the results are for different items for the same construct within the measure.
There are a wide variety of internal consistency measures that can be used:
i.

Average Inter-item Correlation

The average inter-item correlation uses all of the items on the instrument that are designed to
measure the same construct. The correlation between each pair of items is computed first. For
example, if we have six items we will have 15 different item pairings (i.e., 15 correlations). The
average inter item correlation is simply the average or mean of all these correlations.
ii.

Average Item total Correlation

This approach also uses the inter-item correlations. In addition, a total score for the six items is
computed and use that as a seventh variable in the analysis. The figure shows the six item-tototal correlations at the bottom of the correlation matrix.
iii.

Split-Half Reliability

In split-half reliability all items that purport to measure the same construct into two sets are
randomly divided. The entire instrument is administered to a sample of people and the total score
for each randomly divided half is calculated.

iv.

Cronbach's Alpha ()

Cronbachs alpha is a statistic. It is generally used as a measure of internal consistency or


reliability of a psychometric instrument. In other words, it measures how well a set of variables
or items measures a single, one-dimensional latent aspect of individuals.
The value of alpha () may lie between negative infinity and 1. However only positive values of
make sense. Generally, coefficient ranges in value from 0 to 1 and may be used to describe
the reliability of factors extracted from dichotomous (that is, questions with two possible
answers) and/or multi-point formatted questionnaires or scales (i.e., rating scale: 1 = poor, 5 =
excellent).
Validity: The validity of a scale may be defined as the extent to which differences in observed
scale scores reflect true differences among objects on the characteristic being measured, rather
than systematic or random error. Perfect validity requires that there be no measurement error
(XO = XT, XR = 0, XS = 0).

Content validity is a subjective but systematic evaluation of how well the content of a
scale represents the measurement task at hand.

Criterion validity reflects whether a scale performs as expected in relation to other


variables selected (criterion variables) as meaningful criteria.

Construct validity addresses the question of what construct or characteristic the scale is,
in fact, measuring.
Construct validity includes convergent, discriminant, and
nomological validity.

Convergent validity is the extent to which the scale correlates positively with other
measures of the same construct.

Discriminant validity is the extent to which a measure does not correlate with other
constructs from which it is supposed to differ.

Nomological validity is the extent to which the scale correlates in theoretically predicted
ways with measures of different but related constructs.

Q5 The amount of data that can be collected and assembled in a business research study can be
astronomical. Data organization and reduction are two very important aspects of data analysis
that is seldom highlighted. Yet, these steps are crucial to the ability to make sense out of data and
to the ability to make cogent and insightful data interpretation. An impressive array of methods
fordata organization and data reduction are available.
A business researcher may tabulate data or compile frequency. The means or averages and other
measures of dispersion are common ways of analyzing data for which frequency distributions are
available. Very often, advanced statistics and decision models are used to maximize the
information that can be extracted from research data. The following section provides a brief

description of several commonly used statistical tools, decision support models, and optimization
routines
Quantitative Market Research Decision Support Tools
Statistical Methods

Multiple Regression - This statistical procedure is used to estimate the equation with the
best fit for explaining how the value of a dependent variable changes as the values of a
number of independent variables shift. A simple market research example is the estimation of
the best fit for advertising by looking at how sales revenue (the dependent variable) changes
in relation to expenditures on advertising, placement of ads, and timing of ads.

Factor Analysis - This statistical method is used to determine which are the strongest
underlying dimensions of a larger set of variables that are inter-correlated. Where many
variables are correlated, factor analysis identifies which relations are strongest. A Using factor
analysis, a market researcher who wants to know what combination of variables or factors are
most appealing to a particular type of consumer can use factor analysis to reduce the data
down to a few variables are most appealing to consumers.
Conditions for a Factor Analysis Exercise:
It requires metric data. This means that the data should be either interval or ratio scale
in nature.
The size of the sample respondents should be at least four to five times more than the
number of variables.
Initial set of variables should be highly correlated. A correlation matrix of the variables
could be computed and tested for its statistical significance. The test is carried out by
using a Barttlet test of sphericity, which takes the determinant of the correlation matrix
into consideration.
Kaiser-Meyer-Olkinis carried out before performing factor analysis, it takes a value
between 0 and 1.The KMO statistics compares the magnitude of observed correlation
coefficients with magnitudes of partial correlation coefficients.

Discriminant Analysis - This statistical technique is used to for classification of people,


products, or other tangibles into two or more categories. Market research can make use of
discriminant analyses in a number of ways. One simple example is to distinguish
whatadvertising channels are most effective for different types of products.
Major assumptions of Discriminant Analysis are:
the observations are a random sample.
each predictor variable is normally distributed;
each of the allocations for the dependent categories in the initial classi cation are
correctly classi ed;
there must be at least two groups or categories, with each case belonging to only one
group so that the groups are mutually exclusive and collectively exhaustive (all cases
can be placed in a group);
each group or category must be well de ned, clearly differentiated from any other
group(s) and natural. Putting a median split on an attitude scale is not a natural way to

form groups. Partitioning quantitative variables is only justi able if there are easily
identi able gaps at the points of division;
for instance, three groups taking three available levels of amounts of housing loan;
the groups or categories should be de ned before collecting the data;
the attribute(s) used to separate the groups should discriminate quite clearly between
the groups so that group or category overlap is clearly non-existent or minimal;
group sizes of the dependent should not be grossly different and should be at least ve
times the number of independent variables.
Discriminat Analysis is used when:
the dependent is categorical with the predictor IVs at interval level such as age,
income,
attitudes, perceptions, and years of education, although dummy variables can be used
as predictors as in multiple regression. Logistic regression IVs can be of any level of
measurement.
there are more than two DV categories, unlike logistic regression, which is limited to a
dichotomous dependent variable.

Cluster Analysis - This statistical procedure is used to separate objects into a specific
number of groups that are mutually exclusive but that are also relatively homogeneous in
constitution. This process is similar to what occurs in market segmentationwhere the market
researcher is interested in the similarities that facilitate grouping consumers into segments and
is also interested in the attributes that make the market segments distinct.For example, You
need to identify people with similar patterns of past purchases so that you can tailor your
marketing strategies.
The objective of cluster analysis is to identify groups of object that are very similar with regard
to their price consciousness and brand loyalty and assign them into clusters. After having decided
on the clustering variables (brand loyalty and price consciousness), we need to decide on the
clustering procedure to form our groups of objects. This step is crucial for the analysis, as
different procedures require different decisions prior to analysis.
Conjoint Analysis - This statistical method is used to unpack the preferences of consumerswith
regard to different marketing offers. Two dimensions are of interest to the market researcher
in conjoint analysis: (1) The inferred utility functions of each attribute, and (2) the relative
importance of the preferred attributes to the consumers. For example a computer may be
described in terms of attributes such as processor type, hard disk size and amount of memory.
Each of these attributes is broken down into levels - for instance levels of the attribute for
memory size might be 1GB, 2GB, 3GB and 4GB.
These attributes and levels can be used to define different products or product profiles. The first
stage in conjoint analysis is to create a set of product profiles which customers or respondents are

then asked to compare and choose from. Obviously, the number of potential profiles increases
rapidly for every new attribute added, so there are techniques to simplify both the number of
profiles to be tested and the way in which preferences are discovered. Different type of conjoint
analysis (eg choice based conjoint, full-profile conjoint, or adaptive conjoint analysis) have
different approaches to coping with the balance between attribute number and amount of data
that needs to be collected.
By analysing which items are chosen or preferred from the product profiles offered to the
customer it is possible to work out statistically both what is driving the preference from the
attributes and levels shown, but more importantly, give an implicit numerical valuation for each
attribute and level - known as utilities or part-worths and importance scores.
The result is a detailed picture of how customers make decisions and a set of data that can be
used to build market models which can predict market share in new market conditions and test
the impact of product or service changes on the market to see where and how you can gain the
greatest improvements over your competitors. Basic assumptions of conjoint analysis are:

The product is a bundle of attributes

Utility of a product is a simple function of the utilities of the attributes

Utility predicts behavior (i.e., purchases)

Q6 Scale (social sciences)


In the social sciences, scaling is the process of measuring or ordering entities with respect to
quantitative attributes or traits. For example, a scaling technique might involve estimating
individuals' levels of extraversion, or the perceived quality of products. Certain methods of
scaling permit estimation of magnitudes on a continuum, while other methods provide only for
relative ordering of the entities.
See level of measurement for an account of qualitatively different kinds of measurement scales.

Comparative and non comparative scaling


With comparative scaling, the items are directly compared with each other (example : Do you
prefer Pepsi or Coke?). In noncomparative scaling each item is scaled independently of the
others (example : How do you feel about Coke?).
Composite measures
Composite measures of variables are created by combining two or more separate empirical
indicators into a single measure. Composite measures measure complex concepts more
adequately than single indicators, extend the range of scores available and are more efficient at
handling multiple items.
In addition to scales, there are two other types of composite measures. Indexes are similar to
scales except multiple indicators of a variable are combined into a single measure. The index of
consumer confidence, for example, is a combination of several measures of consumer attitudes.
A typology is similar to an index except the variable is measured at the nominal level.
Indexes are constructed by accumulating scores assigned to individual attributes, while scales are
constructed through the assignment of scores to patterns of attributes.
While indexes and scales provide measures of a single dimension, typologies are often employed
to examine the intersection of two or more dimensions. Typologies are very useful analytical
tools and can be easily used as independent variables, although since they are not unidimensional
it is difficult to use them as a dependent variable.
Data types
The type of information collected can influence scale construction. Different types of
information are measured in different ways. See in particular level of measurement.
1. Some data are measured at the nominal level. That is, any numbers used are mere labels:
they express no mathematical properties. Examples are SKU inventory codes and UPC
bar codes.
2. Some data are measured at the ordinal level. Numbers indicate the relative position of
items, but not the magnitude of difference. An example is a preference ranking.
3. Some data are measured at the interval level. Numbers indicate the magnitude of
difference between items, but there is no absolute zero point. Examples are attitude scales
and opinion scales.
4. Some data are measured at the ratio level. Numbers indicate magnitude of difference and
there is a fixed zero point. Ratios can be calculated. Examples include: age, income,
price, costs, sales revenue, sales volume, and market share.

Comparative scaling techniques

Pairwise comparison scale - a respondent is presented with two items at a time and
asked to select one (example: Do you prefer Pepsi or Coke?). This is an ordinal level
technique when a measurement model is not applied.

Rank-order scale - a respondent is presented with several items simultaneously and


asked to rank them (example: Rate the following advertisements from 1 to 10.). This is an
ordinal level technique.

Constant sum scale - a respondent is given a constant sum of money, script, credits, or
points and asked to allocate these to various items (example : If you had 100 Yen to spend
on food products, how much would you spend on product A, on product B, on product C,
etc.). This is an ordinal level technique.

Q-Sort scale - Up to 140 items are sorted into groups based a rank-order procedure.

Guttman scale - This is a procedure to determine whether a set of items can be rankordered on a unidimensional scale. It utilizes the intensity structure among several
indicators of a given variable. Statements are listed in order of importance. The rating is
scaled by summing all responses until the first negative response in the list. The Guttman
scale is related to Rasch measurement; specifically, Rasch models bring the Guttman
approach within a probabilistic framework.

Non-comparative scaling techniques

Continuous rating scale (also called the graphic rating scale) - respondents rate items by
placing a mark on a line. The line is usually labeled at each end. There are sometimes a
series of numbers, called scale points, (say, from zero to 100) under the line. Scoring and
codification is difficult.

Likert scale - Respondents are asked to indicate the amount of agreement or


disagreement (from strongly agree to strongly disagree) on a five- or seven-point scale.
The same format is used for multiple questions.

Phrase completion scales - Respondents are asked to complete a phrase on an 11-point


response scale in which 0 represents the absence of the theoretical construct and 10
represents the theorized maximum amount of the construct being measured. The same
basic format is used for multiple questions.

Semantic differential scale - Respondents are asked to rate on a 7 point scale an item on
various attributes. Each attribute requires a scale with bipolar terminal labels.

Stapel scale - This is a unipolar ten-point rating scale. It ranges from +5 to -5 and has no
neutral zero point.

Thurstone scale - This is a scaling technique that incorporates the intensity structure
among indicators.

The following table classifies the various simple data types, associated distributions,
permissible operations, etc. Regardless of the logical possible values, all of these data types
are generally coded using real numbers, because the theory of random variables often
explicitly assumes that they hold real numbers.

Data Type

binary

Possible
values

0, 1
(arbitrary
labels)

1, 2, ..., K
categorica
(arbitrary
l
labels)

ordinal

integer orre
al
number(arb
itrary

Scale
Level
of
of
Example
relativ Permissible
measu Distribution
usage
e
statistics
remen
differ
t
ences

binary
outcome
("yes/no",
"true/false
",
"success/f
ailure",
etc.)

categorica
l outcome
(specific
blood
type,politi
cal party,
word,
etc.)

logistic,prob
it

Bernoulli

nomin
al
scale

Regression
analysis

incom
mode, Chiparabl
squared
e

categorical

relative
ordinal categorical?? relativ
score,
scale
e
significan
compa
t only for

multinomial
logit,multino
mial probit

ordinal
regression(or
dered
logit,ordered

scale)

binomial 0, 1, ..., N

count

creating a
ranking

probit)

number of
successes
mean, median,
interva binomial,beta
binomial
(e.g. yes
additiv mode,standard
l
-binomial,
regression(lo
votes) out
e??
deviation,corr
scale?? etc.
gistic,probit)
ofN possi
elation
ble

number of
items
(telephon
e calls,
people,
nonnegativ molecules
ratio
eintegers (0 , births,
scale
, 1, ...)
deaths,
etc.) in
given
interval/ar
ea/volum
e

realreal
valuedad number
ditive

rison

All statistics
permitted for
interval scales
plus the
Poisson,negat multip
following:geo
ive binomial, licativ
metric
etc.
e
mean,harmoni
c
mean,coefficie
nt of variation

temperatu interva normal, etc.


re,
l scale (usually
relative
symmetric
distance, l
about
ocation
themean)
parameter
, etc. (or
approxim
ately,
anything
not
varying

Poisson,
negative
binomial
regression

additiv mean, median, standardline


e
mode,standard ar regression
deviation,corr
elation

over a
large
scale)

price,
income,
size,scale
parameter
real, etc.
valuedmu positive rea
ratio
(especiall
ltiplicativ l number
scale
y when
e
varying
over a
large
scale)

All statistics
permitted for
loginterval scales
normal,gamm
plus the
multip
a,exponential,
following:geo
licativ
etc. (usually
metric
e
a skeweddistr
mean,harmoni
ibution)
c
mean,coefficie
nt of variation

generalized
linear
modelwithlo
garithmiclin
k

Q7a) ETHICS IN RESEARCH


Research that involves human subjects or participants raises unique and complex ethical,
legal, social and political issues. Research ethics is specifically interested in the analysis of
ethical issues that are raised when people are involved as participants in research. There are
three objectives in research ethics. The first and broadest objective is to protect human
participants. The second objective is to ensure that research is conducted in a way that
serves interests of individuals, groups and/or society as a whole. Finally, the third objective
is to examine specific research activities and projects for their ethical soundness, looking at
issues such as the management of risk, protection of confidentiality and the process of
informed consent.
For the most part, research ethics has traditionally focused on issues in biomedical research.
The application of research ethics to examine and evaluate biomedical research has been
well developed over the last century and has influenced much of the existing statutes and
guidelines for the ethical conduct of research. However in humanities and social science
research, different kinds of ethical issues arise. New and emerging methods of conducting
research, such as auto-ethnography and participatory action research raise important but
markedly different ethical issues and obligations for researchers.
Research involving vulnerable persons, which may include children, persons with
developmental or cognitive disabilities, persons who are institutionalized, the homeless or
those without legal status, also raises unique issues in any research context.

Research ethicists everywhere today are challenged by issues that reflect global concerns in
other domains, such as the conduct of research in developing countries, the limits of
research involving genetic material and the protection of privacy in light of advances in
technology and Internet capabilities.
Q7b) A proper research report includes the following sections, submitted in the order listed,
each section to start on a new page. Some journals request a summary to be placed at the end of
the discussion. Some techniques articles include an appendix with equations, formulas,
calculations, etc. Some journals deviate from the format, such as by combining results and
discussion, or combining everything but the title, abstract, and literature as is done in the
journal Science. Reports will adhere to the following standard format:

Title

Abstract

Introduction

Experimental Details or Theoretical Analysis

Results

Discussion

Conclusions and Summary

References

Title and Title Page


The title should reflect the content and emphasis of the project described in the report. It should
be as short as possible and include essential key words.
Abstract
The abstract should, in the briefest terms possible, describe the topic, the scope, the principal
findings, and the conclusions. It should be written last to reflect accurately the content of the
report. The lengths of abstracts vary, but seldom exceed 200-300 words.A primary objective of
an abstract is to communicate to the reader the essence of the paper.
Introduction
"A good introduction is a clear statement of the problem or project and why you are studying it."
The nature of the problem and why it is of interest should be conveyed in the opening
paragraphs. This section should describe clearly but briefly the background information on the

problem, what has been done before (with proper literature citations), and the objectives of the
current project. A clear relationship between the current project and the scope and limitations of
earlier work should be made so that the reasons for the project and the approach used will be
understood.
Experimental Details or Theoretical Analysis
This section should describe what was actually done. It is a succinct exposition of the laboratory
notebook, describing procedures, techniques, instrumentation, special precautions, and so on. It
should be sufficiently detailed that other experienced researchers would be able to repeat the
work and obtain comparable results.
In theoretical reports, this section would include sufficient theoretical or mathematical analysis to
enable derivations and numerical results to be checked. Computer programs from the public
domain should be cited. New computer programs should be described in outline form.
If the experimental section is lengthy and detailed, as in synthetic work, it can be placed at the
end of the report or as an appendix so that it does not interrupt the conceptual flow of the report.
Its placement will depend on the nature of the project and the discretion of the writer.
Results
In this section, relevant data, observations, and findings are summarized. Tabulation of data,
equations, charts, and figures can be used effectively to present results clearly and concisely.
Schemes to show reaction sequences may be used here or elsewhere in the report.
Discussion
The crux of the report is the analysis and interpretation of the results. What do the results mean?
How do they relate to the objectives of the project? To what extent have they resolved the
problem? Because the "Results" and "Discussion" sections are interrelated, they can often be
combined as one section.
Conclusions and Summary
A separate section outlining the main conclusions of the project is appropriate if conclusions
have not already been stated in the 'Discussion' section. Directions for future work are also
suitably expressed here.
A lengthy report, or one in which the findings are complex, usually benefits from a paragraph
summarizing the main features of the reportthe objectives, the findings, and the conclusions.
The last paragraph of text in manuscripts prepared for publication is customarily dedicated to
acknowledgments. However, there is no rule about this, and research reports or senior theses
frequently place acknowledgments following the title page.

Citing References
Literature references should be collated at the end of the report and cited in one of the formats
described in The ACS Style Guide or standard journals. Do not mix formats. All references
should be checked against the original literature. Never cite a reference that you have not read
yourself. Double check all journal year, volume, issue, and inclusive page numbers to insure the
accuracy of your citation.

You might also like