Tuesday, 9 July 2013

Paradigms of Approaches to Scales Development

1.                  Definition of Paradigm
1.                   example, pattern; especially : an outstandingly clear or typical example or archetype
2.                  a philosophical and theoretical framework of a scientific school or discipline within which theories, laws, and generalizations and the experiments performed in support of them are formulated; broadly : a philosophical or theoretical framework of any kind
3.                  A typical example or pattern of something; a model.
4.                  A worldview underlying the theories and methodology of a particular scientific subject.
5.                  Definition of approach
1.                  A way of dealing with a situation or problem
2.                  Start to deal with (a situation or problem) in a certain way
3.                  Measurement Scales
Scaling defined:
Procedures for assigning numbers (or other symbols) to properties of an object in order to impart some numerical characteristics to the properties in question.
4.                  Scaling Approaches:
1.                  Unidimensional:
Measures only one attribute of a concept, respondent, or object.
2.                  Multidimensional:
Measures several dimensions of a concept, respondent, or object.
3.                  Types of Scales:
1.                  Noncomparative Scale:
Scales in which judgment is made without reference to another object, concept, or person.
2.                  Comparative Scale:
Scales in which one object, concept, or person is compared with another on a scale.
3.                  Churchill’s paradigms
Here is Figure from "A Paradigm for Developing Better Measures for Marketing Constructs". Gilbert A. Churchill, Jr., Journal of Marketing Research, 16:1. (Feb., 1979)

(Churchill, 1979)(A Paradigm for Developing Better Measures of marketing Constructs)
1.                  Introduction
1.                  Measurements are “rules for assigning numbers to objects to represent qualities of attributes”.
2.                  What is measured? ATTRIBUTES of objects. NOT objects themselves.
3.                  What is the goal? To have measures that are RELIABLE and VALID
4.                  Construct
1.                  Construct, e.g. customer satisfaction
2.                  True level of satisfaction (True score) denoted  Xt
3.                  Observed score X0, rarely similar to Xt due to differences in stable characteristics, transient  personal factors, situational factors etc.
4.                  Validity and Reliability
X0 = Xt + Xs + Xr, where
1.                  Xs – systematic source of error
2.                  Xr – random source of error
Validity: X0 = Xt
Perfect reliability: Xr = 0
1.                  Validity => Reliability
2.                  Reliability is necessary but not sufficient for Validity
3.                  Validity and Reliability (2)
1.                  Objective: find X0 that approximate Xt
2.                  Measures are inferences, their “goodness” is supported by the evidence, that is based on reliability or validity index
3.                  Reliability forms: split-half, test-retest etc.
4.                  Validity forms: face, content, predictive, concurrent, pragmatic, construct, convergent, discriminant.
5.                  Specify domain of the construct
1.                  Exactly defining what is included in the definition and what is excluded
2.                  Consulting the literature
3.                  Widely varying definitions should be avoided
4.                  Example: to measure customer satisfaction
1.                  Measure both expectations at the time of purchase and reactions at some time after the purchase
2.                  Expectations: cost, durability, quality, operating performance, aesthetic features, sales assistance, advertising, availability of competitor’s alternatives,
3.                  Generate sample of items
1.                  Literature searches
2.                  Experience surveys
3.                  Insight-stimulating examples
4.                  Critical incidents and focus groups
5.                  Purify the measure
1.                  Domain sampling model: purpose of any particular measurement is to estimate the score that would be obtained if all the items in the domain were used
2.                  In practice use of SAMPLE of items
3.                  Measurement error due to inadequate sampling
4.                  Correlation matrix of the items in the domain
1.                  Average correlation in the matrix
2.                  Dispersion of the correlation about the average
1.                  Assumption: all items, “if they belong to the domain of the concept, have an equal amount of common core”
2.                  Coefficient Alpha
1.                  Measure of internal consistency of a set of items
2.                  Low coefficient alpha indicates that the sample of items badly describes the construct which motivated the measure
3.                  Procedure by low alpha: some items should be eliminated.
1.                  Calculate correlation of each item with total score
2.                  Plot the correlations by decreasing order of magnitude
3.                  Items with correlations near zero should be eliminated
4.                  Items of substantial drop in the item-to-total correlations also deleted
1.                  Mistake to do split-half reliability
2.                  Purify the measure (2)
1.                  Desirable outcome: high coefficient alpha, dimensions agree with the conceptualized. Then,  additional testing with a new sample of data.
2.                  Second outcome: Factor analysis suggests the overlapping dimensions. Items with pure loadings on the new factor are retained, new alpha calculated.
3.                  Non-desirable outcome: alpha coefficient is low and restructuring of items forming each dimension is unproductive. Loop back to 1. and 2.
4.                  Assess Reliability with new Data
1.                  Source of error within a test or measure is the sampling of items.
2.                  Coefficient alpha is the basic statistic for determining the reliability of a measure based on internal consistency, but it does not estimate errors external to the instrument.
3.                  Collect additional data to rule our the chance possibility of previous findings
4.                  Do not use test-retest reliability
5.                  Assess Construct Validity
1.                  Face or content valid measure has an appropriate sample
2.                  To establish construct validity
1.                  Determine the extent to which the measure correlates with other measures designed to measure the same thing
2.                  Determine whether the measure behaves as expected
3.                  Correlations with Other Measures
1.                  Any construct or trait should be measurable by at least two different methods
2.                  Convergent validity – extent to which it correlates highly with other methods designed to measure the same construct
3.                  Discriminant validity – the extent to which a measure a novel
4.                  Multitrait-multimethod matrix: methods and  traits generating it should be as independent as possible
5.                  Multitrait-multimethod matrix
6.                  Does the measure behave as expected?
1.                  Internal consistency is insufficient condition for construct validity
2.                  Assess whether scale correctly predicts criterion measure (criterion validity)
1.                  The constructs job satisfaction (A) and likelihood of quitting the job (B) are related.
2.                  The scale X provides a measure of A.
3.                  Y provides a measure of B.
4.                  X and Y correlate positively.
1.                  Establish the validity by relating the measure to a number of other constructs and  not only one
2.                  Developing Norms
1.                  Assessing the position of the individual on the characteristic is to compare the person’s score with the scores achieved by other people
2.                  Norm quality depends on both the number of cases on which the average is based and their representativeness
3.                  Anderson (1977) and Garbing’s (1988) paradigms
The purpose of measurement in theory testing and development research is to provide an empirical estimate of each theoretical construct of interest. Because of the limitations inherent in single-item measures (cf. Churchill 1979), respondents usually are administered two or more measures, often referred to as a scale, that are intended to be altemative indicators of the same underlying construct. A composite score defined by the respondent's scores on these measures, generally calculated as an unweighted sum, provides an estimate of the corresponding construct. Our central thesis is that the computation of this composite score is meaningful only if each of the measures is acceptably unidimensional. Unidimensionality refers to the existence of a single trait or construct underlying a set of measures (Hattie 1985; McDonald 1981).' The importance of unidimensionality has been stated succinctly by Hattie (1985 p. 49): "That a set of items forming an instrument all measure just one thing in common is a most critical and basic assumption of measurement theory."
Because the meaning of a measure intended by the researcher may not be the same as the meaning imputed to it by the respondents, the scale development process must include an assessment of whether the multiple measures that defme a scale can be acceptably regarded as altemative indicators of the same construct. Building on the earlier work of Churchill (1979) and Peter (1979, 1981), we outline an updated paradigm for scale development that incorporates confirmatory factor analysis (cf. Bentler 1985; Joreskog and Sorbom 1984) for the evaluation of unidimensionality. The key aspect of this updated paradigm is that confimiatory factor analysis affords a stricter interpretation of unidimensionality than can be provided by more traditional methods such as coefficient alpha, item-total correlations, and exploratory factor analysis and thus generally will provide different conclusions about the acceptability of a scale.
Contributing to the tradition of articles by Churchill (1979) and Peter (1979,1981). We outline an updated Paradigm for scale development that incorporates a more recent methodological development: confirmatory factor analysis. In doing so, we attempt to provide a better understanding of the concept of unidimensional measurement and the ways in which it can be assessed and, in particular, to demonstrate that an explicit evaluation of unidimensionaiity is accomplished with a confirmatory factor analysis of the individual measures as specified by a multiple-indicator measurement model. Coefficient alpha is important in the assessment of reliability, but it does not assess dimensionality. Though item-total correlations and exploratory factor analysis can provide useful preliminary analyses, particularly in the absence of sufficiently detailed theory, they do not directly assess unidimensionality. There as on is that a confirmatory factor analysis makes possible an assessment of the intemal consistency and extemal consistency criteria of unidimensionality implied by the multiple indicator measurement model.
Following the paradigm of scale development outlined here, after the unidimensionality of a set of scales has been acceptably established, one would assess its reliability. Even a perfectly unidimensional scale will not be useful in practice if the resultant scale score has unacceptably low reliability. Because most measures in marketing are administered at a single point in time, coefficient alpha or some other coefficient of equivalence reliability would probably be used for this assessment.
The goal of most research projects is not just to develop unidimensional and reliable measurement scales, but to build and test theory. Essential to this undertaking is the assessment of construct validity. A construct achieves its meaning in two ways (Anderson1987; CronbachandMeehl1955): (1) through observed indicators for which it is posited to be causally antecedent (and through observed measures for which it is not) and (2) through the set of relationships of the construct with other constructs as specified by some theory (the nomoiogical network). Unidimensionality, then, is necessary but not sufficient for construct validity. Not only should all the indicators that define a scale provide estimates of exactly one factor, but the meaning of the underlying factors should correspond to the construct of interest.
The nomological network can be explored with in the
Context of the full structural equation model. One means for accomplishing this is the approach developed byAnderson and Gerbing(1988) that allows an assessment of nomological validity that is asymptotically independent of the assessment of the measurement model. It is called a "two-step" approach because the measurement model first is developed and evaluated separately from the full structural equation model that simultaneously models measurement and structural relations. The measurement model in conjunction with the structural model makes possible a comprehensive confirmatory assessment of construct validity (Bentler1978). Hence, the assessment of unidimensionality provided by a confirmatory factor analysis represents but a first step in the establishment of meaning for the estimated factors.
(Gerbing & Anderson, 1988)
4.                  Loewenthal (1996)  Approach
1.                  Features of Good Psychological Measures
1.                  A statement of what the scale measure;
2.                  Justification for the scale___ its uses and advantages over existing measures
3.                  A description of how the pool of items was drawn up
4.                  A description of the sample used for testing
5.                  An indication of the population  (kind of people) for whom the measure would be appropriate
6.                  Descriptive statistics (norms) means, standard deviations, ranges, different sub scales
7.                  Reliability statistics
8.                  Validity statistics
9.                  The scale itself (introduction, items or examples of items)
10.              Writing
1.                  Defining what you want to measure
The first and very important step is to work out and then write down exactly what you want to measure
2.                  Collecting items
3.                  Producing the preliminary questions or test
4.                  Testing
1.                  Deciding on a sample and reducing sample bias
2.                  Reducing methods
3.                  Testing
4.                  Data and preliminary analysis
1.                  Coding, scoring and data entry
2.                  Selecting reliable items
3.                  Descriptive statistics (norms) for final scale
4.                  Steps for data entry and reliability
5.                  Factor and principal factor analysis
6.                  The final scale and its validation
1.                  Descriptive statistics (norms)
2.                  Validity
3.                  Presenting the scale
9. Eclectic  Approach          
1. Definition
Selecting or choosing from various systems, methodologies, etc.; not following any one system.
Made up of elements selected from various sources: an eclectic philosophy.
2. Eclecticism is a conceptual approach that does not hold rigidly to a single paradigm or set of assumptions, but instead draws upon multiple theories, styles, or ideas to gain complementary insights into a subject, or applies different theories in particular cases. It can sometimes seem inelegant or lacking in simplicity, and eclectics are sometimes criticized for lack of consistency in their thinking. It is, however, common in many fields of study.
The theory of internalization itself is based on the transaction cost theory. This theory says that transactions are made within an institution if the transaction costs on the free market are higher than the internal costs. This process is called internalization.
4.                  Theory
The idea behind the Eclectic Paradigm is to merge several isolated theories of international economics in one approach.
5.                  Mixed-Method Design for Scale Development
Scale development guidelines by Churchill (1979) and DeVellis (2003) clearly  involve two phases, the exploratory (quantitative and qualitative data gathering)  structure, followed by a confirmatory phase (quantitative) involving purification and  confirmation of the scale. The mixed-method approach is a suitable research design to  be applied in scale development research especially when the objective is to discover  in-depth knowledge of a complex phenomenon in a social context (Creswell, 2008;  Bryman, 2008) and test hypotheses (Creswell, 2008).  In exploratory MMR, there are two distinct types of qualitative methods being employed.
i. The first are studies that review the literature togather dimensions or indicators  of service quality and present it to focus groups to revise the existing  instruments.
ii. The second is the use of qualitative data gathering through focus group  interviews to discover new indicators or dimensions of service quality. 
1.                  References
1.                  Churchill, G. A., Jr., (1979). A paradigm for developing better measures of marketing constructs. Journal of Marketing Research, 16(February), 64-73.
2.                  Gerbing, D. W., & Anderson, J. C. (1988). An updated paradigm for scale development incorporating unidimensionality and its assessment. Journal of Marketing Research, 25(2), 186-192.
3.                  Loewenthal, K. M. (2001).An introduction of psychological tests and scales._2nd ed. psychology press Ltd. 27 Church Road, Hove, East Sussex, BN3 2FA, ISBN 1-84169-106-2 (hbk)
4.                  http://www.rasch.org/rmt/rmt222c.htm April 16, 2013
5.                  http://en.wikipedia.org/wiki/Eclectic_paradigm April 16, 2013
6.                  www.psypress.co.uk April 16, 2013
7.                  http://oxforddictionaries.com/definition/english/paradigm April 23, 2013
8.                  http://www.thefreedictionary.com/approachApril 23, 2013
9.                  http://www.thefreedictionary.com/eclectic April 23, 2013 

10.              http://en.wikipedia.org/wiki/Eclecticism April 23, 2013

No comments:

Post a Comment