Preference Uncertainty, Preference Learning, and Paired Comparison Experiments

David C. Kingsley and Thomas C. Brown

Abstract

Results from paired comparison experiments suggest that as respondents progress through a sequence of binary choices they become more consistent, apparently fine-tuning their preferences. Consistency may be indicated by the variance of the estimated valuation distribution measured by the error term in the random utility model. A significant reduction in the variance is shown to be consistent with a model of preference uncertainty allowing for preference learning. Respondents become more adept at discriminating among items as they gain experience considering and comparing them, suggesting that methods allowing for such experience may obtain more well founded values. (JEL C25, D83)

I. Introduction

A fundamental assumption of neoclassical microeconomic theory is that preferences are transitive. This intuitive assumption implies that among a series of binary choices preferences cannot cycle. For example, if a consumer prefers A to B and B to C then it follows that he or she also prefers A to C. Paired comparison experiments involve multiple binary choices between items in the choice set, allowing researchers to test the transitivity axiom.

Past research shows that for all but very small choice sets, respondents' paired choices are rarely fully transitive, but that as respondents progress through a random sequence of paired choices they become more consistent, apparently fine-tuning their preferences (Brown et al. 2008). This fine-tuning implies imprecision of preference, or in other words, preference uncertainty. Preference uncertainty was described by Thurstone (1927) as reflecting an underlying valuation distribution from which an individual randomly draws a value at a given instant. Allowing for preference uncertainty, the respondent becomes a potential source of error within choice models. Respondent error and the existence of preference uncertainty is an increasingly important topic being investigated within choice experiments and valuation studies. Indeed, an emergent theme within nonmarket valuation is to allow respondents to express levels of uncertainty (Alberini, Boyle, and Welsh 2003; Champ et al. 1997; Evans, Flores, and Boyle 2003; Li and Mattsson 1995; Welsh and Poe 1998).

Preference uncertainty implies that the design of the experiment or valuation survey may affect respondent choice. Researchers have examined the effect of experimental design using the error variance of the random utility model as a measure of preference uncertainty. Increasing the complexity of the choice set was found to increase the variance of the error term in a heteroskedastic logit model (DeShazo and Fermo 2002). Deshazo and Fermo hypothesize that the variance of choice models and choice consistency are inversely related. Similarly, it has been shown that the difficulty of the choice, referred to as task demand, has a nonlinear effect on the error term. Both very easy and very difficult choices were more random (Swait and Adamowicz 1996). These papers suggest preference uncertainty but do not address preference learning.

This step was taken in two studies that looked at the effect that repeated choices had on the mean and variance of elicited preferences (Holmes and Boyle 2005; Savage and Waldman 2008). A reduction in the error variance through the choice experiment implies preference learning, while an increase implies respondent fatigue or boredom. Results from Savage and Waldman (2008) were mixed; in their web sample fatigue was supported, but in their mail sample the error was constant. Holmes and Boyle (2005) found that error variance did decline over a sequence of choices, implying that respondents were better able to discriminate between choices made later in the experiment.

In this paper we show that the increasing choice consistency observed by Brown et al. (2008) is accompanied by a significant reduction in the error variance of a random utility model fit to the paired comparison data. We interpret this finding as preference learning. This result implies that the data become less noisy over choice occasions and indicates that respondents are better able to discriminate between items in later choices. Further, we find, as expected, that greater utility difference between items significantly reduces the probability of an inconsistent choice, and that inconsistent choices are likely to be switched when retested at the end of the experiment.

Taken together, these findings suggest that even hypothetical market experience provided through simple paired comparisons may affect respondents' choices and that nonmarket valuation techniques that rely on only one or a few responses may not be obtaining well-founded values. This finding is in line with the recent report by Bateman et al. (2008) that respondents to a dichotomous-choice contingent valuation survey require repetition and experience with the choice task in order to express preferences consistent with economic theory. As described in more detail in the Discussion section (Section V), our finding is also not inconsistent with the discovered preference hypothesis (Plott 1996), which maintains that stable underlying preferences are uncovered through experience with a choice task.

II. Preference Uncertainty and Learning

Random utility models provide a general framework within which researchers investigate individual choice behavior (McFadden 2001). Consistent with economic theory, these models assume that individuals always choose the alternative yielding the highest level of utility (Marschak 1959). Utility is described as a random variable in order to reflect the researcher's observational deficiencies, not individuals' uncertainty about their own preferences (Ben-Akiva and Lerman 1985).

The model that Marschak (1959) proposed was an interpretation of what was probably the first stochastic model of choice, introduced by L. L. Thurstone in 1927 under the name of “the law of comparative judgment.'' Unlike the modern random utility model, in Thurstone's model, utility is represented by a distribution about a fixed point of central tendency (Thurstone 1927). This representation of utility has important implications concerning the source of error in choice models and represents the fundamental difference between these models (Brown and Peterson 2009; Flores 2003). Thurstone's model is now referred to as a constant utility model (Ben-Akiva and Lerman 1985). The constant utility model allows individuals to sample their utility from a distribution; choices are made based on the realization of utility on a particular choice occasion. This uncertainty may cause observed preferences to appear inconsistent (i.e., violate transitivity).

The law of comparative judgment was developed to explain common results from psychometric choice experiments involving binary choices (Bock and Jones 1968; Brown and Peterson 2003; Torgerson 1958). For Thurstone, a choice between two alternatives involved draws from two underlying preference or judgment distributions (McFadden 2001). Subjects might, for example, be presented with numerous pairs of objects and asked, for each pair, to say which object is heavier. The main finding, which dates back at least to Fechner's work (1860), was, not surprisingly, that the closer the items were in weight the more common incorrect selections became.

Allowing for researcher error is common practice in economic models. Although allowing for the existence of uncertain preferences and sources of error beyond the researcher is less common, it has not been ignored. For example, Bockstael and Strand (1987) examined the effect the source of error has on the estimation of economic values in a framework they called “random preferences.'' More recent research suggests that each respondent has an implicit valuation distribution (Wang 1997). For Wang, respondents answer dichotomous choice questions as if their values reflect distributions rather than fixed points. Similarly, Li and Mattsson (1995) assume that respondents have incomplete knowledge of their preferences and thus can give the wrong answer to a dichotomous choice question. They find that respondents are a significant source of error and so exacerbate the standard deviation of the estimated valuation distribution.

This paper assumes that both sources of error, researcher and respondent, are present in individual choice. The term “preference uncertainty'' reflects respondent error, which translates to draws from an underlying valuation distribution unknown to both the respondent and the researcher. These random draws may contribute to inconsistency and increase the noise measured in the data. If respondent uncertainty could be reduced, perhaps through market experience or experimental design, choice consistency would increase and the data would become less noisy. This process will be referred to as preference learning and will be evident through a reduction in the standard deviation of the estimated valuation distribution measured by the error variance in the random utility model.

Dichotomous Choice Contingent Valuation

In a standard dichotomous choice contingent valuation study, respondents are asked to respond “yes'' or “no'' to a question such as, “Would you be willing to pay ti dollars to obtain environmental improvement kŒ'' The individual's valuation function is defined as follows:

Embedded Image[1]

where uik is individual i's unobserved utility of item k, the deterministic component of value is represented by ak, and eik represents the stochastic component. Note that we assume a homogeneous set of individuals with respect to ak. It is common to express ak as linear in parameters, xi'b, where xi is a set of variables describing the characteristics of either the individual or the item. The respondent is assumed to choose “yes'' whenever uik $ ti. Therefore,

Embedded Image[2]

where P indicates probability. Allowing the stochastic error term, eik, to be normally distributed with mean zero and constant variance, se2, we have the following expressions:

Embedded Image[3]

and

Embedded Image[4]

where W is the standard normal cumulative distribution. Then se represents the standard deviation of the estimated valuation distribution, which has mean ak.

It is worth noting that within dichotomous choice contingent valuation settings, the assumption of a symmetric valuation distribution means that the scale of the model has little consequence, such that preference uncertainty leads to no bias in the estimated mean or median. The importance of recognizing preference uncertainty and preference learning becomes evident within choice experiments where respondents make several choices between items. Common examples of such choice experiments include attribute-based methods and paired comparison experiments.

Paired Comparison Experiments

Consider the choice between two items, labeled r and c. The utilities of the items to individual i are distributed as follows:

Embedded Image[5]

and

Embedded Image[6]

Under the assumption that eik is a mean zero random variable distributed i.i.d. normal, the choice between items r and c can be written probabilistically, where Prc is the probability that item r is chosen over item c:

Embedded Image[7]

or

Embedded Image[8]

and

Embedded Image[9]

where 2se is the standard deviation of eic — eir.

Consider the density functions of items r and c depicted in Figures 1 and 2, which represent the underlying valuation distribution assumed to exist for each individual. In expectation item r is preferred to item c, since ar. ac. But for a given choice, individuals act on their instantaneous values, not their expected values, which are unknown. Figure 1 shows the instantaneous value of item r, uir, above (to the right of) the instantaneous value of item c, uic. Thus, Figure 1 depicts a consistent choice because the choice based on these instantaneous values is consistent with the individual's underlying preferences represented by the expected values. As depicted in Figure 2, these two density functions allow for an inconsistent choice, wherein uic. uir despite r being preferred in expectation.

The expression for Prc provides two intuitive results. First, for a given standard deviation, se, the greater the utility difference, arc = ar 2 ac, the more likely the choice will be consistent (item r being chosen over item c), and conversely, the less likely an inconsistent choice becomes

Embedded Image[10]

Second, for a given utility difference, arc, the narrower the distribution (the smaller is se) the more likely a consistent choice becomes, and the wider the distribution the more likely an inconsistent choice becomes:

Embedded Image[11]

In psychometric experiments, inconsistent choices are easily identified because the expected value of each item is objective (e.g., the weight of an object). However, in economic valuation studies the expected value must be estimated, and inconsistency, particularly within an individual, is not easily identified. Peterson and Brown (1998) developed a simple technique (discussed in the next section) used with paired comparison experiments that identifies a respondent's likely set of inconsistent choices.

III. Paired Comparison Methodology

The paired comparison method has successfully been used in nonmarket valuation studies (Champ and Loomis 1998; Kingsley 2006; Loomis et al. 1998; Peterson and Brown 1998).1 In this paper we reanalyze paired comparison data collected by Peterson and Brown (1998). In the Peterson and Brown experiment all items were economic gains. Respondents were instructed to choose the item in each pair they would prefer if they could have either at no cost. The paired choices were drawn from a set of four private goods and six locally relevant public goods (Table 1) along with 11 monetary amounts.2 Items were not paired with themselves, and dollar amounts were not compared (it was assumed that larger dollar amounts were preferred). Each respondent made 155 choices, 45 between items and 110 between an item and a dollar amount. For presentation, pairs were randomized across respondent and choice occasion. The pairs were presented on a personal computer, and the time respondents took to enter each choice was recorded. Three hundred and thirty students from Colorado State University participated in the study. Four were dropped because of missing data, leaving a total of 326 respondents, providing 50,530 individual observations. In addition, the experiment retested 10 consistent and all inconsistent choices within an individual after the initial choices were made. The respondents had not been informed that some choices would be repeated, and there was no break in the presentation of pairs to indicate that a new portion of the experiment had begun.

Table 1.

Items Included by Peterson and Brown (1998)

Given a set of t items, the paired comparison method presents them independently in pairs as (t/2)(t - 1) discrete binary choices. These choices yield a preference score for each item, which is the number of times the respondent prefers that item to other items in the set. A respondent's vector of preference scores describes the individual's preference order among the items in the choice set, with larger integers indicating more-preferred items. In the case of a 21-item choice set, an individual preference score vector with no circular triads contains all 21 integers from 0 through 20. Circular triads (i.e., choices that imply A. B. C. A) cause some integers to appear more than once in the preference score vector, while others disappear.

For a given respondent, a pair's preference score difference (PSD) is simply the absolute value of the difference between the preference scores of the two items of the pair. This integer, which can range from 0 to 20 for a 21-item choice set, indicates on an ordinal scale the difference in value assigned to the two items.

The number of circular triads in each individual's set of binary choices can be calculated directly from the preference scores. The number of items in the set determines the maximum possible number of circular triads. The individual respondent's coefficient of consistency is calculated by subtracting the observed number of circular triads from the maximum number possible and dividing by the maximum.3 The coefficient varies from one, indicating no circular triads in a person's choices, to zero, indicating the maximum possible number of circular triads.

When a circular triad occurs, it is not unambiguous which choice is the cause of the circularity. This is easily seen by considering a choice set of three items whose three paired comparisons produce the following circular triad: A. B. C. A. Reversing any one of the three binary choices removes the circularity of preference; selection of one to label as “inconsistent'' is arbitrary. However, with more items in the choice set, selection of inconsistent choices, though still imperfect, can be quite accurate. For each respondent, we selected as inconsistent any choice that was contrary to the order of the items in the respondent's preference score vector, with the condition that the order of items with identical preference scores was necessarily arbitrary. Simulations show that the accuracy of this procedure in correctly identifying inconsistent choices increases rapidly as the PSD increases. In simulations with a set of 21 items and assuming normal dispersion distributions, the accuracy of the procedure rises quickly from 50% at a PSD of 0 to nearly 100% at a PSD of 5.4

IV. Results And Analysis

In this section, we first report on the likelihood of an inconsistent choice and the likelihood of a preference reversal. This analysis provides support for the notion of preference learning. We then take a closer look at preference uncertainty and preference learning, fitting a heteroskedastic probit model to the paired comparison data.

Probability of an Inconsistent Choice

The proportion of choices identified as inconsistent, using the decision rule described above, falls quickly and then levels off as respondents progress through the experiment (Figure 3). Nearly 14% of respondents' first choices are identified as inconsistent, but by the twentieth choice the proportion has dropped to around 5% to 6%. Because the order in which the pairs were presented was random for each respondent, this result is unrelated to presentation order and thus seems driven by respondents' experience comparing the items.

Figure 3.

Proportion of Choices Identified as Inconsistent

A probit model is estimated to predict the probability of a choice being identified as inconsistent. The dependent variable, yij, equals 1 if the choice is identified as inconsistent and 0 otherwise, where i denotes the individual and j denotes the choice occasion. Exogenous variables include Choice occasion, ln(Choice occasion), PSD, a public good dummy, and the amount of time in seconds required to make the choice. Demographic variables included are Gender, Age, and Education.

The variable Choice occasion represents the order in which a given respondent encountered the pairs of items and made a choice; given Figure 3, the probability of an inconsistent choice is expected to decrease over choice occasion. ln(Choice occasion) captures the decreasing and leveling off relationship between choice inconsistency and choice occasion depicted in Figure 3. PSD is used as an approximate measure of the utility difference, the expectation being that respondents are less likely to commit an inconsistent choice the greater is this difference.5 The public good dummy represents any choice involving a public good (the public good may be paired with another public good, a private good, or a dollar amount). Respondents are assumed to face greater uncertainty when the choice involves a public good, as opposed to choices involving only private goods or a private good and a monetary amount, because they lack experience making public good choices. Greater uncertainty is expected to lead to greater choice inconsistency. Similarly, the amount of time required to make the choice reflects the difficulty of the choice and is expected to be positively related to choice inconsistency. Therefore,

Embedded Image[12]

and

Embedded Image[13]

All 50,530 choices are pooled over individuals i and choice occasions j, so the likelihood function becomes

Embedded Image[14]

Results, shown in Table 2, support the intuition of the model. Note that only marginal effects are reported and that Columns (1) and (2) present models that treat each observation as independent, while Columns (3) and (4) adjust the standard errors to account for the repeated choices made within individual. As expected, controlling for repeated responses per individual raises the standard errors; however, the marginal effects remain significant. A negative and significant PSD indicates, as predicted in the conceptual model, that inconsistency is less likely the greater the utility difference between the items. Additionally, choices including a public good are more likely to be inconsistent, perhaps suggesting that there is greater uncertainty and thus a wider valuation distribution for these goods. Choices that require more time to make are also more likely to be inconsistent, and older students tend to be more consistent.

Table 2.

Probability of an Inconsistent Choice

Importantly, Choice occasion (Columns and 3) is negative and significant, suggesting that an inconsistent choice becomes less likely as respondents progress through the experiment. When Choice occasion is modeled in log form (Columns and 4), Choice occasion is also significant, implying that the probability of an inconsistent choice decreases quickly and levels off. The Akaike information criterion (AIC) statistic reported in Table 2 is a measure of fit; as is shown, the log form provides a substantially lower AIC and therefore provides a better fit to the data. Figure 46 illustrates the two functional forms based on the data in Columns (3) and (4.)

Figure 4.

Specification Effects on the Probability of an Inconsistent Choice

Probability of a Preference Reversal

The experiment retested 10 randomly selected consistent choices and all inconsistent choices made by an individual after the initial 155 choices were made. The data show substantial differences in the reversal rates of originally consistent as opposed to originally inconsistent choices. Of The 3,260 consistent choices retested, 289, or 8.9%, were reversed, whereas 2,250 of 3,679 (61.2%) inconsistent choices were reversed (Table 3). This indicates that respondents were not only striving for consistency with their previous choices but also expressing their preferences.

Table 3.

Choice Switching

The data are further broken out in Table 3 into the types ofchoices made.A choice may be between two public goods, two private goods, a public and private good, a public good and a dollar amount, or a private good and a dollar amount. As is shown, there are only slight differences among the types of choices in the rate of preference reversal, indicating that type of choice did not play a large role in preference reversals.

A probit model was used to test for the significance of factors hypothesized to affect the probability of a preference reversal. As noted above, two distinct sets of data were retested, the set of originally inconsistent choices and a random selection of 10 originally consistent choices. For both sets of choices the dependent variable, yij, equals 1 if the choice was reversed and 0 otherwise. Six exogenous variables are included: PSD, the public good dummy, the choice occasion when the original choice was made, and the three demographic variables.

The results are presented in Table 4. Columns (1) and (2) investigate the probability of reversing a choice that was originally identified as inconsistent, whereas Columns (3) and (4) investigate the probability of reversing a choice that was originally identified as consistent. Columns (2) and (4) adjust the error terms to account for repeated observations within individual. As is shown, accounting for repeated observations has little impact on the marginal effects. Results suggest that the greater the PSD between the items the more (less) likely an originally inconsistent (consistent) choice is reversed. Choices involving a public good are more likely to be reversed in both the inconsistent and consistent subsets. Choice occasion has a significant and negative effect on the probability of reversing an originally inconsistent choice, showing that more recent inconsistent choices are less likely to be reversed. However, Choice occasion has no significant effect on reversing an originally consistent choice.

Table 4.

Probability of A Preference Reversal

Preference Learning

This paper defines preference learning as a significant reduction in the error variance of the random utility model, that is, a significant narrowing of the estimated valuation distribution.7 The valuation function for this analysis is as follows:

Embedded Image[15]

Again the index i denotes the individual and j denotes the choice occasion. Note that the data are set up in rows and columns; as such, the item index will be k = r, c for row or column. The row contains the 10 items and the column contains the 10 items along with the 11 monetary amounts. As before, the model assumes that all respondents are identical (that they have the same valuation on each ak). The error term, eijk, includes an alternative specific error constant and is normally distributed N (0,se2 ). The probability contribution to the likelihood function is denoted Prc (Pcr) for the probability that the row (column) item is chosen over the column (row) item.

First consider the choice between an item and a dollar amount. Each bid amount appears only in the column and is denoted by tijc. The probability that the item is preferred to the dollar amount is

Embedded Image[16]

and thus

Embedded Image[17]

Next, consider the choice between two items

Embedded Image[18]

and thus,

Embedded Image[19]

where (se2 zse2 )1/2 is the standard deviation of ec{er. The dependent variable, yijk, equals 1 if the column item is chosen and 0 otherwise. The likelihood function is written as follows:

Embedded Image[20]

In order to test for preference learning, a heteroskedastic probit model is introduced that allows the standard deviation of the errortermto adjust over Choice occasion as follows:

Embedded Image[21]

Two functional forms are considered: linear sekj ~lkzb(j) and nonlinear sekj ~ lkzb(1=j). Furthermore, this reduced form specification sets the standard deviation to be a function of the three available demographic variables. It has been shown that one can identify the error variance ofchoice models by exploiting the variation in monetary thresholds (Cameron 1988; Cameron and James 1987). In particular, using the variation across dollar amounts in this paired comparison experiment, we are able to identify the standard deviation within the model. Furthermore, using choice occasion, we are able to identify changes to this error structure over the course of the experiment.

This specification, using the standard deviation as a function ofchoice occasion, made it impossible to cluster the error terms by individuals, as the full model was unidentified. However, previous models show little impact of clustering and we do not believe this to be a serious problem.

Interpretation of these parameters is as follows. Researcher error as well as any error generated by the respondent unrelated to choice occasion will be picked up by lk. This term can be interpreted as an alternative specific error constant that is a measure of the error introduced into the model by each item. in both forms a significant b represents a significant change in the scale over choice occasion (i.e., a change in respondent error). A reduction in the scale is interpreted as preference learning, and an increase is interpreted as fatigue or boredom. The hypothesis to be tested is H0 : b = 0, implying that choice occasion has no effect on the scale of the model.

Results for the parameterized error structure are shown in Table 5. The important result is the significance of b in both the linear (Column 1) and nonlinear (Column 2) forms. Both models suggest that the scale of the choice model decreases as respondents progress through a sequence of randomly ordered choices. This implies that the data become less noisy as the respondent continues through the experiment. In order to believe that this reduction stems from researcher error, it would need to be the case that some unobservable characteristics of the choices became less significant to the respondent as the experiment progressed.

Table 5.

Heteroskedastic Probit Model

As shown in Figure 3, the proportion of choices identified as inconsistent drops quickly and levels off, suggesting that the nonlinear form is the more appropriate representation of the preference learning described in these data. This is also supported by the lower AIC statistic reported in Table 5 for the nonlinear model. This evidence supports the inverse relationship between choice consistency and the scale of random utility choice models (DeShazo and Fermo 2002; Savage and Waldman 2008).

The above analysis implies that the error change parameter, b, is equivalent across items. In order to relax this assumption we ran a model allowing each item to have a b. Table 6 presents the b coefficients; to simplify the presentation the 10 l rows are notshown. The b terms for 6of the 10items remain significant; for the 4 others the b is found to be insignificant. It is not entirely surprising that across a variety of items some may benefit from learning while others may not. Why some items are susceptible to learning while others are not is unknown and may be experiment specific. However, it is important to note that the results are not driven to be a single item or even a subset of items (i.e., only the public goods or only the private goods) and that the learning indicated in Table 5 is supported across a variety of items.

Table 6.

Heteroskedastic Probit Model Individual Beta Terms

V. Conclusion And Discussion

This paper combines common notions of preference uncertainty and preference learning found in the literature. Preference uncertainty implies that an individual's choice on a data become less noisy over choice occasion and indicates that the respondents are better able to discriminate between items in later choices. This result supports the ability of individuals to learn their preferences as they progress through a choice experiment in a way consistent with the discovered preference hypothesis (Plott 1996). This hypothesis proposes that stable underlying preferences are “discovered,'' or “learned,'' through experience with a choice task that provides relevant feedback and incentives (see also Braga and Starmer 2005). Respondents in our paired comparison sessions can be said to be learning their underlying preferences in the sense that they refined or fine-tuned them. Further, we observed no indication that the underlying preferences were not stable, since, as reported by Brown et al. (2008), there was no change in mean preference scores over choice occasion.8 It is notable that this refinement occurred even though the paired comparison task provides no feedback or incentives that would encourage respondents to focus more carefully or otherwise alter their behavior. This suggests that mere experience in choosing between pairs of a set of items is sufficient for preference learning to occur.9

The result that respondents are able to learn their preferences as they progress through a choice experiment raises important questions for the economic valuation literature. Notably, one must consider the efficiency, if not the appropriateness, of single-shot surveys that assume well-founded preferences exist a priori. Furthermore, the question of whether elicited preferences may be altered through this process is of central concern to the literature and is deserving of further investigation.

Footnotes

  • The authors are, respectively, assistant professor, Department of Economics and Management, Westfield State College, Westfield, Massachusetts; and research economist, Rocky Mountain Research Station, Fort Collins, Colorado. This paper benefited from discussions with Patricia Champ, Nick Flores, Jason Shogren, and Donald Waldman, as well as conference and seminar participants at the AERE sessions at the ASSA 2006 annual meetings, University of Colorado at Boulder, the Environmental Protection Agency, and Stephen F. Austin State University. All errors remain our own.

  • 1 This paper is primarily concerned with what the measure of the error term of the model of individual choice reveals about the decision making process, and not with the estimated values of the items assessed. All data related to the estimated means are available upon request.

  • 2 Dollar amounts were 1, 25, 50, 75, 100, 200, 300, 400, 500, 600, and 700.

  • 3 The maximum possible number of circular triads, m, is (t/24)(t2 - 1) when t is an odd number, and (t/24)(t2 -4) when t is even, where t is the number of items in the set. Letting ai equal the preference score of the ith item and b equal the average preference score, in other words, (t - 1)/2, the number of circular triads is (David 1988) c~(t=24) (t2{1){1=2 (ai{b)2 : The coefficient of consistency (Kendall and Smith 1940) is then defined as 1-c / m.

  • 4 A thorough explanation of the procedure for specifying inconsistent choices is found in Chapter 4 of Brown and Peterson's book (2009), available at www.treesearch.fs.fed.us/pubs/31863.

  • 5 PSD should clearly be related to the probability of an inconsistent choice. It is included in the model to verify this expectation. Its inclusion raises the question of endogeneity. With this in mind, we ran several other probit models that systematically excluded individual independent variables in order to observe whether results were sensitive to inclusion of PSD or any of the other independent variables. Results remained consistent and the coefficient on PSD varied very little.

  • 6 Figure 4 was created using the estimated coefficients from Table 2 Columns (3) and (4). All variables except dummy variables are set to their averages: PSD = 6.93, Age = 19.8, School = 13.2, Time = 3.2, Gender = 0 (Female), and Public = 0.

  • 7 As noted previously, the primary focus of this paper is the error process over choice occasion and the ability of respondents to discriminate between choices as they progress through the paired comparison experiment. Therefore we limit our presentation here to the error process.

  • 8 An alternative to the discovered preference hypothesis is the constructed preference hypothesis (Gregory, Lichtenstein, and Slovic 1993), which posits that preferences are quite labile and context dependent and that they are constructed during and possibly influenced by the choice task at hand. We presented the paired choices in only one context and, thus, did not test the extent to which respondents' choices were subject to contextual cues. We do note, however, that the paired comparison task is relatively free of contextual cues, in that the choices are binary and randomly ordered and the respondents do not interact. Further, in our implementation of the task, payment—and all the concerns that arise with payment— were not an issue. The method of paired comparisons thus offers a credible way to assess underlying preferences without being overly concerned about the influence of unwanted contextual cues.

  • 9 This finding is also not inconsistent with that of List (2003, 2004), although the kind of field market experience he examined is quite different from the experience gained during a paired comparison exercise.

References