Jump to content
 







Main menu
   


Navigation  



Main page
Contents
Current events
Random article
About Wikipedia
Contact us
Donate
 




Contribute  



Help
Learn to edit
Community portal
Recent changes
Upload file
 








Search  

































Create account

Log in
 









Create account
 Log in
 




Pages for logged out editors learn more  



Contributions
Talk
 



















Contents

   



(Top)
 


1 Related terms  





2 Theoretical background  





3 Distinction between revealed and stated preference studies  





4 History  





5 Relationship with conjoint analysis  





6 Designing a choice model  



6.1  Identifying the good or service to be valued  





6.2  Deciding on what attributes and levels fully describe the good or service  





6.3  Constructing an experimental design that is appropriate for those attributes and levels, either from a design catalogue, or via a software program  





6.4  Constructing the survey  





6.5  Administering the survey to a sample of respondents in any of a number of formats including paper and pen, but increasingly via web surveys  





6.6  Analysing the data using appropriate models, often beginning with the multinomial logistic regression model, given its attractive properties in terms of consistency with economic demand theory  







7 Strengths  





8 Weaknesses  





9 The mean-variance confound  





10 Versus traditional ratings-based conjoint methods  





11 Other types  



11.1  Ranking  





11.2  Bestworst scaling  







12 Uses  



12.1  Occupational choice model  







13 See also  





14 References  





15 External links  














Choice modelling






العربية
فارسی
 

Edit links
 









Article
Talk
 

















Read
Edit
View history
 








Tools
   


Actions  



Read
Edit
View history
 




General  



What links here
Related changes
Upload file
Special pages
Permanent link
Page information
Cite this page
Get shortened URL
Download QR code
Wikidata item
 




Print/export  



Download as PDF
Printable version
 




In other projects  



Wikimedia Commons
 
















Appearance
   

 






From Wikipedia, the free encyclopedia
 

(Redirected from Choice modeling)

Choice modelling attempts to model the decision process of an individual or segment via revealed preferences or stated preferences made in a particular context or contexts. Typically, it attempts to use discrete choices (A over B; B over A, B &C) in order to infer positions of the items (A, B and C) on some relevant latent scale (typically "utility" in economics and various related fields). Indeed many alternative models exist in econometrics, marketing, sociometrics and other fields, including utility maximization, optimization applied to consumer theory, and a plethora of other identification strategies which may be more or less accurate depending on the data, sample, hypothesis and the particular decision being modelled. In addition, choice modelling is regarded as the most suitable method for estimating consumers' willingness to pay for quality improvements in multiple dimensions.[1]

Related terms[edit]

There are a number of terms which are considered to be synonyms with the term choice modelling. Some are accurate (although typically discipline or continent specific) and some are used in industry applications, although considered inaccurate in academia (such as conjoint analysis).[2]

These include the following:

  1. Stated preference discrete choice modeling
  2. Discrete choice
  3. Choice experiment
  4. Stated preference studies
  5. Conjoint analysis
  6. Controlled experiments

Although disagreements in terminology persist, it is notable that the academic journal intended to provide a cross-disciplinary source of new and empirical research into the field is called the Journal of Choice Modelling.[3]

Theoretical background[edit]

The theory behind choice modelling was developed independently by economists and mathematical psychologists. The origins of choice modelling can be traced to Thurstone's research into food preferences in the 1920s and to random utility theory.[4] In economics, random utility theory was then developed by Daniel McFadden[5] and in mathematical psychology primarily by Duncan Luce and Anthony Marley.[6] In essence, choice modelling assumes that the utility (benefit, or value) that an individual derives from item A over item B is a function of the frequency that (s)he chooses item A over item B in repeated choices. Due to his use of the normal distribution Thurstone was unable to generalise this binary choice into a multinomial choice framework (which required the multinomial logistic regression rather than probit link function), hence why the method languished for over 30 years. However, in the 1960s through 1980s the method was axiomatised and applied in a variety of types of study.

Distinction between revealed and stated preference studies[edit]

Choice modelling is used in both revealed preference (RP) and stated preference (SP) studies. RP studies use the choices made already by individuals to estimate the value they ascribe to items - they "reveal their preferences - and hence values (utilities) – by their choices". SP studies use the choices made by individuals made under experimental conditions to estimate these values – they "state their preferences via their choices". McFadden successfully used revealed preferences (made in previous transport studies) to predict the demand for the Bay Area Rapid Transit (BART) before it was built. Luce and Marley had previously axiomatised random utility theory but had not used it in a real world application;[7] furthermore they spent many years testing the method in SP studies involving psychology students.

History[edit]

McFadden's work earned him the Nobel Memorial Prize in Economic Sciences[8] in 2000. However, much of the work in choice modelling had for almost 20 years been proceeding in the field of stated preferences.[9][10] Such work arose in various disciplines, originally transport and marketing, due to the need to predict demand for new products that were potentially expensive to produce. This work drew heavily on the fields of conjoint analysis and design of experiments, in order to:

  1. Present to consumers goods or services that were defined by particular features (attributes) that had levels, e.g. "price" with levels "$10, $20, $30"; "follow-up service" with levels "no warranty, 10 year warranty";
  2. Present configurations of these goods that minimised the number of choices needed in order to estimate the consumer's utility function (decision rule).

Specifically, the aim was to present the minimum number of pairs/triples etc of (for example) mobile/cell phones in order that the analyst might estimate the value the consumer derived (in monetary units) from every possible feature of a phone. In contrast to much of the work in conjoint analysis, discrete choices (A versus B; B versus A, B &C) were to be made, rather than ratings on category rating scales (Likert scales). David Hensher and Jordan Louviere are widely credited with the first stated preference choice models.[10] They remained pivotal figures, together with others including Joffre Swait and Moshe Ben-Akiva, and over the next three decades in the fields of transport and marketing helped develop and disseminate the methods.[11][12] However, many other figures, predominantly working in transport economics and marketing, contributed to theory and practice and helped disseminate the work widely.[13]

Relationship with conjoint analysis[edit]

Choice modelling from the outset suffered from a lack of standardisation of terminology and all the terms given above have been used to describe it. However, the largest disagreement has proved to be geographical: in the Americas, following industry practice there, the term "choice-based conjoint analysis" has come to dominate. This reflected a desire that choice modelling (1) reflect the attribute and level structure inherited from conjoint analysis, but (2) show that discrete choices, rather than numerical ratings, be used as the outcome measure elicited from consumers. Elsewhere in the world, the term discrete choice experiment has come to dominate in virtually all disciplines.[2] Louviere (marketing and transport) and colleagues in environmental and health economics came to disavow the American terminology, claiming that it was misleading and disguised a fundamental difference discrete choice experiments have from traditional conjoint methods: discrete choice experiments have a testable theory of human decision-making underpinning them (random utility theory), whilst conjoint methods are simply a way of decomposing the value of a good using statistical designs from numerical ratings that have no psychological theory to explain what the rating scale numbers mean.[2]

Designing a choice model[edit]

Designing a choice model or discrete choice experiment (DCE) generally follows the following steps:

  1. Identifying the good or service to be valued;
  2. Deciding on what attributes and levels fully describe the good or service;
  3. Constructing an Experimental design that is appropriate for those attributes and levels, either from a design catalogue,[14] or via a software program;[15]
  4. Constructing the survey, replacing the design codes (numbers) with the relevant attribute levels;
  5. Administering the survey to a sample of respondents in any of a number of formats including paper and pen, but increasingly via web surveys;
  6. Analysing the data using appropriate models, often beginning with the Multinomial logistic regression model, given its attractive properties in terms of consistency with economic demand theory.[5]

Identifying the good or service to be valued[edit]

This is often the easiest task, typically defined by:

Deciding on what attributes and levels fully describe the good or service[edit]

A good or service, for instance mobile (cell) phone, is typically described by a number of attributes (features). Phones are often described by shape, size, memory, brand, etc. The attributes to be varied in the DCE must be all those that are of interest to respondents. Omitting key attributes typically causes respondents to make inferences (guesses) about those missing from the DCE, leading to omitted variable problems. The levels must typically include all those currently available, and often are expanded to include those that are possible in future – this is particularly useful in guiding product development.

Constructing an experimental design that is appropriate for those attributes and levels, either from a design catalogue, or via a software program[edit]

A strength of DCEs and conjoint analyses is that they typically present a subset of the full factorial. For example, a phone with two brands, three shapes, three sizes and four amounts of memory has 2x3x3x4=72 possible configurations. This is the full factorial and in most cases is too large to administer to respondents. Subsets of the full factorial can be produced in a variety of ways but in general they have the following aim: to enable estimation of a certain limited number of parameters describing the good: main effects (for example the value associated with brand, holding all else equal), two-way interactions (for example the value associated with this brand and the smallest size, that brand and the smallest size), etc. This is typically achieved by deliberately confounding higher order interactions with lower order interactions. For example, two-way and three-way interactions may be confounded with main effects. This has the following consequences:

Thus, researchers have repeatedly been warned that design involves critical decisions to be made concerning whether two-way and higher order interactions are likely to be non-zero; making a mistake at the design stage effectively invalidates the results since the hypothesis of higher order interactions being non-zero is untestable.[11]

Designs are available from catalogues and statistical programs. Traditionally they had the property of Orthogonality where all attribute levels can be estimated independently of each other. This ensures zero collinearity and can be explained using the following example.

Imagine a car dealership that sells both luxury cars and used low-end vehicles. Using the utility maximisation principle and assuming an MNL model, we hypothesise that the decision to buy a car from this dealership is the sum of the individual contribution of each of the following to the total utility.

Using multinomial regression on the sales data however will not tell us what we want to know. The reason is that much of the data is collinear since cars at this dealership are either:

There is not enough information, nor will there ever be enough, to tell us whether people are buying cars because they are European, because they are a BMW or because they are high performance. This is a fundamental reason why RP data are often unsuitable and why SP data are required. In RP data these three attributes always co-occur and in this case are perfectly correlated. That is: all BMWs are made in Germany and are of high performance. These three attributes: origin, marque and performance are said to be collinear or non-orthogonal. Only in experimental conditions, via SP data, can performance and price be varied independently – have their effects decomposed.

Anexperimental design (below) in a Choice Experiment is a strict scheme for controlling and presenting hypothetical scenarios, or choice sets to respondents. For the same experiment, different designs could be used, each with different properties. The best design depends on the objectives of the exercise.

It is the experimental design that drives the experiment and the ultimate capabilities of the model. Many very efficient designs exist in the public domain that allow near optimal experiments to be performed.

For example the Latin square1617 design allows the estimation of all main effects of a product that could have up to 1617 (approximately 295 followed by eighteen zeros) configurations. Furthermore this could be achieved within a sample frame of only around 256 respondents.

Below is an example of a much smaller design. This is 34 main effects design.

0 0 0 0
0 1 1 2
0 2 2 1
1 0 1 1
1 1 2 0
1 2 0 2
2 0 2 2
2 1 0 1
2 2 1 0

This design would allow the estimation of main effects utilities from 81 (34) possible product configurations assuming all higher order interactions are zero. A sample of around 20 respondents could model the main effects of all 81 possible product configurations with statistically significant results.

Some examples of other experimental designs commonly used:

More recently, efficient designs have been produced.[16][17] These typically minimise functions of the variance of the (unknown but estimated) parameters. A common function is the D-efficiency of the parameters. The aim of these designs is to reduce the sample size required to achieve statistical significance of the estimated utility parameters. Such designs have often incorporated Bayesian priors for the parameters, to further improve statistical precision.[18] Highly efficient designs have become extremely popular, given the costs of recruiting larger numbers of respondents. However, key figures in the development of these designs have warned of possible limitations, most notably the following.[16] Design efficiency is typically maximised when good A and good B are as different as possible: for instance every attribute (feature) defining the phone differs across A and B. This forces the respondent to trade across price, brand, size, memory, etc; no attribute has the same level in both A and B. This may impose cognitive burden on the respondent, leading him/her to use simplifying heuristics ("always choose the cheapest phone") that do not reflect his/her true utility function (decision rule). Recent empirical work has confirmed that respondents do indeed have different decision rules when answering a less efficient design compared to a highly efficient design.[19]

More information on experimental designs may be found here. It is worth reiterating, however, that small designs that estimate main effects typically do so by deliberately confounding higher order interactions with the main effects. This means that unless those interactions are zero in practice, the analyst will obtain biased estimates of the main effects. Furthermore (s)he has (1) no way of testing this, and (2) no way of correcting it in analysis. This emphasises the crucial role of design in DCEs.

Constructing the survey[edit]

Constructing the survey typically involves:

Administering the survey to a sample of respondents in any of a number of formats including paper and pen, but increasingly via web surveys[edit]

Traditionally, DCEs were administered via paper and pen methods. Increasingly, with the power of the web, internet surveys have become the norm. These have advantages in terms of cost, randomising respondents to different versions of the survey, and using screening. An example of the latter would be to achieve balance in gender: if too many males answered, they can be screened out in order that the number of females matches that of males.

Analysing the data using appropriate models, often beginning with the multinomial logistic regression model, given its attractive properties in terms of consistency with economic demand theory[edit]

Analysing the data from a DCE requires the analyst to assume a particular type of decision rule - or functional form of the utility equation in economists' terms. This is usually dictated by the design: if a main effects design has been used then two-way and higher order interaction terms cannot be included in the model. Regression models are then typically estimated. These often begin with the conditional logit model - traditionally, although slightly misleadingly, referred to as the multinomial logistic (MNL) regression model by choice modellers. The MNL model converts the observed choice frequencies (being estimated probabilities, on a ratio scale) into utility estimates (on an interval scale) via the logistic function. The utility (value) associated with every attribute level can be estimated, thus allowing the analyst to construct the total utility of any possible configuration (in this case, of car or phone). However, a DCE may alternatively be used to estimate non-market environmental benefits and costs.[20]

Strengths[edit]

Weaknesses[edit]

The mean-variance confound[edit]

Yatchew and Griliches first proved that means and variances were confounded in limited dependent variable models (where the dependent variable takes any of a discrete set of values rather than a continuous one as in conventional linear regression).[22] This limitation becomes acute in choice modelling for the following reason: a large estimated beta from the MNL regression model or any other choice model can mean:

  1. Respondents place the item high up on the latent scale (they value it highly), or
  2. Respondents do not place the item high up on the scale BUT they are very certain of their preferences, consistently (frequently) choosing the item over others presented alongside, or
  3. Some combination of (1) and (2).

This has significant implications for the interpretation of the output of a regression model. All statistical programs "solve" the mean-variance confound by setting the variance equal to a constant; all estimated beta coefficients are, in fact, an estimated beta multiplied by an estimated lambda (an inverse function of the variance). This tempts the analyst to ignore the problem. However (s)he must consider whether a set of large beta coefficients reflect strong preferences (a large true beta) or consistency in choices (a large true lambda), or some combination of the two. Dividing all estimates by one other – typically that of the price variable – cancels the confounded lambda term from numerator and denominator.[23] This solves the problem, with the added benefit that it provides economists with the respondent's willingness to pay for each attribute level. However, the finding that results estimated in "utility space" do not match those estimated in "willingness to pay space",[24][25] suggests that the confound problem is not solved by this "trick": variances may be attribute specific or some other function of the variables (which would explain the discrepancy). This is a subject of current research in the field.

Versus traditional ratings-based conjoint methods[edit]

Major problems with ratings questions that do not occur with choice models are:

Other types[edit]

Ranking[edit]

Rankings do tend to force the individual to indicate relative preferences for the items of interest. Thus the trade-offs between these can, like in a DCE, typically be estimated. However, ranking models must test whether the same utility function is being estimated at every ranking depth: e.g. the same estimates (up to variance scale) must result from the bottom rank data as from the top rank data.

Best–worst scaling[edit]

Best–worst scaling (BWS) is a well-regarded alternative to ratings and ranking. It asks people to choose their most and least preferred options from a range of alternatives. By subtracting or integrating across the choice probabilities, utility scores for each alternative can be estimated on an interval or ratio scale, for individuals and/or groups. Various psychological models may be utilised by individuals to produce best-worst data, including the MaxDiff model.

Uses[edit]

Choice modelling is particularly useful for:

The section on "Applications" of discrete choice provides further details on how this type of modelling can be applied in different fields.

Occupational choice model[edit]

InEconomics, an occupational choice model is a model that seeks to answer why people enter into different occupations [28] .[29]

In the model, in each moment, the person decides whether to work as in the previous occupation, in some other occupation, or not to be employed. In some versions of the model, an individual chooses that occupation for which the present value of his expected income is a maximum.[30] However, in other versions, risk aversion may drive people to work in the same occupation as before.[31]

See also[edit]

References[edit]

  1. ^ 2001 - Centre for International Economics - Review of willingness-to-pay methodologies
  • ^ a b c Louviere, Jordan J; Flynn, Terry N; Carson, Richard T (2010-01-01). "Discrete Choice Experiments Are Not Conjoint Analysis". Journal of Choice Modelling. 3 (3): 57–72. doi:10.1016/S1755-5345(13)70014-9. hdl:10453/15890.
  • ^ Journal of Choice Modelling. Retrieved 2015-11-05. {{cite book}}: |website= ignored (help)
  • ^ Thurston (1994). "A law of comparative judgment". APA PsycNET. Retrieved 2017-12-08.
  • ^ a b Zarembka, Paul (1974). Frontiers in Econometrics. New York: Academic Press. pp. 105–142.
  • ^ Luce, R. Duncan (1959). Conditional logit analysis of qualitative choice behavior. New York: John Wiley & Sons.
  • ^ Marley, A. A. J. (1968-06-01). "Some probabilistic models of simple choice and ranking". Journal of Mathematical Psychology. 5 (2): 311–332. doi:10.1016/0022-2496(68)90078-3.
  • ^ Economics 2000
  • ^ Louviere, Jordan J.; Woodworth, George (1983-11-01). "Design and Analysis of Simulated Consumer Choice or Allocation Experiments: An Approach Based on Aggregate Data". Journal of Marketing Research. 20 (4): 350–367. doi:10.2307/3151440. JSTOR 3151440.
  • ^ a b Louviere, Jordan J.; Hensher, David A. (1982-01-01). "DESIGN AND ANALYSIS OF SIMULATED CHOICE OR ALLOCATION EXPERIMENTS IN TRAVEL CHOICE MODELING". Transportation Research Record (890). ISSN 0361-1981.
  • ^ a b "Stated Choice Methods". Cambridge University Press. Retrieved 2015-11-04.
  • ^ Discrete Choice Analysis. Transportation Studies. 18 December 1985. ISBN 9780262022170. Retrieved 2015-11-04. {{cite book}}: |website= ignored (help)
  • ^ Baltas, George; Doyle, Peter (2001-02-01). "Random utility models in marketing research: a survey". Journal of Business Research. 51 (2): 115–125. doi:10.1016/S0148-2963(99)00058-2. ISSN 0148-2963.
  • ^ "Orthogonal Arrays". support.sas.com. Retrieved 2015-11-04.
  • ^ "ChoiceMetrics | Ngene | Features". www.choice-metrics.com. Retrieved 2015-11-04.
  • ^ a b Rose, John M.; Bliemer, Michiel C. J. (2009-09-01). "Constructing Efficient Stated Choice Experimental Designs". Transport Reviews. 29 (5): 587–617. doi:10.1080/01441640902827623. ISSN 0144-1647. S2CID 111072085.
  • ^ Street, Deborah J.; Burgess, Leonie (2007-07-20). The Construction of Optimal Stated Choice Experiments: Theory and Methods. John Wiley & Sons. ISBN 9780470148556.
  • ^ [Rossi, P., Allenby, G., McCulloch, R. (2009) Bayesian statistic and marketing. Wiley]
  • ^ Flynn, Terry N (March 2016). "Are Efficient Designs Used In Discrete Choice Experiments Too Difficult For Some Respondents? A Case Study Eliciting Preferences for End-Of-Life Care". Pharmacoeconomics. 34 (3): 273–284. doi:10.1007/s40273-015-0338-z. PMID 26589411. S2CID 207492460.
  • ^ Jeff Bennet University of Queensland https://www.epa.qld.gov.au/publications?id=1585 Archived 2008-08-19 at the Wayback Machine
  • ^ 2001 – The Centre for International Economics – review of willingness-to-pay methodologies
  • ^ Yatchew, Adonis; Griliches, Zvi (1985). "Specification Error in Probit Models". The Review of Economics and Statistics. 67 (1): 134. doi:10.2307/1928444. JSTOR 1928444.
  • ^ Hensher, David; Louviere, Jordan; Swait, Joffre (1998-11-26). "Combining sources of preference data". Journal of Econometrics. 89 (1–2): 197–221. doi:10.1016/S0304-4076(98)00061-X. hdl:2123/19028.
  • ^ Train, Kenneth (2005). Applications of simulation methods in environmental and resource economics. Doredrecht. pp. 1–16.
  • ^ Sonnier, Garrett; Ainslie, Andrew S.; Otter, Thomas (2007). "Heterogeneity Distributions of Willingness-to-Pay in Choice Models". doi:10.2139/ssrn.928412. S2CID 154620696. {{cite journal}}: Cite journal requires |journal= (help)
  • ^ Irwin, Neil (4 September 2014). "Why Colleges With a Distinct Focus Have a Hidden Advantage". The Upshot. The New York Times. Retrieved 9 May 2023.
  • ^ Selingo, Jeffrey J. (September 23, 2015). "When students have choices among top colleges, which one do they choose?". The Washington Post. Retrieved 9 May 2023.
  • ^ Blau, Peter; Gustad, John; Jessor, Richard; Parnes, Herbert; Wilcock, Richard (1956). "Occupational Choice: A Conceptual Framework". ILR Review. 9 (4): 531–543. doi:10.1177/001979395600900401. S2CID 158449519. Retrieved 18 April 2021.
  • ^ McCall, Brian P. (1991). "A dynamic model of occupational choice". Journal of Economic Dynamics and Control. 15 (2): 387–408. doi:10.1016/0165-1889(91)90019-W. Retrieved 18 April 2021.
  • ^ Benewitz, Maurice; Zucker, Albert (1968). "Human Capital and Occupational Choice: A Theoretical Model". Southern Economic Journal. 34 (3): 406–409. doi:10.2307/1055503. JSTOR 1055503.
  • ^ Lang, Kevin (August 2018). "THE DETERMINANTS OF TEACHERS' OCCUPATIONAL CHOICE". NBER Working Paper Series. 24883. Retrieved 18 April 2021.
  • External links[edit]


    Retrieved from "https://en.wikipedia.org/w/index.php?title=Choice_modelling&oldid=1197761231"

    Categories: 
    Choice modelling
    Economics models
    Econometric modeling
    Behavioral economics
    Hidden categories: 
    CS1 errors: periodical ignored
    Webarchive template wayback links
    CS1 errors: missing periodical
    Articles with short description
    Short description matches Wikidata
    Commons category link from Wikidata
     



    This page was last edited on 21 January 2024, at 21:06 (UTC).

    Text is available under the Creative Commons Attribution-ShareAlike License 4.0; additional terms may apply. By using this site, you agree to the Terms of Use and Privacy Policy. Wikipedia® is a registered trademark of the Wikimedia Foundation, Inc., a non-profit organization.



    Privacy policy

    About Wikipedia

    Disclaimers

    Contact Wikipedia

    Code of Conduct

    Developers

    Statistics

    Cookie statement

    Mobile view



    Wikimedia Foundation
    Powered by MediaWiki