26.10.2013 Views

Nonparametric Bayesian Discrete Latent Variable Models for ...

Nonparametric Bayesian Discrete Latent Variable Models for ...

Nonparametric Bayesian Discrete Latent Variable Models for ...

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

4.5 A Choice Model with Infinitely Many <strong>Latent</strong> Features<br />

available to the researcher and one would like to infer them from observed choices.<br />

Although the EBA model can account <strong>for</strong> many choice scenarios, its usefulness <strong>for</strong> the<br />

analysis of choice data has been extremely limited by the fact that inference about the<br />

underlying features is very difficult (Tversky and Sattath, 1979; Batsell et al., 2003).<br />

In the following, we present a <strong>Bayesian</strong> <strong>for</strong>mulation and analysis of the EBA model.<br />

We employ the IBP as a prior over the latent binary features of alternatives in the choice<br />

set and infer the latent features and their weights (importance of each feature to the<br />

choice maker) using MCMC and demonstrate the per<strong>for</strong>mance of the proposed model on<br />

a data set studied in psychology. We describe the non-parametric <strong>Bayesian</strong> <strong>for</strong>mulation<br />

of the EBA model in the next section. The MCMC algorithm used <strong>for</strong> inference on the<br />

model is presented in Section 4.5.2, followed by experimental results in Section 4.5.3,<br />

and we conclude with a discussion in Section 4.5.4.<br />

4.5.1 Model Specification<br />

The EBA model is defined <strong>for</strong> choice from a set of several options but <strong>for</strong> clarity of<br />

presentation we consider only the paired comparison case here which reduces the EBA<br />

model to Restle’s choice model (Restle, 1961; Tversky, 1972). Nevertheless, the inference<br />

techniques we describe below are also valid <strong>for</strong> the general EBA model.<br />

In a paired comparison experiment there is a set of N options but the subject is<br />

presented only with pairs of options at a time. The task of the subject is to indicate<br />

which of the two options i and j she prefers. In the EBA model, options are described<br />

by K-dimensional binary vector of features z, referred to as aspects. The probability of<br />

choosing option i over option j is given as<br />

pij =<br />

<br />

k wkzik(1 − zjk)<br />

<br />

k wkzik(1 − zjk) + <br />

k wkzjk(1 − zik)<br />

, (4.63)<br />

where zik denotes the binary value of the kth feature of option i and wk is the positive<br />

weight associated with the kth feature. The greater the weight of a feature, the heavier<br />

its influence on the choice probabilities. The sum <br />

k wkzik(1 − zjk) collects the weights<br />

<strong>for</strong> all the aspects that option i has but option j does not have. There<strong>for</strong>e, the choice<br />

between two alternatives depends only on the features that are not shared between<br />

the two. Equation (4.63) expresses that the EBA model can account <strong>for</strong> the effects of<br />

similarity on choice. For the above equation, we define the ratio 0<br />

0<br />

= 0.5. That is,<br />

if two alternatives have exactly the same set of features, the choice probability is 0.5.<br />

On the other extreme, if the options are characterized only by unique aspects, i.e. no<br />

option shares any feature with any other option, the BTL model is recovered. Recently,<br />

Ruan et al. (2005) have developed a choice model using a Poisson race model in which<br />

the choice probabilities are mainly determined by the difference in the features of the<br />

alternatives, but the similarities are also allowed to have some effect. This model can<br />

be seen as a generalization of the EBA model.<br />

If one option i has all the features that another alternative j has and more features on<br />

top of these then i will always be preferred over j. This is a reasonable assumption but<br />

in real choice data it can happen that subjects occasionally fail to choose alternative<br />

103

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!