UCSD MGT 100 Week 06
Discrete heterogeneity by segment
Continuous heterogeneity by customer attributes
Individual-level demand parameters
- We'll do 1 & 2
Recall our MNL market share function \(s_{jt}=\frac{e^{x_{jt}\beta-\alpha p_{jt}}}{\sum_{k=1}^{J}{e^{x_{kt}\beta-\alpha p_{kt}}}}\)
What is \(\alpha\)?
What is \(\beta\)?
What limitations does this model have?
Incorporating customer heterogeneity into demand models can enable a rich array of segment-specific or person-specific customer analytics
MNL can estimate quality; Het demand estimates quality & fit
Better “policy experiments” for variables we {manage,can measure,predict sales}
- Pricing: price discrimination, two-part tariffs, fees, targeted coupons
- Customer relationships: Loyalty bonus, referral bonus, freebies
- Social media: Posts, likes, shares, comments, reviews
- Advertising: Ad frequency, content, media, channels
- Product attributes: Targeted attributes, line extensions, brand extensions
- Distribution: Partner selection, intensity/shelfspace, exclusion, in-store environment
Predicts M&A results; oft used in antitrust
Let \(w_{it}\sim F(w_{it})\) be observed customer attributes that drive demand, e.g. usage
\(w_{it}\) is often a vector of customer attributes including an intercept
Assume \(\alpha=\gamma w_{it}\) and \(\beta=\delta w_{it}\)
Then \(u_{ijt}=x_{jt}\delta w_{it}- p_{jt}\gamma w_{it} +\epsilon_{ijt}\) and
\[s_{jt}=\int \frac{e^{x_{jt}\delta w_{it}- p_{jt}\gamma w_{it}}}{\sum_{k=1}^{J}e^{x_{jt}\delta w_{it}- p_{jt}\gamma w_{it}}} dF(w_{it}) \approx \frac{1}{N_t}\sum_i \frac{e^{x_{jt}\delta w_{it}- p_{jt}\gamma w_{it}}}{\sum_{k=1}^{J}e^{x_{jt}\delta w_{it}- p_{jt}\gamma w_{it}}}\]
- We usually approximate this integral with a Riemann sum
What goes into \(w_{it}\)?
What if \(dim(x)\) or \(dim(w)\) is large?
Assume \((\alpha_i,\beta_i)\sim F(\Theta)\)
- Includes the Hierarchical Bayesian Logit from weeks 2&3
Then \(s_{jt}=\int\frac{e^{x_{jt}\alpha_i-\beta_i p_{jt}}}{\sum_{k=1}^{J}e^{x_{jt}\alpha_i-\beta_i p_{jt}}}dF(\Theta)\)
Typically, we assume \(F(\Theta)\) is multivariate normal, for convenience, and estimate \(\Theta\)
Humans choose the model
How do you know if you specified the right model?
- Hints: No model is ever "correct." No assumption is ever "true" (why not?)
How do you choose among plausible specifications?
Pros and cons of model enrichments or simplifications?
Bias-variance tradeoff
- Adding predictors always increases model fit
- Yet parsimony often improves predictions
Many criteria drive model selection
- Modeling objectives
- Theoretical properties
- Model flexibility
- Precedents & prior beliefs
- In-sample fit
- Prediction quality
- Computational properties
Retrodiction = “RETROspective preDICTION”
- Knowing what happened enables you to evaluate prediction quality
- We can compare different models and different specifications on retrodictions
We can even train a model to maximize retrodiction quality (“Cross-validation”)
- Most helpful when the model's purpose is prediction
- More approaches: Choose intentionally simple models
- Penalize the model for uninformative parameters: Lasso, Ridge, Elastic Net, etc.
- You estimate the model K times
- Each estimation uses a different (K-1)/K proportion of the data
- We evaluate retrodiction quality K times, then average them
- When K=N, we call that "leave-one-out" cross-validation
- Important: cross-validation is just one tool in the toolbox
- Final model selection also depends on theory, objectives, other criteria
Heterogeneous demand models enable personalized and segment-specific policy experiments
Demand models can incorporate discrete, continuous and/or individual-level heterogeneity structures
Heterogeneous models fit better, but will predict worse if overfit