Model Building Strategy
Overview
Effective model building combines statistical rigor with business judgment. This guide outlines strategic approaches to building robust, interpretable, and actionable marketing mix models that drive business decisions.
Core Principles
Principle 1: Start Simple, Add Complexity
Begin with Baseline:
- Include only top 5-10 most important variables 
- Use OLS (faster iteration) 
- No transformations initially 
- Establish performance benchmark 
Add Incrementally:
- Add one variable or transformation at a time 
- Evaluate impact on R², coefficients, diagnostics 
- Keep changes that meaningfully improve model 
- Document rationale for each addition 
Why This Works:
- Easier to diagnose problems 
- Understand contribution of each element 
- Avoid over-fitting 
- Build confidence progressively 
Principle 2: Theory Before Statistics
Business Logic First:
- Include variables that should matter theoretically 
- Don't remove variables solely because p-value >0.05 
- Consider business context and domain knowledge 
- Balance statistical significance with practical significance 
Example:
TV_Spend: p=0.08 (not "significant" at 0.05)
But: TV is major channel, large budget, known to drive sales
→ Keep in model, investigate why significance is marginalPrinciple 3: Diagnostics are Not Optional
Always Validate:
- Run diagnostics on every model candidate 
- Don't trust results from models that fail tests 
- Address issues before interpreting coefficients 
- Document which tests passed/failed 
Minimum Required Tests:
- Multicollinearity (VIF) 
- Residual normality 
- Autocorrelation 
- Heteroscedasticity 
Act on Results: Failed tests indicate model problems that must be fixed
Principle 4: Iterate and Compare
Build Multiple Specifications:
- Baseline model 
- Model with adstock 
- Model with saturation curves 
- Model with interactions 
- Simplified "parsimonious" model 
Compare Systematically:
- R², Adjusted R², AIC, BIC 
- Coefficient stability 
- Diagnostic test performance 
- Business reasonableness 
Select Best Overall: Not just highest R², but best balance of fit, parsimony, and interpretability
Step-by-Step Strategy
Step 1: Build Baseline Model
Include:
- KPI as dependent variable 
- Top 5-10 marketing channels (spend or activity) 
- Basic controls (seasonality if obvious) 
Exclude (for now):
- Transformations 
- Interactions 
- Minor variables 
- Complex features 
Evaluate:
- R²: Expect 0.50-0.70 for baseline 
- Coefficients: Check signs make sense 
- Significance: Note which variables are significant 
Example Baseline:
Revenue ~ TV_Spend + Digital_Spend + Print_Spend + Radio_Spend + 
          Promo_Flag + TrendStep 2: Add Control Variables
Seasonality:
- Month indicators or quarter indicators 
- Holiday flags (Thanksgiving, Christmas, etc.) 
- Day-of-week effects (if daily data) 
Trends:
- Linear time trend 
- Year indicators 
- Macro factors (if available) 
External Factors:
- Weather (if relevant) 
- Competitive activity 
- Economic indicators 
Impact: Usually increases R² by 0.05-0.15
Step 3: Apply Adstock Transformations
For Media Variables:
- TV, Radio, Print: Test 50-70% decay 
- Digital, Social: Test 30-50% decay 
- Out-of-Home: Test 60-80% decay 
Process:
- Go to Variable Testing 
- Test multiple adstock rates per variable 
- Select rate with highest t-statistic or contribution 
- Apply to main model 
Evaluation:
- Does adstock improve fit (R²)? 
- Do coefficients make more sense? 
- Are effects more significant? 
Typical Improvement: +0.03 to +0.10 in R²
Step 4: Test Saturation Effects
For High-Spend Channels:
- Channels with >20% of total budget 
- Channels suspected of diminishing returns 
- Channels with wide spending range 
Curves to Test:
- S-curve (ICP): For awareness to consideration 
- Concave (ADBUG): For direct response 
- CDR (Constant Diminishing Returns): General case 
Decision:
- Compare linear vs curved specification 
- Use curve if: - Significantly better fit 
- Passes diagnostic tests 
- Makes business sense 
 
- Otherwise stick with linear 
Advanced: Combine adstock + saturation for same variable
Step 5: Explore Interactions
Test When:
- Theoretical synergy suspected 
- Channels often used together 
- Campaign data suggests interaction 
Create:
- Multiply variables: TV × Digital 
- Test in Variable Testing first 
- Add to model if significant and improves fit 
Be Cautious:
- Interactions increase multicollinearity 
- Harder to interpret 
- Can overfit with small samples 
- Only include if strong evidence 
Step 6: Refine and Simplify
Remove Weak Variables:
- Non-significant with p >0.10 
- Coefficients near zero 
- High VIF (>10) with other variables 
- Don't contribute to business decisions 
Combine Similar Variables:
- Multiple social platforms → Total_Social 
- Similar media → Combined category 
- Reduces multicollinearity 
- Improves stability 
Final Model Should Have:
- 10-30 variables (typical) 
- All VIF <10 (ideally <5) 
- Most variables significant (p <0.10) 
- Clear business interpretation 
Step 7: Validate Thoroughly
Run All Diagnostics:
- Normality: Shapiro-Wilk, Jarque-Bera 
- Autocorrelation: Durbin-Watson 
- Heteroscedasticity: Breusch-Pagan 
- Multicollinearity: VIF 
- Influential points: Cook's Distance 
Sanity Checks:
- Coefficients have expected signs 
- Magnitudes are reasonable 
- Seasonality aligns with business knowledge 
- Top channels match expectations 
Stress Tests:
- Remove one variable at a time - are results stable? 
- Split sample in half - consistent estimates? 
- Compare to previous models - similar findings? 
Model Selection Criteria
Statistical Fit
**R² and Adjusted R²
Last updated