Residual Normality
What Normality Tests Check
Residual normality tests validate the assumption that model errors (residuals) are normally distributed. This assumption is fundamental to regression analysis and affects the validity of confidence intervals, p-values, and hypothesis tests.
Purpose: Validates assumption that errors are normally distributed, which ensures the reliability of statistical inference.
Why Normality Matters
When residuals are normally distributed:
Confidence Intervals are Accurate: The confidence intervals around coefficient estimates are correctly calibrated
P-values are Reliable: Hypothesis tests for variable significance have correct Type I error rates
Predictions are Valid: Prediction intervals accurately reflect uncertainty
Mild violations of normality are often acceptable in Marketing Mix Modeling, especially with larger sample sizes (n > 30), due to the Central Limit Theorem.
Statistical Tests Available
MixModeler provides three complementary normality tests:
Jarque-Bera Test
Tests for normality using skewness and kurtosis
p < 0.05 indicates non-normality
Shapiro-Wilk Test
More powerful normality test for smaller samples
p < 0.05 indicates non-normality
D'Agostino-Pearson Test
Tests based on skewness and kurtosis
p < 0.05 indicates non-normality
Interpretation: If p-value ≥ 0.05, residuals appear normally distributed. If p < 0.05, there is evidence of non-normality.
Visual Diagnostics
MixModeler provides multiple visualizations to assess normality:
Q-Q Plot (Quantile-Quantile Plot):
- Plots residual quantiles against theoretical normal distribution quantiles 
- Good: Points closely follow the diagonal reference line 
- Problem: Systematic deviations from the diagonal line 
Histogram of Residuals:
- Shows the distribution shape of residuals 
- Good: Bell-shaped, symmetric distribution 
- Problem: Heavily skewed or multi-modal distribution 
Distribution Metrics:
- Skewness: Should be close to 0 (symmetric distribution) - Positive skewness: Long right tail 
- Negative skewness: Long left tail 
 
- Kurtosis: Should be close to 3 (normal distribution) - Higher than 3: Heavy tails (more outliers) 
- Lower than 3: Light tails (fewer outliers) 
 
Interpreting Test Results
Passed Tests (✓)
What it means: Residuals appear to be normally distributed (p ≥ 0.05)
Implications:
- Hypothesis tests based on the assumption of normality are valid 
- Confidence intervals are appropriately calibrated 
- Model inference is statistically sound 
Action: No action needed - proceed with using the model
Failed Tests (⚠)
What it means: Residuals may not be normally distributed (p < 0.05)
Implications:
- Hypothesis tests may have incorrect Type I error rates 
- Confidence intervals might be miscalibrated 
- P-values should be interpreted more cautiously 
Common Causes:
- Outliers or influential observations 
- Skewed dependent variable 
- Missing important predictors 
- Non-linear relationships not captured 
- Model misspecification 
What to Do When Tests Fail
If normality tests fail, consider these remedies in order:
1. Check for Outliers
- Review the Influential Points diagnostic 
- Investigate data quality issues 
- Consider removing or down-weighting outliers if justified 
2. Transform the Dependent Variable
- Log transformation for right-skewed data 
- Square root transformation for moderate skewness 
- Box-Cox transformation for optimal normalization 
3. Add Missing Variables
- Include important predictors that might be missing 
- Add interaction terms or non-linear terms 
- Consider time trends or seasonal effects 
4. Use Robust Methods
- Accept minor violations if sample size is large (n > 50) 
- Use robust standard errors in your analysis 
- Consider bootstrapping for inference 
5. When Violations are Acceptable
- Large sample sizes (Central Limit Theorem applies) 
- Business decisions not sensitive to exact p-values 
- Focus is on prediction rather than inference 
- Violations are mild (p-value between 0.01-0.05) 
Practical Guidelines
Acceptable Scenarios:
- Slight non-normality with p-value between 0.01 and 0.05 
- Large datasets (n > 100) with mild skewness 
- Business focus on directional insights rather than precise intervals 
Critical Issues:
- Severe skewness or bimodal distributions 
- Heavy tails with many extreme outliers 
- P-values significantly below 0.01 
- Small sample sizes with clear non-normality 
Example Interpretation
Scenario 1 - Passed:
- Jarque-Bera p-value: 0.23 
- Shapiro-Wilk p-value: 0.18 
- Skewness: -0.12, Kurtosis: 3.05 
Interpretation: Residuals are normally distributed. All normality tests pass. The model satisfies the normality assumption, and hypothesis tests are valid.
Scenario 2 - Failed:
- Jarque-Bera p-value: 0.003 
- Shapiro-Wilk p-value: 0.012 
- Skewness: 1.85, Kurtosis: 8.2 
Interpretation: Residuals show significant positive skewness and heavy tails. This indicates potential outliers or the need for variable transformation. Review influential points and consider log-transforming the KPI.
Related Diagnostics
After reviewing normality:
- Check Influential Points to identify outliers causing non-normality 
- Review Actual vs Predicted plot for systematic patterns 
- Examine Heteroscedasticity tests for variance issues 
Last updated