Marketing Mix Modeling Solutions Provider

Consultancy

Leverage a wide range of Marketing Mix Modeling Consultancy services
Learn More →

Managed Services

Discover a modern approach to running Marketing Mix Modeling Projects
Learn More →

Training

Develop your Marketing Mix Modeling Knowledge
Learn More →

MassTer

End-to-end Marketing Mix Modeling Software
Learn More →

MassFeeds

Data Preparation Software
Learn More →

Insight

Data Optimization Software
Learn More →

Try our Solutions now!

MASS Analytics offers the best-in-class Marketing Mix Modeling Solutions.

Careers

Start Your Journey Today and Join our team!
Learn More →

About MASS Analytics

Where Marketing Meets Measurement
Learn More →

Contact Us

Get Your Questions Answered
Learn More →

Information Security Policy Statement

We Comply with the highest Information Security Standards
Learn More →

#1 Best Place to Work

At MASS Analytics, we place a high value on an atmosphere that encourages innovation and teamwork

Blogs

Discover the ins and outs of MMM

Use Cases

Explore real-life projects

White Papers

Get an in-depth look into MMM

Events & Webinars

Watch a recap of our events

FAQs

Find answers to your MMM questions

Marketing Mix Modeling Guide

Empower yourself with the knowledge and tools to run successful Marketing Mix Modeling projects.

Retail

Marketing Mix Modeling for Multi-regional Retailers
Learn More →

CPG

Marketing Mix Modeling for CPG Companies
Learn More →

E-commerce

Marketing Mix Modeling for E-commerce
Learn More →

Finance

Marketing Mix Modeling for Financial Institutions
Learn More →

7 Stats to Evaluate Your Marketing Mix Model

In the previous part of this MMM Guide series, we explored one of the corner stones of Marketing Mix Modeling, which is Regression Analysis. In this article we cover some of the ways in which you can ensure the robustness of your model and reliability of your results. This step is very important and should be performed at the modeling stage of the MMM workflow. Now that you have built a marketing mix model that makes sense, how do you make sure it’s statistically sound?

There are a number of tests you need to conduct to be sure. Here are the key statistics that should be computed and analyzed to evaluate the robustness of your marketing mix model.

1. R Squared: How much Variance is explained by the model

The coefficient of determination, or R Squared for short (often referred to as R², measures the percentage of the total variance that is explained by the model. For example, if R squared is equal to 85%, it means that 85% of the variance in the Sales data is explained by the variables in the model e.g. media promotions, distribution, seasonality etc.. A modified version of the R² is the adjusted R². One of the drawbacks of the R² is that it increases as the number of variables in the model increases regardless of their explanatory power. The adjusted R-squared on the other hand only increases when the new variables added to the equation improve the model more than what would be expected by chance. It also decreases when a predictor/independent variable improves the model by less than expected. Adjusted R Squared is often used to compare the explanatory power of two models.

2. F Statistic: Overall Model Relevance/Significance

The F test measures the overall significance of the model. H0 that is related to this F test, states that all the coefficients are equal to zero at the same time.

Taking the same example (if R squared is equal to 85%, it means that 85% of the variance in the Sales data is explained by the variables in the model e.g. media promotions, distribution, seasonality etc..), the test is to see whether all the coefficients of media, promotions, seasonality, and distribution (i.e all the variables in the equation) take zero at the same time.

If this is the case, it means that overall, the model is not significant and none of the variables seem to impact the dependent variable significantly. If the F test turns to be significant, then you need to proceed and test every single variable separately and statistically. Be aware though that rejecting H0 of the F stat does not mean that every single variable in the equation is significant.

3. T-Statistic: Relevance Significance of one predictor

The t-stat is a measure of coefficient reliability. It is the ratio between the computed (or the estimated) coefficient and the standard error of the coefficient.

The bigger the t-stat the better, as it means the coefficient is statistically different from 0 (0 indicates absence of relationship between the dependent and the independent variable). In other terms, if another sample is drawn, it is likely to obtain a similar or close measure of the coefficient and that value is ensured to be different enough from 0.

The t-stat is computed for each independent variable in the equation. So as a rule of thumb, if the t value of a variable (search for example) is greater than two, then the variable is statistically significant and search has a significant impact on sales.

4. The Standard Error of The Estimate (SEE) and the MAPE: Model Accuracy

The F test measures the overall significance of the model. H0 that is related to this F test, states that all the coefficients are equal to zero at the same time.

Taking the same example (if R squared is equal to 85%, it means that 85% of the variance in the Sales data is explained by the variables in the model e.g. media promotions, distribution, seasonality etc..), the test is to see whether all the coefficients of media, promotions, seasonality, and distribution (i.e all the variables in the equation) take zero at the same time.

If this is the case, it means that overall, the model is not significant and none of the variables seem to impact the dependent variable significantly. If the F test turns to be significant, then you need to proceed and test every single variable separately and statistically. Be aware though that rejecting H0 of the F stat does not mean that every single variable in the equation is significant.

5. Durbin Watson: Checking for Autocorrelation

Durbin-Watson (DW) is used to test whether the residuals of the model are independent of one another.

Autocorrelation is one of the assumptions of the OLS estimation method This statistic varies between 0 and 4. A value of 2.0 indicates the absence of autocorrelation.

6. The Variance Inflation Factor: Checking for Multicollinearity

The Variance inflation factor, or VIF for short, indicates whether the explanatory (independent) variables are correlated.

In regression analysis, the VIF statistic should be under five for every single variable. If it is not the case, it signals that there is a multicollinearity problem.

Multicollinearity is one of the most challenging issues in MMM projects. In fact, brands tend to deploy several channels at the same time, especially the digital ones (search, display YouTube..).

This results into multicollinearity problems and difficulties in disentangling the impact of each channel.

7. Jarque Bera, Kurtosis, Skewness: Checking for Normality

Jarque Bera, kurtosis, skewness, and others are different ways to test the normality of the residuals. The latter should be normally distributed. 

It’s Not Just About Statistics, Think About the Business Story Too!

Establishing that a model makes sense statistically is crucial. But that is only one half of the story; MMM is about modeling consumer behavior, hence it is also crucial that the model makes commercial and business sense. For example:

  • When modeling a CPG brand, it is not possible to accept that price has a positive impact on sales. One cannot present results that say: the higher the price, the higher the sales. That is counter intuitive.
  • If a business is spending $20 million on advertising, and the model says that the advertising coefficient is negative, then they need to re-evaluate their model to see if they made the right transformations.

Not only should you take the time to evaluate the equation from a statistical standpoint, but you should also assess whether the level of elasticity, seasonality, ROI, and the different parameters that you are estimating make sense business-wise. You should be able to seamlessly defend them when presenting the results back to the client or your internal team.

Modeling is an iterative process. You will often find yourself going through multiple models before you are finally satisfied with the results you’re getting. With rigorous model evaluation, you can be confident you are on the right track in your marketing mix modeling journey!

R Squared

R^2=1-\frac{RSS}{TSS}

OR

R^{2}=\frac{\sum \left ( \widehat{Y}-\bar{Y} \right )^{2}}{\sum \left ({Y}-\bar{Y} \right )^{2}}=\frac{ESS}{TSS}

Adjusted R Squared

\bar{R^{2}}=1-\frac{(1-R^{2})(N-1)}{N-p-1}

OR

\bar{R^{2}}=1-\frac{RSS\times \frac{1}{(n-k)}}{TSS\times \frac{1}{n-1}}

k denotes the number of independent variables (k=p-1)

F Statistic

F=\frac{\sigma_{1}^{2}}{\sigma_{2}^{2}}

where

\sigma^2 = \frac{\sum (x-\mu)^2}{n}

Testing the Overall Significance of the Regression (F-Test)

H_{0}:\beta _{1}=\beta _{2}=...=\beta _{k}=0

H0:not all the parameters are 0

\frac{\sum(\widehat{y}-\bar{y})^2\times \frac{1}{(k-1)} }{\sum\widehat{\mu}^2\times \frac{1}{(n-k)}}=\frac{ESS\times\frac{1}{(k-1)}}{RSS\times\frac{1}{(n-k)}}\sim F_{k-1,n-k}

So if Fcal > Ftable reject H 0

We can calculate an Fcal value from the of value of R Squared

F_{cal}=\frac{R^2\times\frac{1}{(k-1)}}{(1-R^2)\times\frac{1}{(n-k)}}\sim F_{k-1,n-k}

T-Statistic

t_{\hat{\beta}}=\frac{\hat\beta-\beta_{0}}{SE(\hat{beta})}

The Standard Error of the Estimate

SEE =\sqrt{\frac{\sum (x_{i}-\bar{x})}{n-1}}

Simple Regression Model

Y = \beta_{1}+\beta_{2}x

The Standard Error of the Slope

\sigma(\hat{\beta_{2}})=\frac{\hat{\sigma}}{\sum (x_{2i}-\bar{x_{2}})}

The Standard Error of the Regression

\sigma=\sqrt{\frac{RSS}{n-2}}=\sqrt{\frac{TSS-ESS}{n-2}}

Multiple Regression Model

Y=\beta_{0}+\beta_{1}x_{1}+\beta_{2}x_{2}+...+\beta_{k}x_{k}

The Standard Error of the Regression

\hat{\sigma_{\mu}^2}=\frac{\sum \hat{\mu}^2}{(n-k)}=\frac{RSS}{(n-k)}

The Standard Error of the slope

\widehat{Var}(\hat\beta)=\hat\sigma_{\mu}^2(X^{T}X)^{-1}

s.e.(\hat{\beta_{j}})=\hat{\sigma_{\mu}}\sqrt{a_{jj}}

\widehat{Var}\left(\hat{\beta}\right)=\left(\begin{matrix}var\left({\hat{\beta}}_1\right)&cov\left({\hat{\beta}}_1,{\hat{\beta}}_2\right)&cov\left({\hat{\beta}}_1,{\hat{\beta}}_3\right)\\cov\left({\hat{\beta}}_1,{\hat{\beta}}_2\right)&var\left({\hat{\beta}}_2\right)&cov\left({\hat{\beta}}_2,{\hat{\beta}}_3\right)\\cov\left({\hat{\beta}}_1,{\hat{\beta}}_3\right)&cov\left({\hat{\beta}}_2,{\hat{\beta}}_3\right)&var\left({\hat{\beta}}_3\right)\\\end{matrix}\right)

MAPE

M=\frac{1}{n}\sum_{t=1}^{n}\left | \frac{A_{t}-F_{t}}{A_{t}} \right |

Durbin Watson

d=\frac{\sum_{t=2}^{T}(e_t-e_{t-1})^2}{\sum_{t=1}^{T}e^2_t}

The Variance Inflation Factor

VIF=\frac{1}{1-R^2_{i}}

Jarque Bera

JB=\frac{(n-k)}{6}(S^2+\frac{(K-3)^2}{4})

MassTer Corner - MASS Analytics

Model Statistics in MassTer

With MassTer, you can evaluate your model using various statistics. When building and tweaking your model, these statistics will be instantaneously updated and displayed on a table above your model.