Bayesian Time Varying Coefficient Model with Applications to Marketing Mix Modeling
Abstract.
Both Bayesian and varying coefficient models are very useful tools in practice as they can be used to model parameter heterogeneity in a generalizable way. Motivated by the need of enhancing Marketing Mix Modeling (MMM) at Uber, we propose a Bayesian Time Varying Coefficient (BTVC) model, equipped with a hierarchical Bayesian structure. This model is different from other time-varying coefficient models in the sense that the coefficients are weighted over a set of local latent variables following certain probabilistic distributions. Stochastic Variational Inference (SVI) is used to approximate the posteriors of latent variables and dynamic coefficients. The proposed model also helps address many challenges faced by traditional MMM approaches. We used simulations as well as real-world marketing datasets to demonstrate our model’s superior performance in terms of both accuracy and interpretability.
1. Introduction
Marketing as an essential growth driver accounts for sizable investment levels at many companies. Given these large investments, it is not surprising that understanding the return and optimizing the allocation of marketing investment is of foundational importance to marketing practitioners. For many decades Marketing Mix Model (a.k.a. MMM) has been leveraged as one of the most important tools in marketers’ arsenal to address such needs. Recent consumer privacy initiatives (e.g., Apple’s announcement of no-IDFA 111https://developer.apple.com/app-store/user-privacy-and-data-use/ in iOS 14) further underscores the strategic importance of future-proofing any marketing measurement game plan with MMM.
While randomized experiments (Vaver and Koehler, 2011) and causal models (Imbens and Rubin, 2015) are often used for causal inference, they can be either costly or simply infeasible (Lewis and Rao, 2015) under some circumstances. As an alternative, MMM offers a solution by leveraging aggregated time-series data and regression to quantify the relationship between marketing and demand (McCarthy, 1978). MMM could be further tailored and enhanced for different requirements and purposes such as controlling for seasonality, trend, and other control factors (Larsen, 2018) and introducing the geo-level hierarchy (Sun et al., 2017). More importantly, the primary use case of MMM is often not to predict the sales, but rather to quantify the marginal effects of the different marketing tactics.
There are various issues and challenges that have to be accounted for when building MMM (Chan and Perry, 2017). First, advertising media is evolving at a fast pace so it requires modelers to take into account new marketing levers constantly. It ends up forcing modelers to face the “small large ” problem. Second, in order to have actionable insights, modelers tend to pick a high level of data granularity. However, higher level of data granularity may lead to sparse observations and outliers. Practitioners need to strike a balance between the limited amount of reliable historical data and a proper level of data granularity. Third, the sequential nature of the data makes it more susceptible to correlated errors, which violates the basic model assumption of ordinary least squares (Rawlings et al., 2001). Fourth, there are severe endogeneity and multicollinearity concerns due to common marketing planning practices and media dynamics. For instance, setting marketing budget as percent of expected revenue is widely used, which contributes to both endogeneity as well as multicollinearity (i.e., highly correlated channel-level spend) in the models. Self-selection bias especially for the demand-capturing channels such as branded paid search (Blake et al., 2015) can also lead to inflated measurement results if not properly addressed. Fifth, in practice MMM usually involves a large amount of investment and a diverse set of stakeholders with whom alignments need to be secured. As such the bar for model interpretability is very high. Lastly, it is often hard to rely on traditional machine learning approaches such as cross-validation when tuning parameters and choosing models for MMM. Modelers cannot just fall back on that, since there is rarely enough data and/or the holdout periods may not be representative of the challenging series to forecast.
It has been a long journey to build an in-house MMM solution from zero to one at Uber, which takes collaborative efforts across marketers, engineers and data scientists. Throughout this journey, we seek to address all of the above challenges. The preferred modeling solution needs to have the capability of deriving time-varying elasticity along with other temporal patterns on observational studies. More importantly, randomized experimentation results, which are generally deemed as the golden standard in measuring causality, should be incorporated to calibrate the marginal effects in marketing levers. Benefits from both experimentation and regression modeling can be maximized when combined into one holistic framework.
In this paper, we introduce a class of Bayesian time varying coefficient (BTVC) models that power Uber’s MMM solution. Our work brings the ideas of Bayesian modeling and kernel regression together. The Bayesian framework allows a natural way to incorporate experimentation results, and understand the uncertainty of measurement for different marketing levers. The kernel regression is used to produce the time-varying coefficients to capture the dynamics of marketing levers in an efficient and robust way.
The remainder of this paper is organized as follows. In section 2, we describe the problem formulation and the related work. In section 3, we discuss the proposed modeling framework with emphasis on applications to MMM. In section 4, simulations as well as real-case benchmark studies are presented. In section 5, we talk about how to deploy the proposed models using Uber’s modern machine learning platform. section 6 is about the conclusion.
2. Problem Formulation
2.1. Basic Marketing Mix Model
Expressing sales as a function of spending variables with diminishing marginal returns (Farris et al., 2015) is one of the fundamental properties in an attribution or marketing response model. In practice, Han and Gabor (2020), and Lewis and Wong (2018) adopted a similar strategy for their budget allocation, bidding, and attribution. In view of that, our model can be expressed in a multiplicative format as below
(1) |
where are the regressors (i.e., the ads spending variables in our case), is the marketing response, is a time-series process, is the cost curve function, is the number of regressors, and is the number of time points. Choice of is desired to have the following properties such that
-
•
has an explainable structure to decompose into different driving factors,
-
•
temporal effects such as trend and seasonality of are captured,
-
•
is differentiable and monotonically increasing,
-
•
has diminishing marginal returns with respect to .
Equation 1 has an intuitive form as
(2) |
where is the trend component, is the seasonality, and are channel-specific time-varying coefficients.
2.2. Related Work
With a log-log transformation, equation (2) can be re-written as
(3) | ||||
A natural idea is to use state-space models such as Dynamic Linear Model (DLM) (West and Harrison, 2006) or Kalman filter (Durbin and Koopman, 2012) to solve equation (3). However, there are some caveats associated with these approaches, especially given the goal we want to achieve with MMM:
-
•
DLM with Markov Chain Monte Carlo(MCMC) sampling is not efficient and could be costly especially for high-dimensional problems, which requires sampling for a large number of regressors and time steps.
-
•
Although the Kalman filter provides analytical solutions, it has limited room for further customization such as applying restrictions on coefficient signs (e.g., positive coefficient sign for marketing spend) or a t-distributed noise process that is more outlier robust.
Meanwhile, there have been parametric and non-parametric statistical methods proposed (Fan and Zhang, 2008). Wu and Chiang (2000) considered a nonparametric varying coefficient regression model with longitudinal dependent variable and cross-sectional covariates. Two kernel estimators based on componentwise local least squares criteria were proposed to estimate the time varying coefficients. Li et al. (2002) proposed a semiparametric smooth coefficient model as a useful yet flexible specification for studying a general regression relationship with time varying coefficients. It used a local least squares method with a kernel weight function to estimate the smooth coefficient function. Nonetheless, the frequentist approaches can be expensive when doing local estimates with respect to time dimension. There are also no straightforward ways to incorporate information from experimentation results. As such, we are motivated to develop a new approach to derive time varying coefficients under a Bayesian framework for our MMM applications.
3. Methods
3.1. Time Varying Coefficient Regression
In view of the increased complexity of regression problem in practical MMM, we propose a Bayesian Time Varying Coefficient (BTVC) model, as inspired by the Generalized Additive Models (GAM) (Hastie and Tibshirani, 1990) and kernel regression smoothing. The key idea behind BTVC is to express regression coefficients as a weighted sum of local latent variables.
First, we define a latent variable for the -th regressor at time , , , . There are latent variables in total for each regressor. From the perspective of spline regression, can be viewed as a knot distributed at time for a regressor. is a time-based weighting function such that
(4) |
It is intuitive to use a weighting function taking into account the time distance between and ,
(5) |
where is the kernel function, and the denominator is to normalize the weights across knots. In practice, we have different choices for the kernel functions, such as Gaussian kernel, quadratic kernel or any other custom kernels. In subsection 3.3, we will discuss this in more detail.
We can also rewrite Equation 4 into a matrix form
(6) |
where is the coefficient matrix with entries , is the kernel matrix with normalized weight entries , and is the knot matrix with entries
(7) |
where and is the -th row of regressor covariate matrix.
Besides the regression component, we can also apply Equation 4 to other components such as trend and seasonality in Equation 3. Specifically, for the trend component,
(8) | ||||
The trend component can be viewed as a dynamic intercept. For the seasonality component,
(9) | ||||
is the -th row of seasonality covariate matrix derived from Fourier series. In subsection 3.4, we will discuss the seasonality in more detail.
Instead of estimating the local knots directly (i.e., , , and in the above equations) by optimizing an objective function, we introduce the Bayesian framework along with customizable priors to conduct the posterior sampling.
3.2. Bayesian Framework
To capture the sequential dynamics and cyclical patterns, we use the Laplace prior to model adjacent knots
(10) | ||||
The initial values ( and ) can be sampled from a Laplace distribution with mean 0. A similar approach can be found in models implemented in Facebook’s Prophet package (Taylor and Letham, 2018), which uses Laplace prior to model adjacent change points of the trend component.
For the regression component, we introduce a two-layer hierarchy for more robust sampling due to the sparsity in the channel spending data,
(11) | ||||
where the superscript means a folded normal distribution (positive restriction on the coefficient signs).
In the hierarchy, the latent variable depicts the overall mean of a set of knots of a single marketing lever. We can treat this as the overall estimate of a channel coefficient across time. This provides two favorable behaviors for the model: during a period with absence of spending for a channel, coefficient knot estimation of such a channel exhibits a shrinkage effect towards the overall estimate; it helps fight against over-fitting due to the volatile local structure.
The two-layer hierarchy is widely adopted in hierarchical Bayesian models and the shrinkage property is sometimes called the pooling effect on regression coefficients (Gelman et al., 2013). Figure 1 depicts the model flowchart of BTVC. Stochastic Variational Inference (Hoffman et al., 2013) is used to estimate the knot coefficient posteriors from which time varying coefficient estimates can be derived using the formulas above in Equation 7, Equation 8, and Equation 9.
3.3. Kernel Selection
For the kernel function used for trend and seasonality, we propose a customized kernel, i.e.,
when and ,
(12) |
otherwise zero values are assigned. This kernel bears some similarity with the triangular kernel.
For the kernel function used for regression, we adopt the Gaussian kernel, i.e.,
(13) |
where is the scale parameter. Other kernels such as Epanechnikov kernel and quadratic kernel etc., could also be leveraged for the regression component.
3.4. Seasonality
Seasonality is a pattern that repeats over a regular period in a time series. To estimate seasonality, a standard approach is to decompose time-series into trend, seasonality and irregular components using Fourier analysis (De Livera et al., 2011). This method represents the time series by a set of elementary functions called basis such that all functions under study can be written as linear combinations of the elementary functions in the basis. These elementary functions involve the sine and cosine functions or complex exponential. The Fourier series approach describes the fluctuation of time series in terms of sinusoidal behavior at various frequencies.
Specifically, for a given period and a given order , two series and will be generated to capture the seasonality. For example, with daily data, represents the weekly seasonality, while represents the yearly seasonality.
4. Results
4.1. Simulations
4.1.1. Coefficient Curve Fitting
We conduct a simulation study based on the following model
where the trend and are all random walks. The covariates are independent of the error term .
In our study, we compare BTVC with two other time varying regression models available in R CRAN: Bayesian structural time series (BSTS) (Scott and Varian, 2014), and time varying coefficient for single and multi-equation regressions (tvReg) (Casas and Fernandez-Casal, 2021). We set =300 and calculate the average Mean Squared Errors (MSE) against the truth of each regressor across 100 simulations. The estimated coefficient curves of a sample is plotted in Figure 2. The results are reported in Table 1, which demonstrates that BTVC has a better accuracy on the coefficient estimation over the other two models in consideration.
Model | |||
---|---|---|---|
BSTS | 0.0067 | 0.0078 | 0.0080 |
tvReg | 0.0103 | 0.0103 | 0.0096 |
BTVC | 0.0030 | 0.0026 | 0.0029 |
4.1.2. Experimentation Calibration
One appealing property of the BTVC model is its flexibility to ingest any experimentation based priors for any regressors (e.g., advertising channels) since experiments are often deemed as a trustworthy source to tackle the challenges as mentioned in section 1.
To illustrate this feature of BTVC, we first fit a BTVC model on the simulated data, which are generated using a similar simulation scheme as outlined in subsubsection 4.1.1. Next, we assume there is one lift test for the first and third regressors, respectively, and two lift tests for the second regressor. All the tests have a 30-step duration. We use the simulated values as the “truth” derived from the tests, and ingest them as priors into BTVC models. The results are summarized in Figure 3. As expected, the confidence intervals during the ingestion periods and the adjacent neighborhood become narrower, compared to the ones without prior knowledge. Moreover, with this calibration, the coefficient curves are more aligned with the truth around the neighborhood of the test ingestion period. To demonstrate this, in Table 2 we calculated the symmetric mean absolute percentage error (SMAPE) and pinball loss (with 2.5% and 97.5% target quantiles) between the truth and the estimation for the following 30 steps after the prior ingestion period.


SMAPE | Pinball Loss | |||||
w/o priors | w priors | w/o priors | w priors | |||
lower | upper | lower | upper | |||
0.39 | 0.21 | 0.0009 | 0.0032 | 0.0005 | 0.0019 | |
1.37 | 1.25 | 0.0021 | 0. 0019 | 0.0011 | 0.0014 | |
0.30 | 0.18 | 0.0017 | 0.0028 | 0.0014 | 0.0013 |
4.1.3. Shrinkage Property
In real-life MMM data, it is common to observe an intermittent marketing spending pattern, which means that a given advertising channel is characterized by many minuscule or zero spends over time. In BTVC, the coefficient estimation over such sparse periods exhibits a shrinkage effect towards the grand mean of the coefficient curve instead of zero, which is due to the hierarchical structure of the model as discussed in subsection 3.2. Figure 4 is a simulation example to demonstrate this property, where the covariates are plotted along with the estimated coefficients.
4.2. Real Case Studies
4.2.1. Forecasting Benchmark
To benchmark the model’s forecasting accuracy, we conduct a real-case study using Uber Eats’ data across 10 major countries or markets. Each country series consists of the daily number of first orders in Uber Eats by newly acquired users. The data range spans from Jan 2018 to Jan 2021 including a typical Covid-19 period. The scheme change caused by Covid-19 poses a big challenge for modeling.
We compared BTVC with two other time series modeling techniques, SARIMA (Seabold and Perktold, 2010) and Facebook Prophet (Taylor and Letham, 2018). Both Prophet and BTVC models use Maximum A Posterior (MAP) estimates and they are configured as similarly as possible in terms of optimization and seasonality settings. For SARIMA, we fit the structure by maximum likelihood estimation (MLE) where represents the choice of seasonality. In our case, for the weekly seasonality.
We use SMAPE as the performance benchmark metric
where (predicted) or (actual) represents the value measured at time and is the forecast horizon which can also be considered as the “holdout” length in a backtesting process. We used as the forecast horizon with 6 splits (i.e., 6 different cuts of the data with incremental training length) in this exercise.
Figure 5 depicts the SMAPE results across the 10 countries, and Table 3 gives the average and standard deviation of SMAPE values for the three models in consideration. As a result, BTVC outperforms the other two models for the majority of 10 countries in terms of SMAPE.
Model | Mean of SMAPE | Std of SMAPE |
---|---|---|
SARIMA | 0.191 | 0.027 |
Prophet | 0.218 | 0.067 |
BTVC | 0.174 | 0.045 |
4.2.2. Attribution
Simulations in subsubsection 4.1.2 show that ingestion of multiple lift test results can help improve estimation of marketing incrementality. In this section, we use real marketing data to further demonstrate the benefits of MMM calibration with experimentation. Simply put, properly designed and executed experimentation provides the ground truth for marketing measurement. The more valid experiments one can ingest for the measurement of a given marketing lever, the more accuracy one can achieve in terms of understanding its marginal impact, despite experimentation-based insights not being available all the time. BTVC provides a rigorous algorithm that enables informing extrapolation of channel-level elasticity for periods with no experimentation coverage using all available tests, while simultaneously adapting to the new observational data as well as taking into account the differential statistical strength and temporal differences of various tests.
In this real case study, we leverage data and experimentation insights for a given paid channel where three usable experiments are available, covering different but temporally adjacent time periods, i.e., Experiment-1, Experiment-2, Experiment-3. In terms of temporal order, Experiment-1 is the oldest experiment while Experiment-3 is the most recent one.
We construct two attribution validation studies, each addressing a real-life scenario:
-
•
Study 1: Treat Experiment-2 as the unobserved truth while using lift insights from Experiment-1 and Experiment-3 for MMM calibration. This test addresses the scenario where a given paid channel has some experiments but there are non-trivial gaps in between.
-
•
Study 2: Treat Experiment-3 as the unobserved truth while using lift insights from Experiment-1 and Experiment-2 for MMM calibration. This test addresses the scenario where a given paid channel has some old experiments but there are no new experiments planned or the new experiments will take a long time to complete.
For both studies, the following models are fitted:
-
•
Baseline Model 1: No experimentation-based priors will be used.
-
•
Baseline Model 2: Experimentation insights from Experiment-1 will be used.
-
•
Baseline Model 3: Experimentation insights from Experiment-2 will be used (in Study 2 only).
-
•
Champion Model: Experimentation insights from the periods with known experimentation will be ingested as Bayesian priors. Specifically, for Study 1 the insights from Experiment-1 and Experiment-3 will be ingested as priors, while for Study 2 insights from Experiment-1 and Experiment-2 will be ingested as priors.
Scenario | Experiment | Baseline | Baseline | Baseline | Champion |
---|---|---|---|---|---|
Attribution | Model 1 | Model 2 | Model 3 | Model | |
Study 1 | 1889 | 1427 | 1556 | N/A | 1717 |
Study 2 | 1147 | 784 | 855 | 1311 | 1240 |
In Table 4, we provide for each study the attribution insights from various models (i.e., baseline as well as champion models) and the experiments serving as the unobserved truth. By leveraging the results, to some extent, the following hypotheses could be validated:
-
•
For Study 1, attribution accuracy for the gap period in between 2 periods covered with experimentation increases if insights from both adjacent tests are ingested, as attribution based on the champion model (i.e., 1717) is closest to experiment-based attribution (i.e., 1889).
-
•
For Study 2, attribution accuracy for the most recent period with no test coverage is the best when all available older tests are used for model calibration, as attribution from the champion model (i.e., 1240) is closest to experiment-based attribution (i.e., 1147). It also demonstrates that it is helpful to leverage older test in addition to the most recent one, as attribution based on the champion model (i.e., 1240) is closer to experiment-based attribution (i.e., 1147) than baseline model 3 (i.e., 1311).
In summary, ingestion of multiple experiments lead to significant improvement in attribution accuracy. Most notably, the validation results indicate that: contrary to conventional wisdom in marketing measurement, using only the latest experiment for measurement model calibration does not lead to superior attribution accuracy for future periods. Older experimentation, albeit more temporally distant from the attribution period of interest in the future, can still add value. BTVC enables modelers to seamlessly combine insights from new as well as old experiments, producing attribution insights with better accuracy.
5. Architecture
5.1. Implementation
We implemented the BTVC model as a feature branch in our open sourced package Orbit (Ng et al., 2021) by Uber. Orbit is a software package aiming to simplify time series inferences and forecasting with structural Bayesian time series models for real-world cases and research. It provides a familiar and intuitive initialize-fit-predict interface for time series tasks, while utilizing the probabilistic programming languages such as Stan (Carpenter et al., 2017) and Pyro (Bingham et al., 2019) under the hood.
5.2. Deployment
The BTVC deployment system is customized by leveraging Michelangelo (Hermann and Balso, 2017), a machine learning (ML) platform developed by Uber. Michelangelo provides centralized workflow management for end-to-end modeling process.With the help of Michelangelo, BTVC deployment system is able to automate data preprocessing, model training, validations, predictions, and measurement monitoring at scale.
The deployment workflow summarized in Figure 6 consists of three main components:
-
•
Hyperparameter management layer: this is to store and manage the various supplement data needed for BTVC model training such as normalization scalar, adstock (Jin et al., 2017), lift test based priors, as well as model specific hyperparameters.
-
•
Orchestration layer: this is to upload and trigger the model training job.
-
•
Model container: a docker container including all the essential modeling code to be integrated with Michelangelo’s ecosystem.
6. Conclusion
In this paper, we propose a Bayesian Time Varying Coefficient (BTVC) model in particular developed for MMM applications at Uber. By assuming the local latent variables follow certain probabilistic distributions, a kernel-based smoothing technique is applied to produce the dynamic coefficients. This modeling framework entails a comprehensive solution for the challenges faced by traditional MMM. More importantly, it enables marketers to leverage multiple experimentation results in an intuitive yet scientific way. Simulations and real case benchmark studies demonstrate BTVC’s superiority in prediction accuracy and flexibility in experimentation ingestion. We also present the model deployment system, which can serve model training and predictions in real time without human oversight or intervention in a scalable way.
7. Acknowledgments
The authors would like to thank Sharon Shen, Qin Chen, Ruyi Ding, Vincent Pham, and Ariel Jiang for their help on this project, Dirk Beyer, and Kim Larsen for their comments on this paper.
References
- (1)
- Bingham et al. (2019) Eli Bingham, Jonathan P Chen, Martin Jankowiak, Fritz Obermeyer, Neeraj Pradhan, Theofanis Karaletsos, Rohit Singh, Paul Szerlip, Paul Horsfall, and Noah D Goodman. 2019. Pyro: Deep universal probabilistic programming. The Journal of Machine Learning Research 20, 1 (2019), 973–978.
- Blake et al. (2015) Thomas Blake, Chris Nosko, and Steven Tadelis. 2015. Consumer heterogeneity and paid search effectiveness: A large-scale field experiment. Econometrica 83, 1 (2015), 155–174.
- Carpenter et al. (2017) Bob Carpenter, Andrew Gelman, Matthew D Hoffman, Daniel Lee, Ben Goodrich, Michael Betancourt, Marcus Brubaker, Jiqiang Guo, Peter Li, and Allen Riddell. 2017. Stan: A probabilistic programming language. Journal of statistical software 76, 1 (2017).
- Casas and Fernandez-Casal (2021) Isabel Casas and Ruben Fernandez-Casal. 2021. tvReg: Time-Varying Coefficients Linear Regression for Single and Multi-Equations. https://CRAN.R-project.org/package=tvReg R package version 0.5.4.
- Chan and Perry (2017) David Chan and Mike Perry. 2017. Challenges and opportunities in media mix modeling. (2017).
- De Livera et al. (2011) Alysha M De Livera, Rob J Hyndman, and Ralph D Snyder. 2011. Forecasting time series with complex seasonal patterns using exponential smoothing. Journal of the American statistical association 106, 496 (2011), 1513–1527.
- Durbin and Koopman (2012) James Durbin and Siem Jan Koopman. 2012. Time series analysis by state space methods. Oxford university press.
- Fan and Zhang (2008) Jianqing Fan and Wenyang Zhang. 2008. Statistical methods with varying coefficient models. Statistics and its Interface 1, 1 (2008), 179.
- Farris et al. (2015) Paul W Farris, Dominique M Hanssens, James D Lenskold, and David J Reibstein. 2015. Marketing return on investment: Seeking clarity for concept and measurement. Applied Marketing Analytics 1, 3 (2015), 267–282.
- Gelman et al. (2013) Andrew Gelman, John B Carlin, Hal S Stern, David B Dunson, Aki Vehtari, and Donald B Rubin. 2013. Bayesian data analysis. CRC press.
- Han and Gabor (2020) Benjamin Han and Jared Gabor. 2020. Contextual Bandits for Advertising Budget Allocation. https://www.adkdd.org/Papers/Contextual-Bandits-for-Advertising-Budget-Allocation/2020
- Hastie and Tibshirani (1990) Trevor J Hastie and Robert J Tibshirani. 1990. Generalized additive models. Vol. 43. CRC press.
- Hermann and Balso (2017) Jeremy Hermann and Mike Del Balso. 2017. Meet Michelangelo: Uber’s machine learning platform. https://eng.uber.com/michelangelo/.
- Hoffman et al. (2013) Matthew D Hoffman, David M Blei, Chong Wang, and John Paisley. 2013. Stochastic variational inference. Journal of Machine Learning Research 14, 5 (2013).
- Imbens and Rubin (2015) Guido W Imbens and Donald B Rubin. 2015. Causal inference in statistics, social, and biomedical sciences. Cambridge University Press.
- Jin et al. (2017) Yuxue Jin, Yueqing Wang, Yunting Sun, David Chan, and Jim Koehler. 2017. Bayesian methods for media mix modeling with carryover and shape effects. (2017).
- Larsen (2018) Kim Larsen. 2018. Data Science Can’t Replace Human Marketers Just Yet – Here’s Why. https://blog.usejournal.com/data-science-cant-replace-human-marketers-just-yet-here-s-why-d5cdd355d85c
- Lewis and Rao (2015) Randall A Lewis and Justin M Rao. 2015. The unfavorable economics of measuring the returns to advertising. The Quarterly Journal of Economics 130, 4 (2015), 1941–1973.
- Lewis and Wong (2018) Randall A Lewis and Jeffrey Wong. 2018. Incrementality bidding & attribution. Available at SSRN 3129350 (2018).
- Li et al. (2002) Qi Li, Cliff J Huang, Dong Li, and Tsu-Tan Fu. 2002. Semiparametric smooth coefficient models. Journal of Business & Economic Statistics 20, 3 (2002), 412–422.
- McCarthy (1978) E Jerome McCarthy. 1978. Basic marketing: a managerial approach. RD Irwin.
- Ng et al. (2021) Edwin Ng, Zhishi Wang, Huigang Chen, Steve Yang, and Slawek Smyl. 2021. Orbit: Probabilistic Forecast with Exponential Smoothing. arXiv:2004.08492 [stat.CO]
- Rawlings et al. (2001) John O Rawlings, Sastry G Pantula, and David A Dickey. 2001. Applied regression analysis: a research tool. Springer Science & Business Media.
- Scott and Varian (2014) Steven L Scott and Hal R Varian. 2014. Predicting the present with bayesian structural time series. International Journal of Mathematical Modelling and Numerical Optimisation 5, 1-2 (2014), 4–23.
- Seabold and Perktold (2010) Skipper Seabold and Josef Perktold. 2010. statsmodels: Econometric and statistical modeling with python. In 9th Python in Science Conference. Package version 0.11.1.
- Sun et al. (2017) Yunting Sun, Yueqing Wang, Yuxue Jin, David Chan, and Jim Koehler. 2017. Geo-level Bayesian hierarchical media mix modeling. (2017).
- Taylor and Letham (2018) Sean J Taylor and Benjamin Letham. 2018. Forecasting at scale. The American Statistician 72, 1 (2018), 37–45. Package version 0.7.1.
- Vaver and Koehler (2011) Jon Vaver and Jim Koehler. 2011. Measuring ad effectiveness using geo experiments. (2011).
- West and Harrison (2006) Mike West and Jeff Harrison. 2006. Bayesian forecasting and dynamic models. Springer Science & Business Media.
- Wu and Chiang (2000) Colin O Wu and Chin-Tsang Chiang. 2000. Kernel smoothing on varying coefficient models with longitudinal dependent variable. Statistica Sinica (2000), 433–456.