<img align="left" width="120" height="120" src="https://raw.githubusercontent.com/google/lightweight_mmm/main/images/lightweight_mmm_logo_colored_250.png" alt="lightweight_mmm_logo_colored_250"></img>
# Lightweight (Bayesian) Marketing Mix Modeling
### LMMM is a python library that helps organisations understand and optimise marketing spend across media channels.
##### This is not an official Google product.
[![PyPI](https://img.shields.io/pypi/v/lightweight_mmm?logo=pypi&logoColor=white&style=flat-square)](https://pypi.org/project/lightweight_mmm/)
[![GitHub Workflow CI](https://img.shields.io/github/actions/workflow/status/google/lightweight_mmm/ci.yaml?branch=main&label=pytest&logo=python&logoColor=white&style=flat-square)](https://github.com/google/lightweight_mmm/actions/workflows/ci.yaml?branch=main)
[![Read the Docs](https://img.shields.io/readthedocs/lightweight_mmm?label=ReadTheDocs&logo=readthedocs&logoColor=white&style=flat-square)](https://lightweight-mmm.readthedocs.io/en/latest/)
[![Downloads](https://pepy.tech/badge/lightweight-mmm)](https://pepy.tech/project/lightweight-mmm)
[Docs](https://lightweight-mmm.readthedocs.io/en/latest/) •
[Introduction](#introduction) •
[Theory](#theory) •
[Getting Started](#getting-started) •
[References](#references) •
[Community Spotlight](#community-spotlight)
## Introduction
[Marketing Mix Modeling (MMM)](https://en.wikipedia.org/wiki/Marketing_mix_modeling) is used by advertisers to measure advertising effectiveness and inform budget allocation decisions across media channels. Measurement based on aggregated data allows comparison across online and offline channels in addition to being unaffected by recent ecosystem changes (some related to privacy) which may affect attribution modelling. MMM allows you to:
- Estimate the optimal budget allocation across media channels.
- Understand how media channels perform with a change in spend.
- Investigate effects on your target KPI (such as sales) by media channel.
Taking a Bayesian approach to MMM allows an advertiser to integrate prior information into modelling, allowing you to:
- Utilise information from industry experience or previous media mix models using Bayesian priors.
- Report on both parameter and model uncertainty and propagate it to your budget optimisation.
- Construct hierarchical models, with generally tighter credible intervals, using breakout dimensions such as geography.
The LightweightMMM package (built using [Numpyro](https://github.com/pyro-ppl/numpyro) and [JAX](https://github.com/google/jax)) helps advertisers easily build Bayesian MMM models by providing the functionality to appropriately scale data, evaluate models, optimise budget allocations and plot common graphs used in the field.
## Theory
### Simplified Model Overview
An MMM quantifies the relationship between media channel activity and sales, while controlling for other factors. A simplified model overview is shown below and the full model is set out in the [model documentation](https://lightweight-mmm.readthedocs.io/en/latest/models.html). An MMM is typically run using weekly level observations (e.g. the KPI could be sales per week), however, it can also be run at the daily level.
$$kpi = \alpha + trend + seasonality + media\ channels + other\ factors$$
Where *kpi* is typically the volume or value of sales per time period, $\alpha$ is the model intercept, $trend$ is a flexible non-linear function that captures trends in the data, $seasonality$ is a sinusoidal function with configurable parameters that flexibly captures seasonal trends, $media\ channels$ is a matrix of different media channel activity (typically impressions or costs per time period) which receives transformations depending on the model used (see Media Saturation and Lagging section) and $other\ factors$ is a matrix of other factors that could influence sales.
### Standard and Hierarchical models
The LightweightMMM can either be run using data aggregated at the national level (standard approach) or using data aggregated at a geo level (sub-national hierarchical approach).
1. **National level (standard approach).** This approach is appropriate if the data available is only aggregated at the national level (e.g. The KPI could be national sales per time period). This is the most common format used in MMMs.
2. **Geo level (sub-national hierarchical approach).** This approach is appropriate if the data can be aggregated at a sub-national level (e.g. the KPI could be sales per time period for each state within a country). This approach can yield more accurate results compared to the standard approach because it uses more data points to fit the model. We recommend using a sub-national level model for larger countries such as the US if possible.
### Media Saturation and Lagging
It is likely that the effect of a media channel on sales could have a lagged effect which tapers off slowly over time. Our powerful Bayesian MMM model architecture is designed to capture this effect and offers three different approaches. We recommend users compare all three approaches and use the approach that works the best. The approach that works the best will typically be the one which has the best out-of-sample fit (which is one of the generated outputs). The functional forms of these three approaches are briefly described below and are fully expressed in our [model documentation](https://lightweight-mmm.readthedocs.io/en/latest/models.html).
- [Adstock](https://en.wikipedia.org/wiki/Advertising_adstock): Applies an infinite lag that decreases its weight as time passes.
- [Hill-Adstock](https://en.wikipedia.org/wiki/Hill_equation_(biochemistry)): Applies a sigmoid like function for diminishing returns to the output of the adstock function.
- [Carryover](https://static.googleusercontent.com/media/research.google.com/en//pubs/archive/46001.pdf): Applies a [causal convolution](https://paperswithcode.com/method/causal-convolution) giving more weight to the near values than distant ones.
### Flow chart
<img src="https://raw.githubusercontent.com/google/lightweight_mmm/main/images/flowchart.png" alt="flow_chart"></img>
## Getting started
### Installation
The recommended way of installing lightweight_mmm is through PyPi:
```bash
pip install --upgrade pip
pip install lightweight_mmm
```
If you want to use the most recent and slightly less stable version you can install it from github:
`pip install --upgrade git+https://github.com/google/lightweight_mmm.git`
**If you are using Google Colab, make sure you restart the runtime after installing.**
### Preparing the data
Here we use simulated data but it is assumed you have your data cleaned at this
point. The necessary data will be:
- Media data: Containing the metric per channel and time span (eg. impressions
per time period). Media values must not contain negative values.
- Extra features: Any other features that one might want to add to the analysis.
These features need to be known ahead of time for optimization or you would need
another model to estimate them.
- Target: Target KPI for the model to predict. For example, revenue amount, number of app installs. This will also be the metric
optimized during the optimization phase.
- Costs: The total cost per media unit per channel.
```python
# Let's assume we have the following datasets with the following shapes (we use
the `simulate_dummy_data` function in utils for this example):
media_data, extra_features, target, costs = utils.simulate_dummy_data(
data_size=160,
n_media_channels=3,
n_extra_features=2,
geos=5) # Or geos=1 for national model
```
Scaling is a bit of an art, Bayesian techniques work well if the input data is
small scale. We should not center variables at 0. Sales and media should have a
lower bound of 0.
1. `y` can be scaled as `y / jnp.mean(y)`.
2. `media` can be scaled as `X_m / jnp.mean(X_m, axis=0)`, which means the new column mean will be 1.
We provide a `CustomScaler` which can apply multiplications and division scaling
in case the wider used scalers don't fit your use case. Scale your data
accordingly before fitting the model.
Below is an example of usage of this `CustomScaler`:
```python
# Simple split of the data based on time.
split_point = data_size - data_size // 10
media_data_train = media_data[:split_point, :]
target_train = target[:split_point]
extra_features_train = extra_features[:split_point, :]
extra_features_test = extra_features[split_point:, :]
# Scale data
media_scaler = preprocessing.CustomScaler(divide_operation=jnp.mean)
extra_features_scaler = preprocessing.CustomScaler(divide_operation=jnp.mean)
target_scaler = preprocessing.CustomScaler(
divide_operation=jnp.mean)
# scale cost up by N since fit() will divide it by number of time periods
cost_scaler = preprocessing.CustomScaler(divide_operation=jnp.mean)
media_data_train = media_scaler.fit_transform(media_data_train)
extra_features_train = extra_features_scaler.fit_transform(
extra_features_train)
target_train = target_scaler.fit_transform(target_train)
costs = cost_scaler.fit_transform(unscaled_costs)
```
In case you have a variable that has a lot of 0s you can also scale by the mean
of non zero values. For instance you can use a lambda function to do this:
`lambda x: jnp.mean(x[x > 0])`. The same applies for cost scaling.
### Training the model
The model requires the media data, the extra features, the costs of each media
unit per channel and the target. You can also pass how many samples you would
like to use as well as the number of chains.
For running multiple chains in parallel the user would need to set
`numpyro.set_host_device_count` to either the number of chains or the number of
CPUs available.
See an example below:
```python
# Fit model.
mmm = lightweight_mmm.LightweightMMM()
mmm.fit(media=media_data,
extra_features=extra_features,
media_prior=costs,
target=target,
number_warmup=1000,
number_samples=1000,
number_chains=2)
```
If you want to change any prior in the model (besides the media prior which you
are already specifying always), you can do so with `custom_priors`:
```python
# See detailed explanation on custom priors in our documentation.
custom_priors = {"intercept": numpyro.distributions.Uniform(1, 5)}
# Fit model.
mmm = lightweight_mmm.LightweightMMM()
mmm.fit(media=media_data,
extra_features=extra_features,
media_prior=costs,
target=target,
number_warmup=1000,
number_samples=1000,
number_chains=2,
custom_priors=custom_priors)
```
Please refer to our [documentation on custom_priors](https://lightweight-mmm.readthedocs.io/en/latest/custom_priors.html)
for more details.
You can switch between daily and weekly data by enabling
`weekday_seasonality=True` and `seasonality_frequency=365` or
`weekday_seasonality=False` and `seasonality_frequency=52` (default). In case
of daily data we have two types of seasonality: discrete weekday and smooth
annual.
### Model diagnostics
#### Convergence Check
Users can check convergence metrics of the parameters as follows:
```python
mmm.print_summary()
```
The rule of thumb is that `r_hat` values for all parameters are less than 1.1.
#### Fitting check
Users can check fitting between true KPI and predicted KPI by:
```python
plot.plot_model_fit(media_mix_model=mmm, target_scaler=target_scaler)
```
If `target_scaler` used for `preprocessing.CustomScaler()` is given, the target
would be unscaled. Bayesian R-squared and MAPE are shown in the chart.
#### Predictive check
Users can get the prediction for the test data by:
```python
prediction = mmm.predict(
media=media_data_test,
extra_features=extra_data_test,
target_scaler=target_scaler
)
```
Returned prediction are distributions; if point estimates are desired, users
can calculate those based on the given distribution. For example, if `data_size`
of the test data is 20, `number_samples` is 1000 and `number_of_chains` is 2,
`mmm.predict` returns 2000 sets of predictions with 20 data points. Users can
compare the distributions with the true value of the test data and calculate
the metrics such as mean and median.
#### Parameter estimation check
Users can get detail of the parameter estimation by:
```python
mmm.print_summary()
```
The above returns the mean, standard deviation, median and the credible interval
for each parameter. The distribution charts are provided by:
```python
plot.plot_media_channel_posteriors(media_mix_model=mmm, channel_names=media_names)
```
`channel_names` specifies media names in each chart.
#### Media insights
Response curves are provided as follows:
```python
plot.plot_response_curves(media_mix_model=mmm, media_scaler=media_scaler, target_scaler=target_scaler)
```
If `media_scaler` and `target_scaler` used for `preprocessing.CustomScaler()` are given, both the media and target values would be unscaled.
To extract the media effectiveness and ROI estimation, users can do the following:
```python
media_effect_hat, roi_hat = mmm.get_posterior_metrics()
```
`media_effect_hat` is the media effectiveness estimation and `roi_hat` is the ROI estimation. Then users can visualize the distribution of the estimation as follows:
```python
plot.plot_bars_media_metrics(metric=media_effect_hat, channel_names=media_names)
```
```python
plot.plot_bars_media_metrics(metric=roi_hat, channel_names=media_names)
```
### Running the optimization
For optimization we will maximize the sales changing the media inputs such that
the summed cost of the media is constant. We can also allow reasonable bounds
on each media input (eg +- x%). We only optimise across channels and not over
time. For running the optimization one needs the following main parameters:
- `n_time_periods`: The number of time periods you want to simulate (eg. Optimize
for the next 10 weeks if you trained a model on weekly data).
- The model that was trained.
- The `budget` you want to allocate for the next `n_time_periods`.
- The extra features used for training for the following `n_time_periods`.
- Price per media unit per channel.
- `media_gap` refers to the media data gap between the end of training data and
the start of the out of sample media given. Eg. if 100 weeks of data were used
for training and prediction starts 2 months after training data finished we
need to provide the 8 weeks missing between the training data and the
prediction data so data transformations (adstock, carryover, ...) can take
place correctly.
See below and example of optimization:
```python
# Run media optimization.
budget = 40 # your budget here
prices = np.array([0.1, 0.11, 0.12])
extra_features_test = extra_features_scaler.transform(extra_features_test)
solution = optimize_media.find_optimal_budgets(
n_time_periods=extra_features_test.shape[0],
media_mix_model=mmm,
budget=budget,
extra_features=extra_features_test,
prices=prices)
```
### Save and load the model
Users can save and load the model as follows:
```python
utils.save_model(mmm, file_path='file_path')
```
Users can specify `file_path` to save the model.
To load a saved MMM model:
```python
utils.load_model(file_path: 'file_path')
```
## Citing LightweightMMM
To cite this repository:
```
@software{lightweight_mmmgithub,
author = {Pablo Duque and Dirk Nachbar and Yuka Abe and Christiane Ahlheim and Mike Anderson and Yan Sun and Omri Goldstein and Tim Eck},
title = {LightweightMMM: Lightweight (Bayesian) Marketing Mix Modeling},
url = {https://github.com/google/lightweight_mmm},
version = {0.1.6},
year = {2022},
}
```
## References
- [Jin, Y., Wang, Y., Sun, Y., Chan, D., & Koehler, J. (2017). Bayesian Methods for Media Mix Modeling with Carryover and Shape Effects. Google Inc.](https://research.google/pubs/pub46001/)
- [Chan, D., & Perry, M. (2017). Challenges and Opportunities in Media Mix Modeling.](https://research.google/pubs/pub45998/)
- [Sun, Y., Wang, Y., Jin, Y., Chan, D., & Koehler, J. (2017). Geo-level Bayesian Hierarchical Media Mix Modeling.](https://research.google/pubs/pub46000/)
## Support
As LMMM is not an official Google product, the LMMM team can only offer limited support.
For **questions about methodology**, please refer to the References section or to the [FAQ page](https://lightweight-mmm.readthedocs.io/en/latest/faq.html).
For **issues installing or using LMMM**, feel free to post them in the Discussions or Issues tabs of the Github repository. The LMMM team responds to these questions in our free time, so we unfortunately cannot guarantee a timely response. We also encourage the community to share tips and advice with each other here!
For **feature requests**, please post them to the Discussions tab of the Github repository. We have an internal roadmap for LMMM development but do pay attention to feature requests and appreciate them!
For **bug reports**, please post them to the Issues tab of the Github repository. If/when we are able to address them, we will let you know in the comments to your issue.
**Pull requests** are appreciated but are very difficult for us to merge since the code in this repository is linked to Google internal systems and has to pass internal review. If you submit a pull request and we have resources to help merge it, we will reach out to you about this!
## Community Spotlight
- [How To Create A Marketing Mix Model With LightweightMMM](https://forecastegy.com/posts/how-to-create-a-marketing-mix-model-with-lightweightmmm/) by Mario Filho.
- [How Google LightweightMMM Works](https://getrecast.com/google-lightweightmmm/) and [A walkthrough of Google’s LightweightMMM](https://www.youtube.com/watch?v=oRYN0V6sPlM) by Mike Taylor.
Raw data
{
"_id": null,
"home_page": "https://github.com/google/lightweight_mmm",
"name": "lightweight-mmm",
"maintainer": "",
"docs_url": null,
"requires_python": "",
"maintainer_email": "",
"keywords": "",
"author": "Google LLC",
"author_email": "no-reply@google.com",
"download_url": "https://files.pythonhosted.org/packages/a4/f8/e62ca95469c338a8715b9bcde7b92af2f687e47900bf9f4a571a577a8a7a/lightweight_mmm-0.1.9.tar.gz",
"platform": null,
"description": "<img align=\"left\" width=\"120\" height=\"120\" src=\"https://raw.githubusercontent.com/google/lightweight_mmm/main/images/lightweight_mmm_logo_colored_250.png\" alt=\"lightweight_mmm_logo_colored_250\"></img>\n\n# Lightweight (Bayesian) Marketing Mix Modeling\n### LMMM is a python library that helps organisations understand and optimise marketing spend across media channels.\n##### This is not an official Google product.\n\n[![PyPI](https://img.shields.io/pypi/v/lightweight_mmm?logo=pypi&logoColor=white&style=flat-square)](https://pypi.org/project/lightweight_mmm/)\n[![GitHub Workflow CI](https://img.shields.io/github/actions/workflow/status/google/lightweight_mmm/ci.yaml?branch=main&label=pytest&logo=python&logoColor=white&style=flat-square)](https://github.com/google/lightweight_mmm/actions/workflows/ci.yaml?branch=main)\n[![Read the Docs](https://img.shields.io/readthedocs/lightweight_mmm?label=ReadTheDocs&logo=readthedocs&logoColor=white&style=flat-square)](https://lightweight-mmm.readthedocs.io/en/latest/)\n[![Downloads](https://pepy.tech/badge/lightweight-mmm)](https://pepy.tech/project/lightweight-mmm)\n\n[Docs](https://lightweight-mmm.readthedocs.io/en/latest/) \u2022\n[Introduction](#introduction) \u2022\n[Theory](#theory) \u2022\n[Getting Started](#getting-started) \u2022\n[References](#references) \u2022\n[Community Spotlight](#community-spotlight)\n\n## Introduction\n\n[Marketing Mix Modeling (MMM)](https://en.wikipedia.org/wiki/Marketing_mix_modeling) is used by advertisers to measure advertising effectiveness and inform budget allocation decisions across media channels. Measurement based on aggregated data allows comparison across online and offline channels in addition to being unaffected by recent ecosystem changes (some related to privacy) which may affect attribution modelling. MMM allows you to:\n\n- Estimate the optimal budget allocation across media channels.\n- Understand how media channels perform with a change in spend.\n- Investigate effects on your target KPI (such as sales) by media channel.\n\nTaking a Bayesian approach to MMM allows an advertiser to integrate prior information into modelling, allowing you to:\n\n- Utilise information from industry experience or previous media mix models using Bayesian priors.\n- Report on both parameter and model uncertainty and propagate it to your budget optimisation.\n- Construct hierarchical models, with generally tighter credible intervals, using breakout dimensions such as geography.\n\nThe LightweightMMM package (built using [Numpyro](https://github.com/pyro-ppl/numpyro) and [JAX](https://github.com/google/jax)) helps advertisers easily build Bayesian MMM models by providing the functionality to appropriately scale data, evaluate models, optimise budget allocations and plot common graphs used in the field.\n\n## Theory\n\n### Simplified Model Overview\n\nAn MMM quantifies the relationship between media channel activity and sales, while controlling for other factors. A simplified model overview is shown below and the full model is set out in the [model documentation](https://lightweight-mmm.readthedocs.io/en/latest/models.html). An MMM is typically run using weekly level observations (e.g. the KPI could be sales per week), however, it can also be run at the daily level.\n\n$$kpi = \\alpha + trend + seasonality + media\\ channels + other\\ factors$$\n\nWhere *kpi* is typically the volume or value of sales per time period, $\\alpha$ is the model intercept, $trend$ is a flexible non-linear function that captures trends in the data, $seasonality$ is a sinusoidal function with configurable parameters that flexibly captures seasonal trends, $media\\ channels$ is a matrix of different media channel activity (typically impressions or costs per time period) which receives transformations depending on the model used (see Media Saturation and Lagging section) and $other\\ factors$ is a matrix of other factors that could influence sales.\n\n### Standard and Hierarchical models\n\nThe LightweightMMM can either be run using data aggregated at the national level (standard approach) or using data aggregated at a geo level (sub-national hierarchical approach).\n\n1. **National level (standard approach).** This approach is appropriate if the data available is only aggregated at the national level (e.g. The KPI could be national sales per time period). This is the most common format used in MMMs.\n\n2. **Geo level (sub-national hierarchical approach).** This approach is appropriate if the data can be aggregated at a sub-national level (e.g. the KPI could be sales per time period for each state within a country). This approach can yield more accurate results compared to the standard approach because it uses more data points to fit the model. We recommend using a sub-national level model for larger countries such as the US if possible.\n\n### Media Saturation and Lagging\n\nIt is likely that the effect of a media channel on sales could have a lagged effect which tapers off slowly over time. Our powerful Bayesian MMM model architecture is designed to capture this effect and offers three different approaches. We recommend users compare all three approaches and use the approach that works the best. The approach that works the best will typically be the one which has the best out-of-sample fit (which is one of the generated outputs). The functional forms of these three approaches are briefly described below and are fully expressed in our [model documentation](https://lightweight-mmm.readthedocs.io/en/latest/models.html).\n\n- [Adstock](https://en.wikipedia.org/wiki/Advertising_adstock): Applies an infinite lag that decreases its weight as time passes.\n- [Hill-Adstock](https://en.wikipedia.org/wiki/Hill_equation_(biochemistry)): Applies a sigmoid like function for diminishing returns to the output of the adstock function.\n- [Carryover](https://static.googleusercontent.com/media/research.google.com/en//pubs/archive/46001.pdf): Applies a [causal convolution](https://paperswithcode.com/method/causal-convolution) giving more weight to the near values than distant ones.\n\n### Flow chart\n\n<img src=\"https://raw.githubusercontent.com/google/lightweight_mmm/main/images/flowchart.png\" alt=\"flow_chart\"></img>\n\n\n## Getting started\n\n### Installation\n\nThe recommended way of installing lightweight_mmm is through PyPi:\n\n```bash\npip install --upgrade pip\npip install lightweight_mmm\n```\n\nIf you want to use the most recent and slightly less stable version you can install it from github:\n\n`pip install --upgrade git+https://github.com/google/lightweight_mmm.git`\n\n**If you are using Google Colab, make sure you restart the runtime after installing.**\n\n### Preparing the data\nHere we use simulated data but it is assumed you have your data cleaned at this\npoint. The necessary data will be:\n\n- Media data: Containing the metric per channel and time span (eg. impressions\n per time period). Media values must not contain negative values.\n- Extra features: Any other features that one might want to add to the analysis.\n These features need to be known ahead of time for optimization or you would need\n another model to estimate them.\n- Target: Target KPI for the model to predict. For example, revenue amount, number of app installs. This will also be the metric\n optimized during the optimization phase.\n- Costs: The total cost per media unit per channel.\n\n```python\n# Let's assume we have the following datasets with the following shapes (we use\nthe `simulate_dummy_data` function in utils for this example):\nmedia_data, extra_features, target, costs = utils.simulate_dummy_data(\n data_size=160,\n n_media_channels=3,\n n_extra_features=2,\n geos=5) # Or geos=1 for national model\n```\nScaling is a bit of an art, Bayesian techniques work well if the input data is\nsmall scale. We should not center variables at 0. Sales and media should have a\nlower bound of 0.\n\n1. `y` can be scaled as `y / jnp.mean(y)`.\n2. `media` can be scaled as `X_m / jnp.mean(X_m, axis=0)`, which means the new column mean will be 1.\n\nWe provide a `CustomScaler` which can apply multiplications and division scaling\nin case the wider used scalers don't fit your use case. Scale your data\naccordingly before fitting the model.\nBelow is an example of usage of this `CustomScaler`:\n\n```python\n# Simple split of the data based on time.\nsplit_point = data_size - data_size // 10\nmedia_data_train = media_data[:split_point, :]\ntarget_train = target[:split_point]\nextra_features_train = extra_features[:split_point, :]\nextra_features_test = extra_features[split_point:, :]\n\n# Scale data\nmedia_scaler = preprocessing.CustomScaler(divide_operation=jnp.mean)\nextra_features_scaler = preprocessing.CustomScaler(divide_operation=jnp.mean)\ntarget_scaler = preprocessing.CustomScaler(\n divide_operation=jnp.mean)\n# scale cost up by N since fit() will divide it by number of time periods\ncost_scaler = preprocessing.CustomScaler(divide_operation=jnp.mean)\n\nmedia_data_train = media_scaler.fit_transform(media_data_train)\nextra_features_train = extra_features_scaler.fit_transform(\n extra_features_train)\ntarget_train = target_scaler.fit_transform(target_train)\ncosts = cost_scaler.fit_transform(unscaled_costs)\n```\n\nIn case you have a variable that has a lot of 0s you can also scale by the mean\nof non zero values. For instance you can use a lambda function to do this:\n`lambda x: jnp.mean(x[x > 0])`. The same applies for cost scaling.\n\n### Training the model\nThe model requires the media data, the extra features, the costs of each media\nunit per channel and the target. You can also pass how many samples you would\nlike to use as well as the number of chains.\n\nFor running multiple chains in parallel the user would need to set\n`numpyro.set_host_device_count` to either the number of chains or the number of\nCPUs available.\n\nSee an example below:\n\n```python\n# Fit model.\nmmm = lightweight_mmm.LightweightMMM()\nmmm.fit(media=media_data,\n extra_features=extra_features,\n media_prior=costs,\n target=target,\n number_warmup=1000,\n number_samples=1000,\n number_chains=2)\n```\n\nIf you want to change any prior in the model (besides the media prior which you\nare already specifying always), you can do so with `custom_priors`:\n\n```python\n# See detailed explanation on custom priors in our documentation.\ncustom_priors = {\"intercept\": numpyro.distributions.Uniform(1, 5)}\n\n# Fit model.\nmmm = lightweight_mmm.LightweightMMM()\nmmm.fit(media=media_data,\n extra_features=extra_features,\n media_prior=costs,\n target=target,\n number_warmup=1000,\n number_samples=1000,\n number_chains=2,\n custom_priors=custom_priors)\n```\n\nPlease refer to our [documentation on custom_priors](https://lightweight-mmm.readthedocs.io/en/latest/custom_priors.html)\nfor more details.\n\nYou can switch between daily and weekly data by enabling\n`weekday_seasonality=True` and `seasonality_frequency=365` or\n`weekday_seasonality=False` and `seasonality_frequency=52` (default). In case\nof daily data we have two types of seasonality: discrete weekday and smooth\nannual.\n\n### Model diagnostics\n#### Convergence Check\nUsers can check convergence metrics of the parameters as follows:\n\n```python\nmmm.print_summary()\n```\n\nThe rule of thumb is that `r_hat` values for all parameters are less than 1.1.\n\n#### Fitting check\nUsers can check fitting between true KPI and predicted KPI by:\n\n```python\nplot.plot_model_fit(media_mix_model=mmm, target_scaler=target_scaler)\n```\n\nIf `target_scaler` used for `preprocessing.CustomScaler()` is given, the target\nwould be unscaled. Bayesian R-squared and MAPE are shown in the chart.\n\n#### Predictive check\nUsers can get the prediction for the test data by:\n\n```python\nprediction = mmm.predict(\n media=media_data_test,\n extra_features=extra_data_test,\n target_scaler=target_scaler\n)\n```\n\nReturned prediction are distributions; if point estimates are desired, users\ncan calculate those based on the given distribution. For example, if `data_size`\nof the test data is 20, `number_samples` is 1000 and `number_of_chains` is 2,\n`mmm.predict` returns 2000 sets of predictions with 20 data points. Users can\ncompare the distributions with the true value of the test data and calculate\nthe metrics such as mean and median.\n\n#### Parameter estimation check\nUsers can get detail of the parameter estimation by:\n\n```python\nmmm.print_summary()\n```\n\nThe above returns the mean, standard deviation, median and the credible interval\nfor each parameter. The distribution charts are provided by:\n\n```python\nplot.plot_media_channel_posteriors(media_mix_model=mmm, channel_names=media_names)\n```\n\n`channel_names` specifies media names in each chart.\n\n#### Media insights\nResponse curves are provided as follows:\n\n```python\nplot.plot_response_curves(media_mix_model=mmm, media_scaler=media_scaler, target_scaler=target_scaler)\n```\n\nIf `media_scaler` and `target_scaler` used for `preprocessing.CustomScaler()` are given, both the media and target values would be unscaled.\n\nTo extract the media effectiveness and ROI estimation, users can do the following:\n\n```python\nmedia_effect_hat, roi_hat = mmm.get_posterior_metrics()\n```\n\n`media_effect_hat` is the media effectiveness estimation and `roi_hat` is the ROI estimation. Then users can visualize the distribution of the estimation as follows:\n\n```python\nplot.plot_bars_media_metrics(metric=media_effect_hat, channel_names=media_names)\n```\n\n```python\nplot.plot_bars_media_metrics(metric=roi_hat, channel_names=media_names)\n```\n\n### Running the optimization\n\nFor optimization we will maximize the sales changing the media inputs such that\nthe summed cost of the media is constant. We can also allow reasonable bounds\non each media input (eg +- x%). We only optimise across channels and not over\ntime. For running the optimization one needs the following main parameters:\n\n- `n_time_periods`: The number of time periods you want to simulate (eg. Optimize\n for the next 10 weeks if you trained a model on weekly data).\n- The model that was trained.\n- The `budget` you want to allocate for the next `n_time_periods`.\n- The extra features used for training for the following `n_time_periods`.\n- Price per media unit per channel.\n- `media_gap` refers to the media data gap between the end of training data and\n the start of the out of sample media given. Eg. if 100 weeks of data were used\n for training and prediction starts 2 months after training data finished we\n need to provide the 8 weeks missing between the training data and the\n prediction data so data transformations (adstock, carryover, ...) can take\n place correctly.\n\nSee below and example of optimization:\n\n```python\n# Run media optimization.\nbudget = 40 # your budget here\nprices = np.array([0.1, 0.11, 0.12])\nextra_features_test = extra_features_scaler.transform(extra_features_test)\nsolution = optimize_media.find_optimal_budgets(\n n_time_periods=extra_features_test.shape[0],\n media_mix_model=mmm,\n budget=budget,\n extra_features=extra_features_test,\n prices=prices)\n```\n\n### Save and load the model\nUsers can save and load the model as follows:\n\n```python\nutils.save_model(mmm, file_path='file_path')\n```\n\nUsers can specify `file_path` to save the model.\nTo load a saved MMM model:\n\n```python\nutils.load_model(file_path: 'file_path')\n```\n\n## Citing LightweightMMM\n\nTo cite this repository:\n\n```\n@software{lightweight_mmmgithub,\n author = {Pablo Duque and Dirk Nachbar and Yuka Abe and Christiane Ahlheim and Mike Anderson and Yan Sun and Omri Goldstein and Tim Eck},\n title = {LightweightMMM: Lightweight (Bayesian) Marketing Mix Modeling},\n url = {https://github.com/google/lightweight_mmm},\n version = {0.1.6},\n year = {2022},\n}\n```\n\n\n## References\n\n- [Jin, Y., Wang, Y., Sun, Y., Chan, D., & Koehler, J. (2017). Bayesian Methods for Media Mix Modeling with Carryover and Shape Effects. Google Inc.](https://research.google/pubs/pub46001/)\n\n- [Chan, D., & Perry, M. (2017). Challenges and Opportunities in Media Mix Modeling.](https://research.google/pubs/pub45998/)\n\n- [Sun, Y., Wang, Y., Jin, Y., Chan, D., & Koehler, J. (2017). Geo-level Bayesian Hierarchical Media Mix Modeling.](https://research.google/pubs/pub46000/)\n\n## Support\n\nAs LMMM is not an official Google product, the LMMM team can only offer limited support. \n\nFor **questions about methodology**, please refer to the References section or to the [FAQ page](https://lightweight-mmm.readthedocs.io/en/latest/faq.html).\n\nFor **issues installing or using LMMM**, feel free to post them in the Discussions or Issues tabs of the Github repository. The LMMM team responds to these questions in our free time, so we unfortunately cannot guarantee a timely response. We also encourage the community to share tips and advice with each other here!\n\nFor **feature requests**, please post them to the Discussions tab of the Github repository. We have an internal roadmap for LMMM development but do pay attention to feature requests and appreciate them!\n\nFor **bug reports**, please post them to the Issues tab of the Github repository. If/when we are able to address them, we will let you know in the comments to your issue.\n\n**Pull requests** are appreciated but are very difficult for us to merge since the code in this repository is linked to Google internal systems and has to pass internal review. If you submit a pull request and we have resources to help merge it, we will reach out to you about this!\n\n## Community Spotlight\n\n- [How To Create A Marketing Mix Model With LightweightMMM](https://forecastegy.com/posts/how-to-create-a-marketing-mix-model-with-lightweightmmm/) by Mario Filho.\n\n- [How Google LightweightMMM Works](https://getrecast.com/google-lightweightmmm/) and [A walkthrough of Google\u2019s LightweightMMM](https://www.youtube.com/watch?v=oRYN0V6sPlM) by Mike Taylor.\n",
"bugtrack_url": null,
"license": "Apache 2.0",
"summary": "Package for Media-Mix-Modelling",
"version": "0.1.9",
"project_urls": {
"Homepage": "https://github.com/google/lightweight_mmm"
},
"split_keywords": [],
"urls": [
{
"comment_text": "",
"digests": {
"blake2b_256": "e94f4de01b34a2b22303a0b694f98457632ea689767b5e79d0657898de8c2b37",
"md5": "d667ef920163b11fdde71aee6341e36e",
"sha256": "cbf1dba6774710d54137334c76fa94b6a654b12476a79e30d904a97fe00a9899"
},
"downloads": -1,
"filename": "lightweight_mmm-0.1.9-py3-none-any.whl",
"has_sig": false,
"md5_digest": "d667ef920163b11fdde71aee6341e36e",
"packagetype": "bdist_wheel",
"python_version": "py3",
"requires_python": null,
"size": 105221,
"upload_time": "2023-05-23T21:24:17",
"upload_time_iso_8601": "2023-05-23T21:24:17.438617Z",
"url": "https://files.pythonhosted.org/packages/e9/4f/4de01b34a2b22303a0b694f98457632ea689767b5e79d0657898de8c2b37/lightweight_mmm-0.1.9-py3-none-any.whl",
"yanked": false,
"yanked_reason": null
},
{
"comment_text": "",
"digests": {
"blake2b_256": "a4f8e62ca95469c338a8715b9bcde7b92af2f687e47900bf9f4a571a577a8a7a",
"md5": "1707def11bb15e4e9867089510fb7fd7",
"sha256": "3f6562476f08dc30e10bdee569784782edaf79724ee34b64a525ce55e2fbf01c"
},
"downloads": -1,
"filename": "lightweight_mmm-0.1.9.tar.gz",
"has_sig": false,
"md5_digest": "1707def11bb15e4e9867089510fb7fd7",
"packagetype": "sdist",
"python_version": "source",
"requires_python": null,
"size": 86102,
"upload_time": "2023-05-23T21:24:19",
"upload_time_iso_8601": "2023-05-23T21:24:19.678199Z",
"url": "https://files.pythonhosted.org/packages/a4/f8/e62ca95469c338a8715b9bcde7b92af2f687e47900bf9f4a571a577a8a7a/lightweight_mmm-0.1.9.tar.gz",
"yanked": false,
"yanked_reason": null
}
],
"upload_time": "2023-05-23 21:24:19",
"github": true,
"gitlab": false,
"bitbucket": false,
"codeberg": false,
"github_user": "google",
"github_project": "lightweight_mmm",
"travis_ci": false,
"coveralls": false,
"github_actions": true,
"lcname": "lightweight-mmm"
}