# A guide to dynamic pricing algorithms

Price setting is one of the most important problems in retail because any price setting error directly results in lost profit. However, traditional price management methods almost never achieve optimal pricing because they are designed for traditional environments, where the frequency of price changes is inherently limited (e.g., brick-and-mortar stores), and the complexity of pricing models is constrained by the capabilities of off-the-shelf tools and manual processes.

Dynamic pricing algorithms help to increase the quality of pricing decisions in e-commerce environments by leveraging the ability to change prices frequently and collect the feedback data in real time. These capabilities enable a company to respond to demand changes more efficiently, reduce forecasting errors, and automate price management for catalogs with hundreds of millions of items.

This article is a deep dive into dynamic pricing algorithms that use reinforcement learning and Bayesian inference ideas, and were tested at scale by companies like Walmart and Groupon. We focus on the engineering aspects through code snippets and numerical examples; the theoretical details can be found in the referenced articles.

## Overview

Traditional price optimization requires knowing or estimating the dependency between the price and demand. Assuming that this dependency is known (at least at a certain time interval), the revenue-optimal price can be found by employing the following equation:

$$

p^* = \underset{p}{\text{argmax}}\ \ p \times d(p)

$$

where $p$ is the price and $d(p)$ is a demand function. This basic model can be further extended to incorporate item costs, cross-item demand cannibalization, competitor prices, promotions, inventory constraints and many other factors. The traditional price management process assumes that the demand function is estimated from the historical sales data, that is, by doing some sort of regression analysis for observed pairs of prices and corresponding demands $(p_i, d_i)$. Since the price-demand relationship changes over time, the traditional process typically re-estimates the demand function on a regular basis. This leads to some sort of dynamic pricing algorithm that can be summarized as follows:

- Collect historical data on different price points offered in the past as well as the observed demands for these points.
- Estimate the demand function.
- Solve the optimization problem similar to the problem defined above to find the optimal price that maximizes a metric like revenue or profit, and meets the constraints imposed by the pricing policy or inventory.
- Apply this optimal price for a certain time period, observe the realized demand, and repeat the above process.

The fundamental limitation of this approach is that it passively learns the demand function without actively exploring the dependency between the price and demand. This may or may not be a problem depending on how dynamic the environment is:

- If the product life cycle is relatively long and the demand function changes relatively slowly, the passive learning approach combined with organic price changes can be efficient, as the price it sets will be close to the true optimal price most of the time.
- If the product life cycle is relatively short or the demand function changes rapidly, the difference between the price produced by the algorithm and the true optimal price can become significant, and so will the lost revenue. In practice, this difference is substantial for many online retailers, and critical for retailers and sellers that extensively rely on short-time offers or flash sales (Groupon, Rue La La, etc.).

The second case represents a classical exploration-exploitation problem: in a dynamic environment, it is important to minimize the time spent on testing different price levels and collecting the corresponding demand points to accurately estimate the demand curve, and maximize the time used to sell at the optimal price calculated based on the estimate. Consequently, we want to design a solution that optimizes this trade-off, and also supports constraints that are common in real-life environments. More specifically, let's focus on the following design goals:

- Optimize the exploration-exploitation trade-off given that the seller does not know the demand function in advance (for example, the product is new and there is no historical data on it). This trade-off can be quantified as the difference between the actual revenue and the hypothetically possible revenue given that the demand function is known.
- Provide the ability to limit the number of price changes during the product life cycle. Although the frequency of price changes in digital channels is virtually unlimited, many sellers impose certain limitations to avoid inconsistent customer experiences and other issues.
- Provide the ability to specify valid price levels and price combinations. Most retailers restrict themselves to a certain set of price points (e.g., $25.90, $29.90, ..., $55.90), and the optimization process has to support this constraint.
- Enable the optimization of prices under inventory constraints, or given dependencies between products.

In the remainder of this article, we discuss several techniques that help to achieve the above design goals, starting with the simplest ones and gradually increasing the complexity of the scenarios.

## Scenario 1: Limited price experimentation

We first consider a scenario where the demand remains constant during the product life cycle, but the number of price changes is limited by the seller’s pricing policy. This scenario is often a valid approximation of flash sales or time-limited deals. For instance, a variant of the algorithm described below was tested at Groupon with very positive results. ^{[1]}

Assuming that the total duration of the product life cycle $T$ is known to the seller in advance, the goal is to sequentially optimize prices for $m$ time intervals, and also optimize the durations $\tau_i$ of these intervals:

In an extreme case, only one price change is allowed — a seller starts with an initial price guess, collects the demand data during the first period of time (exploration), computes the optimized price, and sells at this new price during the second time period that ends with the end of the product life cycle (exploitation).

It can be shown that in these settings, the optimal durations of the price intervals have to be exponentially increasing, so that a seller starts with short intervals to explore and learn, and gradually increases the intervals until the final price is set for the last and the longest interval, which is focused purely on exploitation:

$$\tau_i = \alpha \log^{(m-i)}T$$ where $\log^{(n)}x$ stands for $n$ iterations of the logarithm, $\log(\log(...\log x))$, and $\alpha$ is a coefficient that depends on the demand distribution. For practical purposes, $\alpha$ can be chosen empirically because the parameters of the demand may not be known. This layout is illustrated in the figure below:

Next, we need to specify how the prices are generated for each time interval. One simple but flexible approach is to generate a set of parametric demand functions (hypotheses) in advance, pick the hypothesis that most closely corresponds to the observed demand at the end of each time interval, and optimize the price for the next interval based on this hypothesis. In practice, the set of hypotheses can be generated based on the historical demand functions for similar products or categories (we just need to generate a reasonably dense “grid” of demand curves that covers the range where the true demand function is likely to be located).

The complete algorithm can be summarized as follows:

**Input**: Number of hypothesis $k$, number of time intervals $m$

- Generate a set of $k$ demand functions $d_1, \ldots, d_k$
- Compute the optimal price for each demand function, so the set of optimal prices is $p_1^ *, \ldots, p_k^ *$
- Pick random $p_i^* $ as the initial price $p_1$
- For each interval $1 \le i \le m-1$:
- Offer price $\ p_i\ $ for $\ \alpha \log^{m-i}(T)\ $ time units
- Observe the average demand per time unit $D_i$
- Find $d_j$ that minimizes $|d_j(p_i) - D_i|$
- Pick $p_j^* $ as the next price $p_{i+1}$

- Offer price $p_m$ until the end of the product life cycle

#### Implementation of the iterative hypothesis testing

Next, let's implement the above algorithm and run a simulation. We use a linear demand model to generate the hypotheses (and it is a reasonable choice for many practical applications as well), but any other parametric demand model, such as the constant-elasticity model, can also be used.

For a linear model, the revenue-optimal price can be calculated by taking a derivative of the revenue with respect to price, and equating it to zero: $$

\begin{aligned}

d(p) &= b + a\cdot p \\

p_{\text{opt}} &:\ \frac{\delta}{\delta p}\ p\cdot d(p) = 0 \\

p_{\text{opt}} &= -\frac{b}{2a}

\end{aligned} $$ This logic can be implemented as follows:

We use this code to generate a sample set of demand functions and the corresponding optimal prices:

For the runtime portion of the algorithm, we generate the price interval schedule in advance, and use it to determine whether or not we need to generate a new price at every time step (as we mentioned earlier, the schedule depends on the properties of the demand distribution, which is unknown to the seller, so the fixed schedule is a heuristic approximation):

## Click to expand the code sample (36 lines)

The execution of this algorithm is illustrated in the animation below. The top chart shows the true demand function as the dotted line, the realized demands at each time step as red crosses (sampled from the true demand function with additive noise), and the black lines as the selected hypotheses. The bottom plot shows the price and demand for every time step, with the price intervals highlighted with different bar colors.

## Scenario 2: Continuous experimentation under pricing rules

The algorithm described in the previous section is a simple yet efficient solution for settings where the demand function can be assumed to be stationary. In more dynamic settings, we need to use more generic tools that can continuously explore the environment, while also balancing the exploration-exploitation trade-off. Fortunately, reinforcement learning theory offers a wide range of methods designed specifically for this problem.

In this section, we will discuss a very flexible framework for dynamic pricing that uses reinforcement learning ideas and can be customized to support an extensive range of use cases and constraints.^{[2]} A variant of this framework was tested by Walmart with positive results.^{[3]}

Let's start with an observation that the approach used in the previous section can be improved in the following two areas:

- First, we can expect to build a more flexible and efficient framework by utilizing Bayesian methods for demand estimation. Using a Bayesian approach will enable us to accurately update the demand distribution model with every observed sample, as well as quantify the uncertainty in the model parameter estimates.
- Second, we should replace the fixed price change schedule with continuous exploration. Again, a Bayesian approach can help to better control the exploration process, as the time allocated for exploration and the breadth of exploration can be derived from the uncertainty of the demand estimates.

These two ideas are combined together in Thompson sampling, a generic method that belongs to a large and well-researched family of algorithms for the multi-armed bandit problem (a particular formulation of the exploration-exploitation problem). First, let's review a generic description of the Thompson sampling algorithm for demand estimation, and then refine it with more details:

**Initialization**:

- Specify the demand distribution $p(d\ |\ \theta)$ conditioned on some parameter
- Specify the prior distribution of the demand model parameters $p(\theta)$

**Execution**:

- For each time step $t$:
- Sample the demand parameters $\theta_t \sim p(\theta)$
- Find the optimal price for the sampled demand parameters: $$p^* = \underset{p}{\text{argmin}}\ \ p \times \mathbb{E}[d(p;\ \theta_t)]$$
- Offer the optimal price and observe the demand
- Update the posterior distribution with the observed price-demand pair $$ p(\theta) \leftarrow p(\theta) \times p(d\ |\ \theta) $$

The main idea of Thompson sampling is to control the amount of exploration by sampling the model parameters for a probabilistic distribution that is refined over time. If the variance of the distribution is high, we will tend to explore a wider range of possible demand functions. If the variance is low, we will mostly use functions that are close to what we think is the most likely demand curve (that is, the curve defined by the mean of the distribution), and explore more distant shapes just occasionally. Note that the demand distribution incorporates both the dependency between the price and demand (which can be comprised of deterministic and random components), as we illustrate in the next paragraph.

To make the above algorithm concrete, we need to specify a probabilistic model for the demand. One possible way to accomplish this task is to use a linear, constant-elasticity or some other continuous model that treats the slope coefficient or elasticity coefficient as a random parameter $\theta$. Another way is to use a model with a discrete set of price levels. The latter approach is preferable in many environments because many companies, especially retailers, have a pricing policy that prescribes a certain set of price levels (e.g., $5.90, $6.90, etc.). The demand model in this case represents a table with $k$ price levels, and each price level is associated with its own demand probability density function (PDF) specified by some parameters, so that the overall demand curve can be visualized by plotting the price levels and their mean demands:

Thus, the curve can have an arbitrary shape and can approximate a wide range of price-demand dependencies, including linear and constant-elasticity models.

Next, we need to specify the demand distributions for individual price levels. For illustrative purposes, we assume that there is no correlation between prices. In this case, parameter $\theta$ can simply be the mean demand at the corresponding price level. In other words, the demand distribution model conditioned on $\theta$ becomes trivial because the shape of the demand curve is already captured point by point, and we can simply sample the mean demand at each point to optimize the price. Let us assume that the observed demand samples have a Poisson distribution (a natural choice because each sample represents the number of purchases per unit of time):

$$

d_1, \ldots, d_n \sim \text{poisson}(\theta)

$$

The prior $\theta$ distribution can be chosen to be gamma because it is conjugate to the Poisson distribution:

$$

p(\theta)=\text{gamma}(\alpha, \beta) = \frac{\beta^\alpha}{\Gamma(\alpha)} \theta^{\alpha-1} e^{-\beta\theta}

$$

The likelihood given the observed samples for a certain price is:

$$

p(d\ |\ \theta) = \prod_{i=1}^n \frac{e^{-\theta} \theta^{d_i}}{d_i!} = \frac{e^{-n\theta} \theta^{\sum_i d_i}}{\prod_i d_i!}

$$

Finally, the update rule for the posterior distribution of the parameter $\theta$ is obtained as a product of the prior and likelihood:

$$

p(\theta) \leftarrow p(\theta)\cdot p(d\ |\ \theta) =\text{gamma}(\alpha + \sum d_i,\ \beta+n)

$$

In words, we update the prior distribution at a certain price point by adding the number of times this price was offered to hyperparameter $\beta$, and the total demand observed during these times to the hyperparameter $\alpha$.

Given the above assumptions, we can rewrite the Thompson sampling algorithm as follows:

- Prior distribution $p(\theta)=\text{gamma}(\alpha, \beta)$
- For each time step $t$:
- Sample the mean demand from $d \sim p(\theta)$
- Find the optimal price: $$ p^* = \underset{p}{\text{argmin}}\ \ p \times d $$
- Offer the optimal price and observe the demand $d_t$
- Update the posterior distribution: $$ \begin{aligned} \alpha &\leftarrow \alpha + d_t \\\\ \beta &\leftarrow \beta + 1 \end{aligned} $$

This version of the algorithm is detailed enough to handle more dynamic pricing, and can be implemented straightforwardly.

#### Implementation of Thompson sampling for dynamic pricing

An example of a dynamic pricing implementation with Thompson sampling is shown in the code snippet below. This snippet includes both the algorithm and the parts needed to run a simulation.

## Click to expand the code sample (40 lines)

By running this implementation and recording how the parameters of the distributions are changing over time, we can observe how the algorithm explores and learns the demand function:

In the beginning, the demand parameters are the same for all price levels. The algorithm actively explores different prices (the red line in the bottom chart), becomes certain that the price of $3.99 provides the best revenue (the yellow curve in the middle chart), and starts to choose it most of the time, exploring other options only occasionally.

We conclude this section with a note that Thompson sampling is not the only choice for dynamic price optimization; there are a wide range of alternative algorithms that can be used in practice, and generic off-the-shelf implementations of such algorithms are readily available.^{[4]} Many of these algorithms are designed for advanced formulations of multi-armed bandit problems, such as contextual bandit problems, and can improve their performance by using additional pieces of information, such as customer profile data. The basic Thompson sampler can also be extended in many ways (see, for example, ^{[5]} for a detailed treatment). In particular, we can dramatically increase the flexibility of demand modeling using Markov Chain Monte Carlo (MCMC) methods, as we will discuss later in this article.

## Scenario 3: Multiple products or time intervals

The framework described in the previous section is a flexible tool that can be extended to support various constraints and features. For example, one can add inventory constraints to the routine that finds optimal prices to exclude the options where the demand exceeds the available inventory. The framework can also be extended to estimate demands and optimize prices for multiple products, and optimization typically remains straightforward until dependencies between products or time intervals appear (the optimization problem can be solved separately for each product). Such dependencies can make the optimization problem much more challenging. In this section, we discuss the scenarios with dependencies between products or time intervals, and the optimization methods that can help to handle such use cases.

Consider a scenario where a seller offers multiple products in some category or group, so that the products are fully or partly substitutable. In the general case, the demand function for each product depends on all individual prices of other products that can be challenging to accurately estimate and optimize, especially in the dynamic pricing settings. One possible simplification is to use a demand function that depends not on the individual prices of other products, but on the average price within a group of substitutable products.^{[6]} This can be an accurate approximation in many settings, because the ratio between a product’s own price and the average price in the group reflects the competitiveness of the product and quantifies demand cannibalization. In this case, we can assume a demand model that estimates not just one demand value for each product-price pair, but multiple values for each possible average price (the set of possible average prices is finite because the set of valid price levels is discrete). This assumption leads to the following optimization problem:

$$

\begin{aligned}

\max \ \ & \sum_k \sum_i p_k \cdot d_{ik} \cdot x_{ik} \\

\text{subject to} \ \ & \sum_k x_{ik} = 1, \quad \text{for all } i \\

&\sum_k \sum_i p_k \cdot x_{ik} = c \\

&x_{ik} \in \{0,1\}

\end{aligned}

$$

where $d_{ik}$ is the demand for product $i$, given that it is assigned price $k$, and $x_{ij}$ is a binary dummy variable that is equal to one if price $k$ is assigned to product $i$, and zero otherwise. The first constraint ensures that each product has only one price, and the second constraint ensures that all prices sum up to some value $c$: that is, the average price is fixed. In solving this problem for each possible value of $c$ and picking the best result, we obtain the set of variables $x$ that defines the revenue-optimal assignment of prices to products.

The problem defined above is an integer programming problem, because the decision variables $x$ are either ones or zeros. It can be computationally intractable to solve this problem, even for medium size categories, especially if prices need to be updated frequently. We can work around this problem by replacing the original integer programming problem with a linear programming problem where variables $x$ are assumed to be continuous:

$$

\begin{aligned}

\max \ \ & \sum_k \sum_i p_k \cdot d_{ik} \cdot x_{ik} \\

\text{subject to} \ \ & \sum_k x_{ik} = 1, \quad \text{for all } i \\

&\sum_k \sum_i p_k \cdot x_{ik} = c \\

&0 \le x_{ik} \le 1

\end{aligned}

$$

This technique is known as linear relaxation. The resulting linear program can be solved efficiently, even if the number of products and possible average prices is high. It can be shown that the solution of the linear program gives a good linear bound for the optimal solution of the integer program.^{[6:1]} This boundary can be used to reduce the set of price sums $c$ for which the integer problem needs to be solved. In practice, the number of integer programs that need to be solved can be reduced very sharply (e.g., from hundreds to less than ten).

Another approach is to set prices directly based on the solution of the linear program. In this case, each product can have more than one non-zero variables $x$, and the operational model needs to be adjusted to account for this. For example, a time interval for which one price is offered can be divided into multiple sub-intervals in proportion, specified by variables $x$. For instance, if there are two non-zero elements equal to 0.2 and 0.8, then the corresponding prices can be offered for 20% and 80% of the time, respectively.

The approach above using integer programming or linear relaxation can be applied to a range of scenarios, including the following:

- Price optimization for multiple products that have inventory dependencies. For example, a manufacturer can assemble different products from parts drawn from one or several shared pools of resources. In this case, the optimization problem will have a constraint that the total number of parts needed to assemble all products must not exceed the corresponding level of in-stock inventory.
- Price optimization for multiple time intervals. Consider the case of a seasonal product that is purchased by a retailer at the beginning of the season and has to be sold out by the end of the season. In this case, we might be interested not only in forecasting the demand and optimizing the price for the next time interval, but in estimating the demand functions for all time intervals until the end of the season and optimizing prices under the constraint that the sum of the demands for all intervals needs to converge to the available inventory (i.e., the product needs to be sold out or the unsold units will be lost). The optimization problem for one product can then be defined as follows:

$$

\begin{aligned}

\max \ \ & \sum_t \sum_k p_k \cdot d_{it} \cdot x_{it} \\

\text{subject to} \ \ & \sum_k x_{tk} = 1, \quad \text{for all } t \\

&\sum_t \sum_k p_k \cdot x_{tk} = c \\

&x_{tk} \in \{0,1\}

\end{aligned}

$$

where $t$ iterates over time intervals within the season.

#### Implementation of linear relaxation for price optimization

For illustrative purposes, we will implement the solver for the linear relaxation problem with multiple products, as described in the previous section. The solver uses a standard routine for linear programming from the SciPy library that requires the input problem to be defined in the following vector form:

$$

\begin{aligned}

\max \ \ & \mathbf{r} \cdot \mathbf{x} \\

\text{subject to} \ \ & \mathbf{A}\cdot \mathbf{x} = \mathbf{b}

\end{aligned}

$$

We use the following design of the inputs to impose constraints on the sum of the prices and price weights for each product:

In others words, the cost vector $r$ consists of revenues for all possible price assignments, and each row of matrix $A$ ensures that the price weights sum to 1 for any given product, except the last row that ensures that all prices sum to the required level $c$.

The following code snippet shows the actual implementation and an example test run:

## Click to expand the code sample (38 lines)

The algorithm produces a vector of the price weights for each product that can be used to reduce the number of integer programs that need to be solved, or set the prices directly, as described in the previous section.

This solver can be straightforwardly adapted to other cases, such as shared pools of resources or multiple time intervals. Such solvers can then be plugged into any dynamic pricing algorithm described in this article, including the iterative offline learning and Thompson sampling algorithms.

## Scenario 4: Complex demand models

Although the demand models used in practice are often simple (linear or constant-elasticity), the development of probabilistic models for Thompson sampling and other similar algorithms can be complicated. Even in our simple implementation of the Thompson sampling algorithm that uses a standard Poisson-Gamma model, we had to do some math and manually implement updated rules for the distribution parameters. This process can be even more complicated if we need to use multivariate distributions for dependent products, or need to customize the model based on business requirements and constraints. We can work around this issue by using probabilistic programming frameworks that allow us to specify models in a declarative manner and abstract the inference procedure. Under the hood, these frameworks use generic MCMC methods to infer the model parameters.

The probabilistic programming approach can be illustrated with a couple of examples that utilize the PyMC3 framework. To start, let us re-implement the Poisson-Gamma model used in Scenario 2 to draw the demand samples:

In the code snippet above, we just declare that the mean demand (theta) has a prior gamma distribution and that the observed demand samples have a Poisson distribution, and point the model to an array with all demand samples observed since selling began. After that, we just draw ten thousand samples from the model, and plot the histogram that follows the posterior distribution of the mean demand:

This implementation can be plugged directly into the Thompson sampler — we associate each price level with an instance of the above model, and draw one sample from each of these models at every time step to solve the price optimization problem. This is a striking simplification compared to the manual updates of the posterior distribution parameters we implemented in the Scenario 2 section.

We can use the flexibility of this approach to sample the parameters needed for the Thompson sampler from more complex demand models, both discrete and continuous. As a second example, consider a constant-elasticity model defined as follows:

$$

d(p) = b\cdot p^a

$$

We first rewrite this model in the additive (logarithmic) form for the sake of computational stability and ease of modification:^{[7]}

$$

\log d(p) = \log b + a \log p

$$

The implementation of this model with PyMC3 is straightforward (although we omit some details, like data centering, for the sake of simplicity):

We can now sample the parameters of the constant-elasticity model, and visualize multiple realizations of the demand function as follows:

This approach can help to build and test even more complex demand models. It can be particularly useful for multiple related products with correlated demand functions. In this case, the correlated parameters of different demands (e.g., elasticity coefficients) can be drawn from a single multivariate distribution, and probabilistic programming frameworks can significantly help to specify and infer such hierarchical models.

## Conclusions

This article describes several algorithms and techniques that address different aspects of dynamic pricing — experimentation and active learning, optimization with and without pricing policy constraints, and demand modeling. These methods together constitute a comprehensive toolkit that can be used to build dynamic pricing systems and customize them based on business requirements and needs.

In practice, dynamic pricing techniques may have a major impact on sales volume and revenue. It is not unusual to see revenue uplift in the range of 10 to 20 percent, and sales volume uplift as high as 80 to 200 percent depending on the product category and business model.^{[1:1]}^{[2:1]} This is the reason that many market leaders, including Amazon and Walmart, extensively research and utilize dynamic pricing, which, in turn, has heavily influenced the retail market as a whole, driving the frequency of price changes up over the last decade.^{[8]} This makes retail price management increasingly more challenging, and has made algorithmic price management methods, including dynamic pricing, become an increasingly important source of competitive advantage.

## References

W. Cheung, D. Simchi-Levi, and H. Wang, Dynamic Pricing and Demand Learning with Limited Price Experimentation, February 2017 ↩︎ ↩︎

K. Ferreira, D. Simchi-Levi, and H. Wang, Online Network Revenue Management Using Thompson Sampling, November 2017 ↩︎ ↩︎

R. Ganti, M. Sustik, T. Quoc, B. Seaman, Thompson Sampling for Dynamic Pricing, February 2018 ↩︎

D. Russo, B. Roy, A. Kazerouni, I. Osband, Z. Wen, A Tutorial on Thompson Sampling, November 2017 ↩︎

K. J. Ferreira, B. Lee, and D. Simchi-Levi, Analytics for an Online Retailer: Demand Forecasting and Price Optimization, November 2015 ↩︎ ↩︎

C. Scherrer, Bayesian Optimal Pricing, May 2018 ↩︎

A. Cavallo, More Amazon Effects: Online Competition and Pricing Behaviors, September 2018 ↩︎

Data ScienceMachine Learning and Artificial Intelligence

## Leave us a comment, we would love to know what you think