Hostname: page-component-784d4fb959-xb4ml Total loading time: 0 Render date: 2025-07-15T08:41:18.894Z Has data issue: false hasContentIssue false

Algorithmic price recommendations and collusion: Experimental evidence

Published online by Cambridge University Press:  07 July 2025

Matthias Hunold
Affiliation:
Faculty III, Universität Siegen, Unteres Schloß 3, Siegen, NW, Germany
Tobias Werner*
Affiliation:
Center for Humans and Machines, Max Planck Institute for Human Development, Lentzeallee 94, Berlin, BE, Germany
*
Corresponding author: Tobias Werner; Email: tobias.felix.werner@gmail.com
Rights & Permissions [Opens in a new window]

Abstract

This paper investigates the collusive and competitive effects of algorithmic price recommendations on market outcomes. These recommendations are often non-binding and common in many markets. We develop a theoretical framework and derive two algorithms that recommend collusive pricing strategies. Utilizing a laboratory experiment, we find that sellers condition their prices on the recommendation of the algorithms. The algorithm with a soft punishment strategy lowers market prices and has a pro-competitive effect. The algorithm that recommends a subgame perfect equilibrium strategy increases the range of market outcomes, including more collusive ones.

Information

Type
Original Paper
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2025. Published by Cambridge University Press on behalf of Economic Science Association.

1. Introduction

We explore the potential of non-binding algorithmic price recommendations to facilitate tacit collusion among three sellers in a Bertrand game. The algorithms recommend collusive pricing strategies and vary in the severity of punishment phases after deviations from the recommended price. We focus on whether these recommendations can coordinate pricing behavior despite being non-binding and allowing for deviations.

There are various examples of non-binding price recommendations by a common intermediary in practice. For instance, various platforms provide sellers on their marketplaces with an algorithmic recommendation on how to set the price for the products they sell based on historical market dataFootnote 1. While it can increase market efficiency, platforms can have incentives to foster tacit collusion in the downstream market, depending on their monetization scheme, as they usually receive a share of the revenues from the seller marketFootnote 2. Besides sales platforms, there exists a market for third-party providers of price recommendations, where the same developer provides the same recommendation algorithm to multiple competing firms. Those developers can have an incentive to make the recommendations from the algorithm less competitive as it increases the profits of the supplied firms and thereby increases the demand for the algorithm itself (Harrington, Reference Harrington2022).

Competition authorities are concerned that price recommendation algorithms by a common intermediary can dampen competition as it might make coordination between sellers easier (Competition and Markets Authority, 2021, Bundeskartellamt & Autorité de la concurrence, 2019). For example, according to reporters of ProPublica, price recommendation software allegedly led to coordination effects in the U.S. rental market, especially in regions where a few property managers control a large share of the apartmentsFootnote 3. Moreover, recent investigations by the Federal Trade Commission and the U.S. Department of Justice suggest that hotels in New Jersey were using the same third-party price recommendation tools to coordinate their behavior (Federal Trade Commission and U.S. Department of Justice, 2024). The hotels argued that the recommendations were non-binding, but enforcers noted that even non-binding recommendations can facilitate collusion.

To investigate the effects of price recommendations on the prices of competing sellers, we derive two rule-based algorithms from economic theory and behavioral insightsFootnote 4. These algorithms aim at increasing sellers’ profits and recommend collusive strategies with different punishment mechanics. They provide the same price recommendation to all sellers in a market during a given round and adapt their recommendations across rounds based on the past pricing decisions of the sellers. The recommendations are non-binding and do not change the game’s action space and payoff functions, as they do not provide fundamentally new information. In practice, however, collusive outcomes can be challenging to achieve without communication among competitors or another form of coordination (see, for instance,Fonseca & Normann, Reference Fonseca and Normann2012).

The theory-based algorithm recommends actions consistent with a collusive trigger strategy and Nash reversion. If a seller undercuts the collusive price level, it recommends competitive prices for several periods to all sellers until it returns to recommending collusive prices. All firms following the recommendations constitutes a subgame perfect Nash equilibrium. Additionally, we consider an algorithm that is motivated by behavioral findings whereby firms often do not use harsh punishment strategies (see, for instance, Wright, Reference Wright2013). This algorithm recommends brief punishment phases with prices at the level of the deviating price and returns to recommending high prices when all sellers comply.

The subjects in our laboratory experiments resemble competing sellers who repeatedly set prices and receive price recommendations from an algorithm in each round. Across treatments, we vary whether participants receive a recommendation or not, as well as the type of recommendation algorithm. Subjects in the experiment are informed that the recommendation algorithm aims to symmetrically maximize sellers’ profits without favoring any particular seller.

The algorithmic price recommendations positively influence individual pricing decisions in the sense that higher recommended sales prices induce sellers to set higher individual prices. The estimated “pass-on rate” from recommended prices to sales prices is between 0.22 and 0.57, depending on the recommendation algorithm. The pass-on rate is higher for the theory-based algorithm recommending a collusive trigger strategy with temporary Nash reversal.

The effects on the realized market prices and profits differ sharply between the distinct recommendation algorithms. We find insightful price patterns for the theory-based algorithm even though the average market prices do not differ from the control treatment without any price recommendation. The substantial heterogeneities can explain the absence of an average treatment effect in market outcomes. The collusive effect of the algorithm depends on the seller’s characteristics. In markets where sellers have low levels of negative reciprocity, the recommendation algorithm decreases market prices. Thus, if participants are usually unwilling to punish unfair behavior, the recommendation lowers market prices.

The behaviourally motivated algorithm also recommends the monopoly price but differs in the reaction to the deviation of a seller. For this algorithm, we find lower market prices than without any recommendation. Participants repeatedly deviate downwards from the recommendation, which triggers a downward spiral that leads to lower prices. We find no evidence that this algorithm fosters collusion for any subgroup. This is particularly interesting against the backdrop of observations where humans prefer soft punishments for deviations from collusion in experiments (see, for instance, Wright, Reference Wright2013).

Related literature

Our article relates to the literature on the collusive effects of algorithmic pricing. There exists evidence that algorithms can foster collusion and lead to anti-competitive prices (Klein, Reference Klein2021, Calvano et al., Reference Calvano, Calzolari, Denicolo and Pastorello2020, Hansen et al., Reference Hansen, Misra and Pai2021, Brown & MacKay, Reference Brown and MacKay2023). Johnson et al. Reference Johnson, Rhodes and Wildenbeest(2023) focus on tacit collusion among self-learning algorithms on sales platforms and discuss how the platform’s design choices influence it. Ezrachi and Stucke Reference Ezrachi and Stucke2017 and Harrington Reference Harrington(2022) discuss the possibility that multiple competing firms use the same pricing algorithm and and how it can lead to seller coordination. Normann and Sternberg Reference Normann and Sternberg(2023) and Werner Reference Werner(2022) show experimentally that algorithms may raise market prices even above the price level usually observed in human markets. We differ from this approach as we consider algorithms that only give recommendations but do not compete with the sellers.

Our study also relates to the literature on recommended retail prices. These are pricing recommendations that a manufacturer provides to its retailers. In theory, they can act as a coordination device (Faber & Janssen, Reference Faber and Janssen2019, Buehler & Gärtner, Reference Buehler and Gärtner2013) and can make markets more collusive (Foros & Steen, Reference Foros and Steen2013, Gill & Thanassoulis, Reference Gill and Thanassoulis2016). Furthermore, they can also influence demand by setting a reference point for the consumers (Bruttel, Reference Bruttel2018), and manufacturers may use them to influence and guide the search process of consumers (see, for instance, Lubensky Reference Lubensky(2017) and Janssen and Reshidi Reference Janssen and Reshidi(2022)). Algorithmic price recommendations, while similar in providing common recommendations to competing firms, differ significantly. Unlike the relatively static recommended retail prices, which are often distributed in print format, algorithmic recommendations are highly dynamic and digitally distributed. Furthermore, they are not visible to consumers, influencing demand solely through the pricing decisions of sellers.

Recommendations and requests influence the decisions of participants in various experimental games. They can increase contributions to public goods, reduce tax evasion, and facilitate coordination in games with correlated equilibria (Silverman et al., Reference Silverman, Slemrod and Uler2014, Croson & Marks, Reference Croson and Marks2001, Cadsby, Maynes and Trivedi, Reference Cadsby, Maynes and Umashanker Trivedi2006, Duffy & Feltovich, Reference Duffy and Feltovich2010). Various papers also study the effect of price announcements on collusion. Those announcements can be understood as a recommendation of one participant to other market participants. While they can temporarily foster collusion, the effect usually fades as the game progresses (e.g.,Holt & Davis, Reference Holt and Davis1990, Harstad et al., Reference Harstad, Martin and Normann1998, Harrington et al., Reference Harrington and Kujal2016). Sonntag and Zizzo Reference Sonntag and John Zizzo(2015) show that authoritarian quantity recommendations can lower quantities in a Cournot market gameFootnote 5. Schotter and Sopher Reference Schotter and Sopher(2003) shows that intergenerational advice passed from one participant to the next in a sequence can help coordinate actions and improve outcomes across various games and experimental setups. Our approach is distinct as an algorithm provides dynamic recommendations instead of previous subjects.

The remainder of the article is structured as follows. Section 2 introduces the experimental design and discusses the rule-based algorithms we consider. Furthermore, we derive our hypotheses. In Section 3, we present the results. We discuss the implications of our results in Section 4 and provide avenues for future research. Online Appendix A contains the theoretical derivations underlying our algorithms. Online Appendix B contains the instructions and post-experimental questionnaire. We document various robustness checks and further algorithm variations in Online Appendix C. In Online Appendix D, we demonstrate that a monopoly platform can benefit from collusive price recommendations, highlighting that it can be a suitable example for collusive price recommendations in practice.

2. Experimental design

2.1. Market environment

To experimentally investigate the collusive effect of price recommendations, we consider a market setup similar to Dufwenberg and Gneezy Reference Dufwenberg and Gneezy(2000) and Fonseca and Normann Reference Fonseca and Normann(2012). There are n = 3 sellers, each represented by a participant. The market size is chosen such that tacit collusion is unlikely without any recommendation (Huck et al., Reference Huck, Normann and Oechssler2004)Footnote 6. All firms sell the same homogeneous good and have no capacity constraintsFootnote 7. The demand side consists of k = 30 computerized consumers with perfectly inelastic unit demand. Firms compete in an indefinitely repeated game with a discount rate of 95%. In each round of the game, all participants choose their prices from the set of integers in the range between the competitive Nash price of $p^N =1$ and the monopoly price of $p^M=10$ independently. The seller with the lowest price in a given period supplies the entire market. If multiple sellers have the lowest price, they share the market equally. There is no direct communication between the participants.

2.2. Price recommendations and treatments

We consider two additional treatments next to a Baseline treatment in which we do not provide any price recommendation to the participants. Across those treatments, we vary which type of algorithm provides the price recommendations. The algorithms are designed to foster tacit collusion. While they do not provide fundamentally new information to the seller, they could help coordinate market behavior. Each participant in a market receives a price recommendation at the beginning of each round. Crucially, this price recommendation is identical for all participants within the same market in a given round. After each participant selects a price, the participants receive information about the pricing decision of the other participants and their payoff in the given round. Furthermore, the recommended price is again shown to the respective treatment participants.

It is commonly understood that for collusion to be successful, firms have to anticipate that collusion takes place. The fact that an algorithm provides collusive price recommendations can support this anticipation. Moreover, the firms must obtain a common understanding of the collusive strategy. The collusive price is not necessarily the monopoly price of pM but could be any price above the competitive price pN. Additionally, collusion also depends on a shared understanding of how to punish deviations from the collusive price. It includes a punishment price but also an understanding of how many periods this price is set before, possibly, the firms return to a collusive price. Recommendations can act as a coordination device that addresses all these issues. The idea behind a recommendation algorithm is that sellers may expect other sellers to behave according to the recommendation. It makes it incentive-compatible to do the sameFootnote 8.

We focus on rule-based algorithms as they are highly tractable and allow us to derive clear, theoretically guided hypotheses. Furthermore, in digital markets, many algorithms are simple as well. Hanspach et al. (Reference Hanspach, Sapi and Wieting2024) and Musolff (Reference Musolff2022) show that real-world pricing algorithms are often rule-based and follow straightforward conditional processes. Moreover, although alternative methods like reinforcement learning algorithms have more complex routines to learn a pricing strategy, they eventually often converge to strategies that simple rules can describe (Werner, Reference Werner2022, Klein, Reference Klein2021, Kasberger et al., Reference Kasberger, Martin, Normann and Werner2023)Footnote 9. Hence, our focus on those algorithms is attractive from a methodological perspective and realistic regarding the tools used in actual markets.

In the following, we outline the recommendation algorithms we consider. We provide further details on the theoretical foundations of the algorithms and the different equilibria they induce in Online Appendix AFootnote 10.

2.2.1. Algorithm that recommends Nash equilibrium actions

The following algorithm, labeled RecTheory, recommends prices according to a trigger strategy. It relies on two state variables, “collusive” and “punitive”, which are determined by the sellers’ past pricing behavior. Based on these states, the algorithm recommends either the monopoly price (pM) or the stage game Nash equilibrium price (pN). It operates as follows:

Algorithm 1. (RecTheory)

  • At the beginning of period t = 1, initialize:

    • Set the state variable St to “collusive”.

    • Set the punishment counter τt to 0.

  • At the beginning of any period t > 1, check for a state transition:

    • If $ S_{t-1} = \text{''collusive''} $:

      • * If all firms set a price of pM in period t − 1, set St to “collusive”.

      • * Otherwise:

        • - Set St to “punitive”.

        • - Set the punishment counter to be equal to the punishment duration T ( $ \tau_t = T $).

    • If $ S_{t-1} = \text{''punitive''} $:

      • * Decrease the punishment counter by 1 ( $ \tau_t = \tau_{t-1} - 1 $).

      • * If the punishment counter equals 0 ( $ \tau_t = 0$), set St to “collusive”.

  • In each period t, provide a recommendation based on St:

    • If $ S_t = \text{''collusive''} $, recommend a price of pM.

    • If $ S_t = \text{''punitive''} $, recommend a price of pN.

Following our parameterization of the environment with $p^M=10$ and $p^{N}=1$, RecTheory recommends actions that constitute a subgame perfect Nash equilibrium for a punishment phase of $T\geq3$ (see Online Appendix A). We choose the minimal feasible punishment length of T = 3 for the experiments. In summary, RecTheory recommends the monopoly price to all firms. If all firms continue to play the monopoly price, the recommendation remains unchanged in the next period. If one or more firms deviate from this recommendation, the algorithm recommends punishing this deviation jointly for three periods and reverting to the monopoly price afterward.

2.2.2. Behaviourally motivated soft punishment algorithms

Empirical and experimental evidence indicates that punishment is often less harsh than in theory models with trigger or even grim-trigger strategies. For instance, Wright Reference Wright(2013) finds that only a small fraction of subjects in market experiments use optimal or grim punishment strategies. Most punishment strategies are softer and more gradual. It concerns both the punishment length, as well as by how much prices are reduced in a punishment phase. Similarly, Dal Bó and Fréchette Reference Bó, Pedro and Fréchette(2019) show that humans often use tit-for-tat strategies in the iterated prisoners’ dilemma, which is strategically similar to our stylized market environment.

To reflect these practices, we set up a behaviourally motivated recommendation algorithm. It works as follows:

Algorithm 2. (RecSoft)

  • At the beginning of period one, recommend a price of pM.

  • At the beginning of any period after period one:

    • If all sellers set the same price in the previous period, recommend the monopoly price of pM in the current period.

    • If not all sellers set the same price in the previous period, recommend a punitive price equal to the lowest price from the previous period (e.g., min(10,10,9)=9).

In view of the behavioral insights cited above, such a recommendation mechanism may be superior to the algorithm implementing a subgame perfect Nash equilibrium with trigger strategies. The recommendation is similar to a tit-for-tat algorithm as it mimics the firms’ decisions in the previous period. However, it also proactively tries to increase prices after all firms arrive at the same price level by recommending the monopoly price again to everyone.

2.3. Hypothesis

We test the following hypotheses in the experiment.

Hypothesis 1.

Recommendations positively influence individual prices. A higher recommended price leads to higher individual prices.

As the recommendation may act as a coordination device, we expect that firms factor it into their pricing decision, and we hypothesize that higher recommendations lead to higher individual prices. This is necessary for any sensible algorithm to have a collusive effect.

Hypothesis 2.

The RecTheory recommendation algorithm leads to higher market prices than the absence of a recommendation algorithm.

Hypothesis 2 expresses the rationale that RecTheory acts as a coordination device among the firms and thereby indeed facilitates collusion.

Similarly, but for different reasons, RecSoft can have a pro-collusive effect. Following the recommendation might be behaviourally attractive as no harsh punishment needs to be implemented compared to RecTheory. With k-level reasoning, for instance, a seller might rationalize that other sellers prefer to punish if it bears little cost and it yields an expected price soon after. Suppose sellers anticipate punishment under the current soft punishment algorithm. In that case, it may deter them from departing from the collusive priceFootnote 11. Furthermore, if sellers deviated in the past, the algorithm promotes cooperation as it again recommends the monopoly price once sellers agree on a joint price level. These arguments lead to

Hypothesis 3.

The RecSoft recommendation algorithm leads to higher market prices than the absence of a recommendation algorithm.

It is noteworthy that, in contrast to RecTheory, following the recommendations from RecSoft does not constitute a subgame perfect Nash equilibrium at the parameters used in the experiment. We show this formally in Online Appendix A. Following the soft recommendation algorithm may nevertheless be more attractive than the recommendation involving Nash reversion in punishment phases. It depends on the willingness of the sellers to implement drastic and longer-lasting punishments and their beliefs about the behavior of the other market participants. Nevertheless, sellers might find the soft punishment not harsh enough. Which recommendation algorithm performs better thus remains an ex-ante open question.

2.4. Procedure

The experiments were conducted between February 2020 and August 2021 in the University of Duesseldorf DICE Lab. We used ORSEE (Greiner, Reference Greiner2015) to recruit the subject for the experiments. The experiment was programmed in oTree (Chen et al., Reference Chen, Schonger and Wickens2016). We utilized a between-subject design, and each subject participated only once.

At the beginning of each experiment session, participants were randomly assigned to a computer in the lab and could read the instructions on the computer screen. Additionally, the participants received a printed version of the instructions. The instructions were the same for each subject. The original and the translated version are in Online Appendix B. After the subjects read the instructions, they answered several control questions to ensure they understood the setupFootnote 12. In case a participant failed to answer all control questions correctly, the software asked the participant to reread the instructions and allowed the participant to ask the experimenter clarifying questions in private.

In RecTheory and RecSoft, the instructions describe the objective of the algorithms to the participants. Specifically, we explain that the recommendation algorithm aims at increasing the sellers’ joint long-term profits. We also explain that the recommendations reflect this long-run objective rather than achieving the highest possible profit in any single round. One control question specifically assesses whether participants comprehend the design purpose of the algorithm. The answers are affirmative and confirm that the participants have the same understanding of the algorithm’s objective of maximizing the sellers’ joint profits.

Following the design of the algorithm described in Section 2.2, all firms in a market receive the same recommendation in a given round. We emphasize this clearly in the instructions to ensure participants understand this central aspect of the experimental setup. The recommended price is identical for all firms, both when the algorithm suggests collusion and when it recommends punishment. The instructions also point out that the price recommendation is non-binding, so each subject can choose a different price. This approach is motivated by the price suggestions that sellers receive in online marketplaces.

In real-world markets, sellers usually do not know how the algorithms work exactly and how they are programmed. We mimic this information structure in our experiment and do not describe the exact strategy of the algorithm besides the explanation that it aims at maximizing the joint profits of all sellers. Crucially, participants can learn the algorithm’s behavior over time. Hence, participants can learn that following the recommendation may benefit them in the long run. This again imitates the setup in real markets.

To mimic an infinitely repeated game, each round of the game has a continuation probability of 95%. Thus, with a probability of 5% each game terminates after a given round. Within this setup, the continuation probability is equivalent to the discount rate of δ = 0.95 (Roth & Murnighan, Reference Roth and Keith Murnighan1978). The game is repeated for three supergames to observe possible learning effectsFootnote 13. Within each supergame, the group composition is fixed. We use a perfect stranger matching scheme across supergames. Hence, the participants know they will meet each participant only once during the entire experiment. It rules out possible reputation effects that might arise otherwise. At the end of the experiment, the participants answered a post-experimental questionnaire (see Online Appendix B.3).

In total, we allocated 162 participants evenly across the three main treatmentsFootnote 14. The market and matching group sizes determine the number of independent observations. In the first supergame, there are no spillovers from one market to another. Hence, each of the 18 markets per treatment is an independent observation. In later supergames, participants are rematched with other participants from the same matching group. Each perfect stranger matching group consists of nine subjects. The markets are not independent anymore due to possible spillovers created by previous supergames. Therefore, the number of independent observations in later supergames is lower. To account for this dependency, we either cluster the standard errors at the matching group level or aggregate the respective outcome variable at the matching group level if we use nonparametric tests. Table 1 contains an overview of the number of independent observations.

Table 1. Number of observations by treatment

Note: The number of independent observations in later supergames is determined by the matching group size which always consists of nine participants.

We used an experimental currency unit (ECU) with an exchange rate of 100 ECU = EUR 1. On average, the participants received a payoff of EUR 10.73 plus a show-up fee of EUR 4Footnote 15. The average session length was 45 minutes.

3. Results

3.1. The influence of price recommendations on individual prices

Hypothesis 1 states that price recommendations influence individual prices as participants base their pricing decisions on the recommendations. To test this hypothesis, we regress the individual prices ( $p_t^i$) on the recommended prices ( $p_t^R$). The results of the linear regressions are shown in Table 2.

Table 2. Individual prices explained by the recommendation in a linear regression

Note: The coefficients are from a linear regression. Model specifications without fixed effects are estimated with a constant. For the fixed effects regression, we include dummies for each of the rounds and supergames.

*** Signif. codes: < 0.01, **< 0.05, *< 0.1; Clustered (matching group) standard errors in parentheses.

In all five columns, price recommendations positively and significantly affect individual prices. The effect remains significant when we control for lagged prices (column 2) and time-fixed effects (column 3). Furthermore, the effect size is more extensive for RecTheory than for RecSoft (column 4)Footnote 16. In specifications (3) and (4), we furthermore control for a set of individual-specific control variablesFootnote 17.

Notably, the additional effect of RecTheory diminishes in the final supergame (column 5). After the first two supergames, 97% of participants observed punishment by the algorithms at the Nash equilibrium stage game at least once and learned their mechanisms. Therefore, once participants understand how the recommendations work, both algorithms exhibit a similarly positive impact on prices. This supports Hypothesis 1, indicating that recommendations positively influence prices.

Result 1.

Sellers condition their prices on the recommendation of the algorithms. Price recommendations positively influence the individual sales prices of the participants.

In all regression specifications, the coefficient of the price recommendation is below one. It indicates that the price recommendation only translates partially into the individual price. Increasing the price recommendation by one only increases the individual price by 0.20 to 0.57, depending on the model specification and treatment. Thus, although prices change with the recommendations, it appears that, on average, participants do not fully follow them.

In both treatments, the modal action is to pick the recommended price. However, the average differences between the recommended and individual prices are 2.06 in RecSoft and -0.53 in RecTheory. Thus, in RecSoft, participants tend to undercut the recommendation, and in RecTheory, they do not always adhere to the punishment recommendation of the algorithm.

Furthermore, adherence to the algorithm differs depending on whether the recommendation is punitive or collusive. To illustrate the magnitude of these deviations for each state of the algorithms, we define an indicator variable that equals one if the recommended price is 10 (collusive state) and zero otherwise (punitive state). We regress the difference between the recommended and individual prices on this indicator variable. The results of this additional analysis, outlined in detail in Appendix C.1, show that participants choose a price 3.55 below the recommendation in the collusive state and exceed the recommendation by 1.77 in the punitive state. As such, in absolute terms, participants deviate from the recommendation more when the algorithms recommend colluding than when they recommend punishing. This difference across states is highly statistically significant in the first supergame (p < 0.01) and statistically significant at the 10% level across all supergames. However, the effect fades and is not statistically significant at any conventional level in the last supergame (separate Wald tests)Footnote 18. Overall, this additional analysis indicates that participants tend to undercut recommendations when collusion is recommended but exceed them during punishment. It highlights that although participants follow the recommendations, they do so imperfectly.

3.2. Collusive effects of price recommendations

Building on the finding that subjects use the algorithms’ recommendations for their pricing decisions, we now investigate whether the recommendations effectively foster collusion. Therefore, we compare the mean market prices in the treatments featuring recommendations with outcomes in the baseline treatment of no price recommendations. Note that the market price has a 1:1 relation with the sum of the sellers’ profits, so an analysis of the market price is equivalent to an analysis of the aggregate profits.

According to Hypotheses 2 and 3, price recommendations foster collusion as they provide a common reference point and simplify coordination on common punishment strategies after the deviation of a firm.

Figure 1 shows the mean market prices by treatment pooled across supergamesFootnote 19. The average market prices in Baseline and RecTheory are similar. There are no statistically significant differences (p-value= 0.818, two-sided Mann–Whitney U test). Thus, we find no evidence that, on average, the RecTheory recommendation algorithm fosters collusion.

Figure 1. Market price for the main treatments. The error bars represent 95% confidence intervals.

One of our initial conjectures was that the RecTheory recommendation algorithm, while constituting a subgame perfect Nash equilibrium, might feature too harsh punishments from the perspective of human players. We, therefore, designed the softer recommendation algorithm RecSoft. On balance, this algorithm, however, does not foster collusion either. In fact, the market prices are on average lower than in Baseline (p-value< 0.05, two-sided Mann–Whitney U test). In other words, the algorithm makes competitive market outcomes more likely even though the initial design objective was to make markets more collusive. It is in contrast to the consideration provided in Section 2.2 and to Hypotheses 2 and 3. Furthermore, average market prices in RecSoft are also lower than in RecTheory (p-value< 0.1, two-sided Mann–Whitney U test).

Result 2.

RecSoft leads to statistically significantly lower prices than Baseline and RecTheory. We do not find a statistically different mean price of RecTheory relative to Baseline.

Result 2 summarizes the findings about the average treatment effects. While we find support for the hypothesis that recommendations influence individual prices (Result 1), we find no evidence that, on average, price recommendations foster collusionFootnote 20. This effect also remains robust if we use linear regression of the market price on the different treatment indicators, including time-fixed effects or further control variablesFootnote 21. Furthermore, the average treatment effects are similar across supergames and remain present after participants have had the chance to learn the game and how the algorithms work. We report those additional results in Table C.10 and C.11 in the online appendix.

We do find evidence that price recommendations can lead to lower prices. Notably, this result is not driven by the fact that RecSoft leads to lower price recommendations than RecTheory. The average recommendation is similar in both treatments (5.40 vs. 5.94, p-value= 0.31, two-sided Mann-Whitney U test)Footnote 22. Thus, there appears to be a difference in participants’ responses to the different recommendation mechanisms. We explore the dynamics that drive behavior in the following section.

Our results indicate that fostering collusion purely through price recommendations may be difficult. On the contrary, price recommendations can make markets more competitive and lower market prices relative to a baseline without any recommendations. In the following section, we explore the mechanism for the price-decreasing effects of the soft recommendation algorithm. Furthermore, we discuss heterogeneous treatment effects for the RecTheory treatment. While the difference between Baseline and RecTheory could potentially become statistically significant with greater statistical power, the heterogeneity analysis highlights that the lack of significance mainly arises from opposing responses to the recommendations across markets.

3.3. Exploratory analysis: Heterogeneity and mechanisms

In this section, we explore heterogeneity among participants to uncover some of the mechanisms driving the results that we report in the previous section. In other words, we move beyond comparing treatment differences and conduct exploratory analyses within specific population subgroups based on the questionnaire items elicited at the end of the experiment. First, we document the heterogeneity in market outcomes across treatments and highlight specific stylized facts contributing to this heterogeneity. Then, we examine the price patterns that emerge in the different treatments.

3.3.1. Heterogeneous response to recommendations

There are substantial differences in market outcomes in RecTheory across matching groups. While the variance in average market prices in Baseline ( $\sigma^2=0.65$) and RecSoft ( $\sigma^2=0.34$) is small, there exists a large variation in RecTheory ( $\sigma^2=4.85$). Those differences in variances are statistically significant (p< 0.05, two separate Bartlett-tests)Footnote 23. This indicates that the recommendation algorithm RecTheory, which recommends strategies that constitute a subgame perfect Nash equilibrium, fosters more heterogeneous market outcomes.

To study the origin of the differences in variances, we show the maximal, minimal, and median average market price across matching groups for each treatment in Table 3. In line with the previous analysis, the median market price in RecSoft is small, and the maximal price is even below the median of the other treatments. Interestingly, although the median prices in baseline and RecTheory are similar, market prices are more spread out in RecTheory than in Baseline. The recommendations in RecTheory make specific markets more collusive, whereas they make others more competitive.

Table 3. Market price statistics by treatment

We confirm this by dividing the observations for each treatment into subgroups that are above (HIGH) and below (LOW) the treatment-specific median market price at the matching group level across all rounds. We observe that the average market prices for the RecTheory-High subgroup are statistically significantly higher than in Baseline-High, although only at the 10% level (two-sided MWU test). Also, the market prices in Baseline-Low are higher than in RecTheory-Low. Nevertheless, those differences are not statistically significant, likely due to a lack of statistical power because of the sample split (p=0.4, two-sided MWU test).

Result 3.

The variance in market outcomes is larger in RecTheory compared to RecSoft and Baseline.

Result 3 summarizes these findings. It provides some context for the lack of statistically significant differences between RecTheory and Baseline. While, on average, there is no significant difference in market prices between the two treatments, this does not imply that the recommendation itself has no effect. Instead, the average treatment effect masks an increase in heterogeneity, where some markets become more collusive while others become more competitive compared to the baseline treatment.

3.3.2. Relationship between negative reciprocity and the effect of recommendations

To understand the heterogeneity in market outcomes, we regress market prices on negative reciprocity. In market games, negative reciprocity measures the willingness to punish deviations from a set price level, which is critical for sustaining collusion. Especially with recommendations that actively promote punishment, participants may show varying sensitivity to the recommendation based on different levels of negative reciprocity. The observed heterogeneity appears to stem from differences in participants’ responses to the treatment, conditional on distinct levels of negative reciprocity, rather than from overall issues with randomization. In the online appendix, we provide randomization checks in Table C.2.

We elicited the economic preferences on the subject level at the end of the experiment using the validated post-experimental questionnaire by Falk et al. Reference Falk, Becker, Dohmen, Huffman and Sunde(2023)Footnote 24. We apply a min-max normalization to the variables on the individual level. Thus, all measures are between zero and one. Furthermore, we average them on the market level for the subsequent analysis.

Differences in negative reciprocity lead to vastly different market outcomes across treatments (see Table 4). In the Baseline treatment without any price recommendations, higher degrees of negative reciprocity lead to lower market prices, as indicated by the negative coefficient of Neg. Rec. in model specification 2. In other words, markets tend to exhibit lower prices if the participants are more inclined to punish each other when they feel maltreated. The ability to coordinate on collusive prices after a punishment phase may be missing for successful collusion.

Table 4. Market price explained by negative reciprocity and treatments

Note: The coefficients are from a linear regression. Model specifications without fixed effects are estimated with a constant. For the fixed effects regression, we include dummies for each of the rounds and supergames.

*** Signif. Codes: < 0.01, **< 0.05, *< 0.1; Clustered (Matching group) standard errors in parentheses.

For the treatments with price recommendations, this pattern is indeed different. The coefficients of the interaction terms with negative reciprocity are positive and statistically significant. Thus, as the degree of negative reciprocity increases, market prices increase in RecTheory and RecSoftFootnote 25. This suggests that the collusive recommendations and the willingness to punish deviations are complementary. Interestingly, the price level is lower for lower levels of negative reciprocity in RecTheory and RecSoft compared to Baseline, as indicated by the negative coefficients of RecTheory and RecSoft.

We interpret negative reciprocity as a willingness to punish deviations in this context. The recommendations harm collusion in markets, with sellers usually unwilling to punish. Possibly, the recommendations lead to harsh punishments that would not have happened without them. If participants are unable to recover from the punishment, the recommendations reduce the market prices below the level that is observed in markets without recommendations but with similarly low levels of negative reciprocity. The postulated mechanisms can explain around 10.5 % of the variation in the data, as indicated by the R-squared in column 3. We conclude that those heterogeneous treatment effects can explain lower prices than in Baseline for the treatments with a price recommendation. Result 4 summarizes our findings regarding negative reciprocity and the effectiveness of the recommendations.

Result 4.

Variations in negative reciprocity can explain the heterogeneous market outcomes. Recommendations reduce prices in markets where sellers have low negative reciprocity.

Those differences make intuitive sense and explain the considerable heterogeneity in market outcomes discussed in Result 3. The result emphasizes that algorithms can be pro-collusive for particular subgroups, even though we do not find statistically significant pro-collusive effects on average. Hence, if the algorithm’s provider understands their users and targets the recommendation to the specific population of sellers, the algorithm could increase the price level.

3.3.3. Price patterns across treatments

In Figure 2, we plot the market prices for each treatment by supergame and round. In the initial round, market prices in RecSoft and RecTheory are higher than in Baseline following the recommendation of $p^R_{t=1}=10$ (p-value=0.052 & p< 0.05, two-sided Mann–Whitney U tests)Footnote 26. Yet, there are deviations from the recommendation in 86.1% of all markets in the first round. As a result, the treatment-specific punishment recommendations are triggered in the subsequent round.

Figure 2. Market price for each treatment by supergame and round.

Let us focus first on the pattern of RecTheory in Figure 2. In response to deviations in the first round, the market prices drop for the following three rounds. At the end of this punishment phase, the prices increase sharply as the algorithm reverts to recommending the monopoly price. However, the prices do not stabilize entirely at this level. In the following rounds, there are reoccurring deviations after a recommendation of the monopoly price. This results in clearly visible spikes in the price pattern. In the second and third supergames, the spikes become less frequent, and the price patterns are more similar to Baseline.

The recurring deviations in RecTheory are almost entirely driven by matching groups with below median market prices (RecTheory-Low) as discussed in Section 3.3.1. This becomes clear when assessing the market price patterns for RecTheory for both subgroups separately. Whereas there are deviations in both subgroups in the first round, the market prices stabilize in RecTheory-High after the initial punishment phase. In RecTheory-Low, the share of markets with deviations from the collusive recommendations is significantly higher after the first round, which results in price spikes (p-value< 0.05, two-sided Mann–Whitney U test)Footnote 27. We visualize those price patterns in the online appendix in Figure C.3Footnote 28.

Matching groups with below-median market prices show repeated deviation patterns in the first supergame. They do not recover from this experience as average market prices remain lower throughout the rest of the experimentFootnote 29. Hence, we find suggestive evidence that the recommendation in RecTheory works as expected for specific subgroups. However, other participants repeatedly deviate from the recommendation, which leads to lower market prices than in Baseline for this subgroup.

For RecSoft, the price patterns in Figure 2 are also interesting. We designed this recommendation algorithm to be forgiving to slight deviations as it does not immediately punish at the stage game Nash equilibrium price of $p^{N}=1$ (see Section 2.2). We expected the punishment to be softer and, possibly, short compared to RecTheory. However, the data does not support this claim. After an initial deviation from the recommended price, the following recommendation is usually above the stage game Nash equilibrium ( $\overline{p}^R_{t=2} = 5.33$). Hence, in contrast to RecTheory, there are again profitable deviations from the recommendation in the following round. Participants repeatedly deviate from the recommendation. This triggers a downward spiral as the recommendation for the next period is again the deviation price. There are, on average, 5.44 rounds with a recommendation below the monopoly price after the first deviation in the first supergame. This initial punishment period is significantly longer than in RecTheory, which always punishes for three periods (p-value< 0.05, one-sided one-sample t-test). Due to those frequent deviations from the recommendation, market prices deteriorate in the first rounds and only recover insufficiently in the subsequent rounds. As a result, the average prices in the treatment RecSoft are low, and markets are even more competitive than in Baseline. It appears that the algorithms lead the participants to learn pro-competitive behavior as the effect remains present even in the last supergame as visualized in the most right panel of Figure 2 and respective regression in Table C.11.

In RecSoft, for $p^R_t \gt 5$, the mode of the difference between the recommended and individual prices is 9, while it is 0 for $p^R_t \leq 5$. This makes intuitive sense, given the discussed price patterns. Although the recommendation algorithm attempts to increase the price level by suggesting the monopoly price after all firms choose the same price, participants eventually disregard these upward recommendations. Many markets sustain a joint price of 1 after the initial deviations and punishments. As all prices are the same, the algorithm continues to recommend $p^R_t = 10$ despite the low price level. This leads to a significant discrepancy between the recommendations and prices for these high price recommendationsFootnote 30.

Result 5.

Repeated deviations from the recommendation in RecTheory lead to lower market prices for specific markets. The recommendation in RecSoft offers repeated deviation opportunities that decrease market prices.

Result 5 again emphasizes the adverse effects of recommendations for a third party that likes to foster collusion if they are not designed appropriately. Furthermore, it suggests that recommendations can be used to decrease sellers’ prices. It can be attractive in specific scenarios, for instance, to avoid excessive double marginalization. Sales intermediaries like platforms could specifically design a recommendation algorithm to foster competition among the sellers. Our results suggest those price recommendations are feasible using recommendation patterns as in RecSoft.

4. Concluding remarks

We derive two rule-based recommendation algorithms and study their effects on seller collusion in a stylized Bertrand market environment. Both algorithms aim to foster collusion compared to a baseline without any recommendation. The recommendation of the RecTheory algorithm uses harsh punishment phases after deviations from the recommended price and aims at implementing a subgame perfect Nash equilibrium. Motivated by experimental evidence, we also design a recommendation algorithm (RecSoft) that recommends softer punishments after a seller deviates from the collusive price. We test both algorithms in a laboratory experiment where each participant represents a seller.

We find evidence that the recommendations influence the sales prices in the sense that higher recommended sales prices induce sellers to set higher individual prices. The algorithm RecTheory, which recommends collusive trigger strategies with temporary Nash reversal, does not lead to higher prices on average. However, we find extensive and interesting heterogeneity in market outcomes. The variance in market prices increases in RecTheory compared to the baseline treatment, suggesting that while some markets are more collusive, others become more competitive. This indicates that the null effect is not primarily due to a lack of statistical power but instead reflects a divergence into opposing market outcomes, which masks the impact of the recommendations in the average. Furthermore, RecTheory lowers market prices in markets with low levels of negative reciprocity among sellers. For the behaviourally motivated algorithm RecSoft, which can recommend brief punishment phases with moderate price levels, we find lower market prices compared to the case without any recommendation. Participants frequently deviate from the recommendation, which starts a downward spiral that lowers prices. Similarly to RecTheory, market prices are lower than in Baseline for markets with sellers that have low negative reciprocity. There is no evidence that RecSoft facilitates collusion for any subgroup.

Our findings contrast with previous findings from targeted punishment in Cournot (Roux & Thöni, Reference Roux and Thöni2015) or public goods games (see, for instance, Fehr & Gächter, Reference Fehr and Gächter2000). In those experiments, participants can punish specific participants by explicitly reducing their payoff. This is highly effective at fostering cooperation. In our setup, the possibility of punishment works through coordination on a punishment strategy instead, an inherently less targeted approach. At the same time, the recommended trigger strategies theoretically have a considerable potential to foster collusion and are a standard modeling assumption when studying collusion. Our findings are surprising in this sense and suggest that coordination on non-targeted punishment strategies is insufficient to foster collusion.

While the results are, on balance, not alarming regarding the collusive risks of recommendation algorithms, we do provide reasons for potential concern. We find that recommendation algorithms can facilitate seller collusion in certain circumstances. Recommendations may foster collusion and harm consumers if the algorithm provider understands the sellers’ characteristics and targets the recommendation based on these characteristicsFootnote 31. Furthermore, in practice, recommendations can provide additional information on demand or help with pricing more generally, which could make sellers more likely to follow them. We chose to abstract from these factors in order to isolate the pure coordination effect. However, we suspect that the collusive potential of recommendations may be higher when there are other reasons for sellers to follow them. Therefore, we believe our experiment is relatively conservative in terms of demonstrating collusive effects. We consider it fruitful for future research to study the collusive effects of recommendations that incorporate these additional factors.

In other cases, we find that recommendation algorithms may even decrease prices despite being designed and intended to facilitate collusion. The finding is consistent with the theoretical insight that all players following this behaviorally motivated algorithm does not constitute a subgame perfect Nash equilibrium. One interpretation of the price-reducing effects is that platforms may be able to use recommendation algorithms to make the sellers’ offers more competitive. Under certain circumstances, such as excessive double marginalization or a dynamic pricing strategy, this could be in the interest of a sales platform. A caveat applies, as we told our experiment participants that the algorithm would aim to increase prices and profits, which aligned with our expectations. On average, the opposite, however, turned out to be the case for this algorithm. Over time, sellers may thus lose trust in following the algorithm’s recommendations. More research in this regard would be desirable.

In the experiment, we abstract from the contracts between the recommendation provider, the sellers, and the developers, thus excluding their incentives. From the perspective of vertical relations theory, it raises the question of why a provider or developer would need to facilitate seller collusion when other mechanisms could influence sales prices. For instance, from an ex-ante perspective, it is unclear why a sales platform would provide collusive price recommendations instead of charging higher commission rates. In Online Appendix D, we study contracting between such an intermediary and the sellers theoretically, showing that an intermediary can benefit from collusive recommendations. Future research could explore the role of the recommendation provider and developer as strategic players more explicitly in experiments.

Supplementary material

The supplementary material for this article can be found at https://doi.org/10.1017/eec.2025.9.

Acknowledgements

We thank participants of the ESA World 2022, DICE Brown Bag, EARIE 2023 and, in particular, Lisa Bruttel, Johannes Muthers, and Hans-Theo Normann for their helpful comments and suggestions. We thank Robin Bitter, Leon Heidelbach, Vera Konrad, and, in particular, Mara Siegesmund for excellent research assistance. The replication material for the study is available at https://doi.org/10.17605/OSF.IO/V2TMG.

Statements and Declarations

The authors declare that they have no competing financial or non-financial interests.

Footnotes

1 For instance, Airbnb uses price recommendation algorithms that utilize historical and geographical data and combine machine learning methods with human intuition. Furthermore, the algorithmic price recommendation changes daily for the upcoming dates. See Hill Reference Hill(2015) for details.

2 For instance, price recommendations might also reduce information asymmetries between the platform and sellers, which makes them potentially attractive from a business perspective beyond collusion. For instance, platforms may have better demand information than individual sellers, and they can use price recommendations to transmit this information (see Pavlov & Berman, Reference Pavlov and Berman2019, Lefez, Reference Lefez2021). For a discussion on the overall incentives of platforms to foster collusion in the seller market, see, for instance, Teh Reference Teh(2022), Schlütter (Reference Schlütter2022), and Online Appendix D.

3 See Vogell, Coryne & Little, “Technology Rent Going Up? One Company’s Algorithm Could Be Why.”, https://www.propublica.org/article/yieldstar-rent-increase-realpage-rent, last accessed March 17, 2025.

4 Recent studies by Musolff (Reference Musolff2022) and Hanspach et al. (Reference Hanspach, Sapi and Wieting2024) highlight that price algorithms on sales platforms often follow simple rule-based logic. Similarly, even complex reinforcement learning algorithms often converge to strategies that simple rules can describe (see, for instance, Werner, Reference Werner2022).

5 While we consider neutrally framed dynamic recommendation algorithms that depend on the history of the game, we also conducted a static recommendation treatment, similar to this aspect of the study by Sonntag and Zizzo Reference Sonntag and John Zizzo(2015). We document it in Appendix C.4.

6 Horstmann et al. Reference Horstmann, Krämer and Schnurr(2018) provide a meta-study of the relationship between the number of firms in the market and tacit collusion. They find that collusion is more challenging to sustain when moving from two to three-firm markets. Also, the meta-study by Engel Reference Engel(2015) highlights that market prices are decreasing in the number of firms. Freitag et al. Reference Freitag, Roux and Thöni(2021) report no evidence of collusive outcomes within a multi-market environment with three firms.

7 Differentiation between products or sellers is a typical feature in many markets. As is common in many models and experiments, however, we abstract from this for the sake of simplicity. Qualitatively analogous results for the recommendation algorithms can be obtained, at least theoretically, when introducing differentiation. The detailed derivations are in Online Appendix A.

8 We show that it is incentive-compatible to follow the recommendations in our setup in Online Appendix A.

9 The Q-learning algorithms in Klein Reference Klein(2021) punish for a certain number of periods before reverting to the monopoly price. In Werner Reference Werner(2022), they learn one-period punishment strategies similar to a win-stay lose-shift strategy where algorithms revert to the stage game Nash equilibrium for punishment.

10 We consider two additional mechanisms as a robustness check that we report in Online Appendix C.4. In RecStatic, the algorithm provides a static price recommendation at the monopoly price similar to Sonntag and Zizzo Reference Sonntag and John Zizzo(2015). Additionally, we analyze an algorithm similar to RecTheory but with a shorter punishment phase. Both additional algorithms do not foster collusion compared to Baseline.

11 We also consider a recommendation algorithm without any punishment in the Online Appendix C.4.

12 All control questions are in Online Appendix B.3.

13 The exact number of rounds was pre-drawn with a random number generator to allow for the same supergame length across different experimental sessions. The round numbers are 27 (Supergame 1), 8 (Supergame 2), and 18 (Supergame 3).

14 For details on the additional control treatments see Online Appendix C.4. The control treatments highlight the effects of algorithms without any punishment or fewer punishment periods.

15 During the COVID-19 pandemic, we paid each participant an additional EUR 4. This bonus was announced after the end of the session. Thus, it does not influence the behavior in the experiment itself.

16 The effect is not driven by a difference in the average price recommendation across both treatments. We discuss this in the following section.

17 These include economic preferences and measures of socioeconomic status. We provide a list in the Online Appendix B.3.

18 Further details on the statistical tests are provided in Appendix C.1.

19 Furthermore, we provide an overview of the development of market prices across time in Figure 2.

20 Moreover, algorithms without any punishment mechanism or with fewer punishment periods do not increase market prices. See Online Appendix C.4 for details.

21 The control variables were elicited on an individual level using a post-experimental questionnaire (see Online Appendix B.3). We aggregate them on the group level by calculating the mean across all group members.

22 We also provide a plot of the distribution of recommendations in Figure C.2 in the online appendix.

23 As in the previous analysis, we aggregate the market prices at the matching group level. Thereby, we account for dependencies that arise by rematching participants at the end of each supergame. It allows for correct statistical inference. We provide an overview of the number of independent observations in Table 1.

24 Next to negative reciprocity, the questionnaire also includes positive reciprocity, time preferences, risk aversion, and measures of altruism and trust. We report the results regarding those variables in Online Appendix C.3.

25 The average marginal effect of the treatment dummies is not statistically significant at the 10%-level if Neg. Rec. is equal to one (see Table C.8 in the online appendix). Note that one is the maximal value that Neg. Rec. can take due to the normalization we apply.

26 Market prices in RecTheory and RecSoft are similar in the first round following the same initial recommendation (p=0.25, two-sided Mann–Whitney U test), which confirms that randomization into treatments worked.

27 We test this by restricting the data to the first supergame and to rounds in which the monopoly price was recommended. Then, we calculate for each market in RecTheory-High and RecTheory-Low the share of rounds in which at least one participant deviated from the recommendation. We test for differences in this variable across the two subgroups. Rematching only occurs after the first supergame, so each market constitutes an independent observation, allowing correct inference.

28 For the respective analysis for RecSoft see Figure C.4.

29 Similarly, Dal Bó and Fréchette Reference Bó, Pedro and Fréchette(2018) show that participants’ initial experience in the infinitely repeated Prisoners Dilemma is essential for their cooperation behavior in subsequent supergames.

30 Note that in RecTheory, the modal price is always the recommendation, independent of whether the recommendation is to pick a high or low price. Further analyses on the deviation from the recommendation are provided in Appendix C.1.

31 For example, accommodation platforms may gather more and more data about their hosts and guests over time and thus could condition their recommendations on seller characteristics in specific local markets to make them more effective.

References

, D., Pedro, , & Fréchette, G. R. (2018). On the determinants of cooperation in infinitely repeated games: A survey. Journal of Economic Literature, 56(1), 60114.Google Scholar
, D., Pedro, , & Fréchette, G. R. (2019). Strategy choice in the infinitely repeated Prisoner’s Dilemma. American Economic Review, 109(11), .Google Scholar
Brown, Z. Y., & MacKay, A. (2023). Competition in pricing algorithms. American Economic Journal: Microeconomics, 15(2), 109156.Google Scholar
Bruttel, L. (2018). The effects of recommended retail prices on consumer and retailer behaviour. Economica, 85(339), 649668.10.1111/ecca.12218CrossRefGoogle Scholar
Buehler, S., & Gärtner, D. L. (2013). Making sense of nonbinding retail-price recommendations. American Economic Review, 103(1), .10.1257/aer.103.1.335CrossRefGoogle Scholar
Bundeskartellamt and Autorité de la concurrence. (2019) Algorithms and competition. Available at, https://www.bundeskartellamt.de/SharedDocs/Publikation/EN/Berichte/Algorithms_and_Competition_Working-Paper.pdf.Google Scholar
Cadsby, C. B., Maynes, E., & Umashanker Trivedi, V. (2006). Tax compliance and obedience to authority at home and in the lab: A new experimental approach. Experimental Economics, 9(4), 343359.10.1007/s10683-006-7053-8CrossRefGoogle Scholar
Calvano, E., Calzolari, G., Denicolo, V., & Pastorello, S. (2020). Artificial intelligence, algorithmic pricing, and collusion. American Economic Review, 110(10), .10.1257/aer.20190623CrossRefGoogle Scholar
Chen, D. L., Schonger, M., & Wickens, C. (2016). oTree-An open-source platform for laboratory, online, and field experiments. Journal of Behavioral and Experimental Finance, 9, 8897.10.1016/j.jbef.2015.12.001CrossRefGoogle Scholar
Competition and Markets Authority, Algorithms: How they can reduce competition and harm consumers, Available at: https://www.gov.uk/government/publications/algorithms-how-they-can-reduce-competition-and-harm-consumers 2021.Google Scholar
Croson, R., & Marks, M. (2001). The effect of recommended contributions in the voluntary provision of public goods. Economic Inquiry, 39(2), 238249.10.1111/j.1465-7295.2001.tb00063.xCrossRefGoogle Scholar
Duffy, J., & Feltovich, N. (2010). Correlated equilibria, good and bad: An experimental study. International Economic Review, 51(3), 701721.10.1111/j.1468-2354.2010.00598.xCrossRefGoogle Scholar
Dufwenberg, M., & Gneezy, U. (2000). Price competition and market concentration: An experimental study. International Journal of Industrial Organization, 18(1), 722.10.1016/S0167-7187(99)00031-4CrossRefGoogle Scholar
Engel, C. (2015). Tacit collusion: The neglected experimental evidence. Journal of Empirical Legal Studies, 12(3), 537577.10.1111/jels.12080CrossRefGoogle Scholar
Ezrachi, A., & Stucke, M. E. (2017). Artificial intelligence & collusion: When computers inhibit competition. U. Ill. L. Rev., 2017, 1775.Google Scholar
Faber, R. P., & Janssen, M. C. W. (2019). On the effects of suggested prices in gasoline markets. The Scandinavian Journal of Economics, 121(2), 676705.10.1111/sjoe.12289CrossRefGoogle Scholar
Falk, A., Becker, A., Dohmen, T., Huffman, D., & Sunde, U. (2023). The preference survey module: A validated instrument for measuring risk, time, and social preferences. Management Science, 69(4), 19351950.10.1287/mnsc.2022.4455CrossRefGoogle Scholar
Federal Trade Commission and U.S. Department of Justice (2024). Statement of Interest of the United States. Case No., 1(23), March :-cv-02536-KMW-EAP.Google Scholar
Fehr, E., & Gächter, S. (2000). Cooperation and punishment in public goods experiments. American Economic Review, 90(4), 980994.10.1257/aer.90.4.980CrossRefGoogle Scholar
Fonseca, M. A., & Normann, H. -T. (2012). Explicit vs. tacit collusion-The impact of communication in oligopoly experiments. European Economic Review, 56(8), 17591772.10.1016/j.euroecorev.2012.09.002CrossRefGoogle Scholar
Foros, O., & Steen, F. (2013). Vertical control and price cycles in gasoline retailing. The Scandinavian Journal of Economics, 115(3), 640661.10.1111/sjoe.12024CrossRefGoogle Scholar
Freitag, A., Roux, C., & Thöni, C. (2021). Communication and market sharing: an experiment on the exchange of soft and hard information. International Economic Review, 62(1), 175198.10.1111/iere.12480CrossRefGoogle Scholar
Gill, D., & Thanassoulis, J. (2016). Competition in posted prices with stochastic discounts. The Economic Journal, 126(594), 15281570.10.1111/ecoj.12294CrossRefGoogle Scholar
Greiner, B. (2015). Subject pool recruitment procedures: organizing experiments with ORSEE. Journal of the Economic Science Association, 1(1), 114125.10.1007/s40881-015-0004-4CrossRefGoogle Scholar
Hansen, K. T., Misra, K., & Pai, M. M. (2021). Frontiers: Algorithmic collusion: Supra-competitive prices via independent algorithms. Marketing Science, 40(1), 112.10.1287/mksc.2020.1276CrossRefGoogle Scholar
Hanspach, P., Sapi, G., & Wieting, M. (2024). Algorithms in the marketplace: An empirical analysis of automated pricing in e-commerce. Information Economics and Policy, 69, 101111.10.1016/j.infoecopol.2024.101111CrossRefGoogle Scholar
Harrington, J. E. (2022). The effect of outsourcing pricing algorithms on market competition. Management Science, 68(9), 68896906.10.1287/mnsc.2021.4241CrossRefGoogle Scholar
Harrington, R. H. G., & Kujal, P. (2016). The relative efficacy of price announcements and express communication for collusion: Experimental findings. Journal of Economic Behavior & Organization, 128, 251264.10.1016/j.jebo.2016.05.014CrossRefGoogle Scholar
Harstad, R., Martin, S. and Normann, H. -T. (1998). Intertemporal Pricing schemes: Experimental Tests of Consciously Parallel Behavior in oligopoly Applied Industrial Organization. Cambridge University Press.10.1017/CBO9780511522048.008CrossRefGoogle Scholar
Hill, D. (2015). How much is your spare room worth?. IEEE Spectrum, 52(9), 3258.10.1109/MSPEC.2015.7226609CrossRefGoogle Scholar
Holt, C. A., & Davis, D. (1990). The effects of non-binding price announcements on posted-offer markets. Economics letters, 34(4), 307310.10.1016/0165-1765(90)90136-OCrossRefGoogle Scholar
Horstmann, N., Krämer, J., & Schnurr, D. (2018). Number effects and tacit collusion in experimental oligopolies. The Journal of Industrial Economics, 66(3), 650700.10.1111/joie.12181CrossRefGoogle Scholar
Huck, S., Normann, H. -T., & Oechssler, J. (2004). Two are few and four are many: number effects in experimental oligopolies. Journal of Economic Behavior & Organization, 53(4), 435446.10.1016/j.jebo.2002.10.002CrossRefGoogle Scholar
Janssen, M., & Reshidi, E. (2022). Regulating recommended retail prices. International Journal of Industrial Organization, 85, 102872.10.1016/j.ijindorg.2022.102872CrossRefGoogle Scholar
Johnson, J. P., Rhodes, A., & Wildenbeest, M. (2023). Platform Design When Sellers Use Pricing Algorithms. Econometrica, 91(5), 18411879.10.3982/ECTA19978CrossRefGoogle Scholar
Kasberger, B., Martin, S., Normann, H. -T., & Werner, T. (2023). Algorithmic Cooperation. Available at SSRN 4389647.Google Scholar
Klein, T. (2021). Autonomous algorithmic collusion: Q-learning under sequential pricing. The RAND Journal of Economics, 52(3), 538558.10.1111/1756-2171.12383CrossRefGoogle Scholar
Lefez, W. (2021) Price recommendations and the value of data: A mechanism design approach. Working Paper.10.2139/ssrn.4154435CrossRefGoogle Scholar
Lubensky, D. (2017). A model of recommended retail prices. The RAND Journal of Economics, 48(2), 358386.10.1111/1756-2171.12179CrossRefGoogle Scholar
Musolff, L. (2022). Algorithmic pricing facilitates tacit collusion: Evidence from e-commerce. In Proceedings of the 23rd ACM Conference on Economics and Computation, July (pp. 3233). https://dl.acm.org/doi/abs/10.1145/3490486.3538239CrossRefGoogle Scholar
Normann, H. -T., & Sternberg, M. (2023). Human-algorithm interaction: Algorithmic pricing in hybrid laboratory markets. European Economic Review, 152, 104347.10.1016/j.euroecorev.2022.104347CrossRefGoogle Scholar
Pavlov, V. and Berman, R. (2019) Price manipulation in Peer-to-Peer markets and the sharing economy. Working Papers 19-10, NET Institute September.10.2139/ssrn.3468447CrossRefGoogle Scholar
Roth, A. E., & Keith Murnighan, J. (1978). Equilibrium behavior and repeated play of the prisoner’s dilemma. Journal of Mathematical Psychology, 17(2), 189198.10.1016/0022-2496(78)90030-5CrossRefGoogle Scholar
Roux, C., & Thöni, C. (2015). Collusion among many firms: The disciplinary power of targeted punishment. Journal of Economic Behavior & Organization, 116, 8393.10.1016/j.jebo.2015.03.018CrossRefGoogle Scholar
Schlütter, F. (2022) Managing seller conduct in online marketplaces and platform most-favored nation clauses. Working Paper.Google Scholar
Schotter, A., & Sopher, B. (2003). Social learning and coordination conventions in intergenerational games: An experimental study. Journal of Political Economy, 111(3), 498529.10.1086/374187CrossRefGoogle Scholar
Silverman, D., Slemrod, J., & Uler, N. (2014). Distinguishing the role of authority “in” and authority “to”. Journal of Public Economics, 113, 3242.10.1016/j.jpubeco.2014.02.003CrossRefGoogle Scholar
Sonntag, A., & John Zizzo, D. (2015). Institutional authority and collusion. Southern Economic Journal, 82(1), 1337.10.1002/soej.12065CrossRefGoogle Scholar
Teh, T. -H. (2022). Platform governance. American Economic Journal: Microeconomics, 14(3), 213254.Google Scholar
Werner, T. (2022). Algorithmic and human collusion. Available at SSRN, 3960738.Google Scholar
Wright, J. (2013). Punishment strategies in repeated games: Evidence from experimental markets. Games and Economic Behavior, 82, 91102.10.1016/j.geb.2013.06.012CrossRefGoogle Scholar
Figure 0

Table 1. Number of observations by treatment

Figure 1

Table 2. Individual prices explained by the recommendation in a linear regression

Figure 2

Figure 1. Market price for the main treatments. The error bars represent 95% confidence intervals.

Figure 3

Table 3. Market price statistics by treatment

Figure 4

Table 4. Market price explained by negative reciprocity and treatments

Figure 5

Figure 2. Market price for each treatment by supergame and round.

Supplementary material: File

Hunold and Werner supplementary material

Hunold and Werner supplementary material
Download Hunold and Werner supplementary material(File)
File 1.1 MB