Hostname: page-component-cb9f654ff-lqqdg Total loading time: 0 Render date: 2025-09-09T12:00:37.912Z Has data issue: false hasContentIssue false

Optimal stopping with variable attention

Published online by Cambridge University Press:  09 September 2025

Jukka Lempa*
Affiliation:
University of Turku
Harto Saarinen*
Affiliation:
Turku School of Economics
Wiljami Sillanpää*
Affiliation:
University of Turku
*
*Postal address: Department of Mathematics and Statistics, University of Turku, Vesilinnantie 5, Turku, FI-20014, Finland.
***Postal address: Department of Economics, Turku School of Economics, Rehtorinpellonkatu 3, Turku, FI-20014, Finland. Email: hoasaa@utu.fi
*Postal address: Department of Mathematics and Statistics, University of Turku, Vesilinnantie 5, Turku, FI-20014, Finland.
Rights & Permissions [Opens in a new window]

Abstract

We consider an optimal stopping problem of a linear diffusion under Poisson constraint where the agent can adjust the arrival rate of new stopping opportunities. We assume that the agent may switch the rate of the Poisson process between two values. Maintaining the lower rate incurs no cost, whereas the higher rate requires effort that is captured by a cost function c. We study a broad class of payoff functions, cost functions and diffusion dynamics, for which we explicitly characterize the solution to the constrained stopping problem. We also characterize the case where switching to the higher rate is always suboptimal. The results are illustrated with two examples.

Information

Type
Original Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution, and reproduction in any medium, provided the original work is properly cited.
Copyright
© The Author(s), 2025. Published by Cambridge University Press on behalf of Applied Probability Trust

1. Introduction

Consider an agent with an option to invest in the style of [Reference McDonald and Siegel30]. That is, the agent can make an irreversible investment in a project, where the returns follow a continuous-time stochastic process. In standard models of this style it is often assumed that the option can be exercised at any time. This reflects the idea that the agent is attentive to the relevant state variable at all times. In practice, this type of constant attention can be very costly, even prohibitively so. The issue of costly information acquisition and/or limited information processing capacity in economic decision making is addressed in the economics literature under the title ‘rational inattention’. Rational inattention can be described so that when there is uncertainty about future events regarding a decision making problem and obtaining and processing information comes with a cost (at least a utility or an opportunity cost), it is rational for agents to update their information and strategies only sporadically. We refer the reader to the recent survey by Maćkowiak, Matĕjka and Wiederholt [29] for a comprehensive overview on the rational inattention literature.

Mathematical models of rational inattention can be roughly divided into two main categories: structural models and models of reduced type. In structural models the information constraint is modelled directly and its effects on the resulting decision making problem are studied; see, e.g., [Reference Sims36, Reference Sims37]. In these papers the observations on the system are acquired through a communication channel with a finite information processing capacity. As a result, rational agents are not able to utilize all the available information instantaneously and, hence, cannot optimally update their strategies in continuous time.

In models of reduced type one introduces a simplified proxy for attention into the model to study its implications on the optimal strategies. The proxy is usually chosen to be a deterministic or random discrete set of admissible decision making times. Papers of this type include, e.g., [Reference Huang and Liu17, Reference Liu, Kadan and Yang28, Reference Moscarini32, Reference Reis33]. In [Reference Moscarini32], estimation of an Ornstein–Uhlenbeck process with discrete periodic noisy observations is considered. Expectation formation by consumers is studied in [Reference Reis33] in the case where consumers update their plans only sporadically and remain inattentive otherwise. A similar idea is adapted in [Reference Huang and Liu17], where portfolio selection with only discrete periodic news updates is considered. Executive option valuation is considered in [Reference Liu, Kadan and Yang28], where the proxy for attention is an independent Poisson process and the option can be exercised only at the jump times of this process.

Our study is related to the models of reduced type. Consider an agent observing the value of a dynamic stochastic variable at discrete irregularly spaced time points. The agent is facing an optimal stopping problem, where the objective is to maximize the expected present value of the exercise payoff contingent on the observations. The irregular observations stem from a linear diffusion process, which is observed at the arrival times of two independent Poisson processes with rates $\lambda_1 < \lambda_2$ . The observation mechanism is set up as follows. At the initial time, the agent may stop or choose one of the Poisson processes and acquire the next observation of the diffusion at the first arrival of the chosen Poisson process. If the agent chooses the Poisson process with higher rate, she has to pay a potentially state-dependent cost c. At the arrival of the chosen Poisson process, the agent can choose between the following: she can either stop, continue following the same Poisson process or switch to the other Poisson process. The procedure is then repeated perpetually. The rate parameter of each Poisson process is understood as the mean number of observations per unit time contingent on the agent’s decision. We call these parameters attention rates since they refer to the agent’s level of attention. Lower attention can be maintained for free, whereas higher attention requires additional effort that is modelled by the cost c.

The decision rule of our agent is seen to be two fold. It is a combination of the stopping time and the sequence of decisions modulating the attention rate. Thus, our problem may be cast as a problem of stochastic control with an additional optimal stopping component. These types of problems are also known as stochastic control problems with discretionary stopping and there exists a wealth of literature on them. We refer the interested reader to [Reference Beneš, Gaitsgori and Karatzas7] for a recent literature review. In most papers that have been written on the topic (such as [Reference Davis and Zervos9, Reference Karatzas and Sudderth18, Reference Karatzas and Zamfirescu20]), the controlling aspect is formulated as a control that directly affects either the state or the dynamics of the underlying stochastic process. A notable exception is the cooperative game set-up of [Reference Karatzas and Zamfirescu19] in which controlling is understood as choosing a probability measure under which the expected payoff is calculated. Our model conforms to the latter framework since the attention rate control influences only the probability distribution of the next controlling/stopping opportunity.

As an application of our model, consider the optimal liquidation of a reserve of a renewable natural resource. The standard models (see, e.g., [Reference Alvarez4, Reference Alvarez and Shepp5, Reference Song, Stockbridge and Zhu38]) usually assume that the level of reserve is observable at all times. In practice, this quantity can be difficult to monitor due to different physical constraints. For instance, the reserve can be distributed spatially over a large area and it takes time and effort to quantify its level for decision making purposes. This setting can be captured by the proposed models. The reserve level is no longer observable continuously in time but rather for a given number of times per unit time on average. The rate $\lambda_1$ is the average number of observations per time unit that can be maintained without additional costs, whereas maintaining a higher average number of observations $\lambda_2$ incurs an extra cost c.

The proposed model is related to the recent paper by Hobson and Zeng [Reference Hobson and Zeng15]. In [Reference Hobson and Zeng15] the agent observes the diffusion and has the capacity to control the rate of the Poisson process continuously in time. Moreover, maintaining a rate will incur an instantaneous cost that is a function of the maintained rate; maintaining a higher rate is more costly than maintaining a lower rate. We consider an irregularly spaced discrete time version of this problem, where observations and rate adjustments are made only at the arrival times. However, in many cases studied in [Reference Hobson and Zeng15] the optimal adjustment rule is to switch between two rates, from ‘low’ to ‘maximal’.

In this view, our restriction to modulate between two rates is reasonable. In [Reference Hobson and Zeng15] the authors consider examples with Brownian motion and geometric Brownian motion (GBM), and note that their results could be generalized to more general diffusions. The main contribution of the study at hand is to present a broad class of models with easily verifiable conditions for which the optimal solution can be characterized in terms of the minimal r-excessive functions of the diffusion process. This complements the existing literature. We also characterize the case where switching to the higher rate is always suboptimal.

Furthermore, models in which the underlying evolves in continuous time but the admissible control times are restricted have been studied in the literature over the past two decades. They are often called Poisson constrained optimal stopping problems. These problems were introduced by [Reference Dupuis and Wang10], where the diffusion is a GBM and the exercise payoff is of an American call option type. The results of [Reference Dupuis and Wang10] were extended in [Reference Lempa23] to cover a broad class of payoff functions and linear diffusion dynamics. Menaldi and Robin [Reference Menaldi and Robin31] provide further generalizations, going so far as to have a not necessarily time-homogeneous Markov process with a locally compact state space and independent and identically distributed intervention times. However, at this level of generality not much can be said about the optimal stopping times besides the usual characterization and well-posedness results. Other papers in this vein include [Reference Arai and Takenaka6, Reference Gapeev11Reference Hobson, Liang and Wang16, Reference Lange, Ralph and Støre21, Reference Lempa and Saarinen24Reference Liang and Wei27, 34]. Arai and Takenaka [Reference Arai and Takenaka6] study Poisson constrained optimal stopping with a regime switching GBM. In [Reference Guo and Liu12], optimal stopping of the maximum of the GBM at Poisson arrivals is considered. The shape properties of the value function for a class of Poisson stopping problems with diffusion dynamics are analysed in [Reference Hobson and Zeng14]. In [Reference Lange, Ralph and Støre21], optimal stopping with Poisson constraint is considered for a multidimensional state variable. Optimal stopping games with Poisson constraints under various degrees of generality are addressed in [Reference Gapeev11, Reference Hobson, Liang and Wang16, Reference Lempa and Saarinen24Reference Liang and Sun26]. Lastly, Liang and Wei [Reference Liang and Wei27] study optimal switching at Poisson arrivals and Rogers and Zane [34] consider portfolio optimization under the Poisson constraint. There is also a wide variety of Poisson-constrained control problems that are not purely stopping problems. For a comprehensive and recent review on this topic, see [Reference Saarinen35, Section 3.4]. We point out that in all these papers, the rate of the Poisson process is not a decision variable.

The structure of the paper is as follows. In Section 2 we formulate the main stopping problems. In Section 3 we derive the candidate value functions and associated policies. The candidate values are proved to be the actual values in Section 4. Section 5 contains a brief discussion on the asymptotic properties of the main stopping problem. The findings are illustrated in Section 6.

2. Problem formulation

2.1. The dynamics

Let $(\Omega, \mathcal{F}, \mathbb{F} = (\mathcal{F}_t )_{t\geq 0}, \mathbb{P})$ be a filtered probability space supporting a diffusion X with a state space I that is an interval in $\mathbb{R}$ such that $\overline{I} = [\mathfrak{l}, \mathfrak{r} ], -\infty \leq \mathfrak{l} < \mathfrak{r} \leq \infty$ . We denote by $\mathbb{P}_x$ the probability measure $\mathbb{P}$ conditioned on the event that the diffusion is started from the state $x \in I$ . The corresponding expectation is denoted by $\mathbb{E}_x[\cdot ]$ . We assume that the boundaries $\mathfrak{l}, \mathfrak{r}$ are natural. Thus, the lifetime of X

\begin{align*}\zeta &= \inf \{t\geq0\colon X_t \notin I\}\end{align*}

satisfies $\zeta = \infty$ , $\mathbb{P}$ -almost surely. As usual, we denote by m and S the speed measure and the scale function of X, which we assume to be absolutely continuous with respect to the Lebesgue measure on I and twice continuously differentiable. The infinitesimal generator $\mathcal{A}\colon C^2(I) \rightarrow C_b(I)$ of X can be written as

\begin{align*} \mathcal{A} &= \dfrac{\sigma^2(x)}{2} \dfrac{d^2}{dx^2} + \mu(x)\dfrac{d}{dx},\end{align*}

and the infinitesimal parameters $\mu\colon I \rightarrow \mathbb{R}$ , $\sigma^2\colon I \rightarrow \mathbb{R}_+ $ are defined by

\begin{align*} m'(x) = \dfrac{2}{\sigma^2(x)}\textrm{e}^{B(x)},\qquad S'(x) = \textrm{e}^{-B(x)},\end{align*}

where $B(x) = -\int^x({2\mu(y)}/{\sigma^2(y)})\,\textrm{d}y$ . Note that we use the convention $\mathbb{R}_+ = (0, \infty)$ here. Let $r > 0$ . We denote by $\boldsymbol{\psi_{r}}$ ( $\boldsymbol{\varphi}_{r}$ ) the unique increasing (decreasing) fundamental solution to the ordinary differential equation $\mathcal{A}u = r u$ . These fundamental solutions can be identified as minimal r-excessive mappings for the diffusion X. Let $L^1_{r}(I)$ be the set of measurable functions $f\colon I \rightarrow \mathbb{R}$ such that

\begin{align*} \mathbb{E}_x\left[ \int_0^{\infty}\textrm{e}^{-r s}|f(X_s)| \,\textrm{d}s \right] &< \infty\end{align*}

and, for any $f \in L^1_{r}(I)$ , define the resolvent $\textbf{R}_{r}$ as

(1) \begin{align} (\textbf{R}_{r}f)(x) &= \mathbb{E}_x\left[ \int_0^{\infty}\textrm{e}^{-r s}f(X_s) \,\textrm{d}s \right].\end{align}

By combining the relevant identities from Sections I.2.7, II.1.4 and II.1.11 in [Reference Borodin and Salminen8] and making use of Fubini’s theorem (which holds for the integral $(R_rf)$ when $f \in L^1_{r}(I)$ ), we see that the resolvent $\textbf{R}_{r}$ can be expressed in terms of the minimal r-excessive functions $\boldsymbol{\psi}_{r}$ and $\boldsymbol{\varphi}_{r}$ as

(2) \begin{align} (\textbf{R}_{r}f)(x)&= B_{r}^{-1} \left( \boldsymbol{\varphi}_{r}(x) (\boldsymbol{\Psi}_{r}f)(x) + \boldsymbol{\psi}_{r}(x) (\boldsymbol{\Phi}_{r}f)(x) \right).\end{align}

Here, $B_{r}$ is the (constant) Wronskian

(3) \begin{align} B_{r} &= \dfrac{\boldsymbol{\psi}^{\prime}_{r}(x)\boldsymbol{\varphi}_{r}(x) - \boldsymbol{\psi}_{r}(x)\boldsymbol{\varphi}^{\prime}_{r}(x)}{S'(x)}\end{align}

and the functionals $\Psi_{r}$ and $\Phi_{r}$ are defined as

\begin{align*} (\boldsymbol{\Psi}_{r}f)(x) &= \int_{\mathfrak{l}}^x\boldsymbol{\psi}_{r}(z)f(z)m'(z)\,\textrm{d}z,\\ (\boldsymbol{\Phi}_{r}f)(x) &= \int_x^{\mathfrak{r}}\boldsymbol{\varphi}_{r}(z)f(z)m'(z)\,\textrm{d}z.\end{align*}

We also recall the resolvent equation (see, e.g., [Reference Borodin and Salminen8, p. 4])

(4) \begin{equation} \textbf{R}_q \textbf{R}_r= \frac{\textbf{R}_r-\textbf{R}_q}{q-r},\end{equation}

where $q>r>0$ .

Lastly, we make use of the formulae (see, e.g., [Reference Alvarez2, Corollary 3.2])

(5) \begin{align} \frac{f'(x)\boldsymbol{\psi}_r(x) - \boldsymbol{\psi}^{\prime}_{r}(x)f(x)}{S'(x)} = \int_{\mathfrak{l}}^x\boldsymbol{\psi}_r(z)((\mathcal{A} - r)f)(z)m'(z) \,\textrm{d}z,\end{align}
(6) \begin{align} \frac{\boldsymbol{\varphi}^{\prime}_{r}(x)h(x) - h'(x)\boldsymbol{\varphi}_r(x)}{S'(x)} = \int_x^{\mathfrak{r}}\boldsymbol{\varphi}_r(z)((\mathcal{A} - r)h)(z)m'(z) \,\textrm{d}z,\end{align}

which hold for functions $f,h \in C^2(I)$ such that $\lim_{x \downarrow \mathfrak{l}}f(x)/\boldsymbol{\varphi}_r(x) = \lim_{x \uparrow \mathfrak{r}}h(x)/\boldsymbol{\psi}_r(x) =0$ and $(\mathcal{A} - r)f,\: (\mathcal{A} - r)h \in L^1_r(I)$ . Note that the last two conditions are equivalent to the integrability conditions of [Reference Alvarez2, Cor. 3.2]. On the other hand, the lower boundary $\mathfrak{l}$ was assumed to be natural so $\lim_{x \downarrow \mathfrak{l}}\boldsymbol{\varphi}_r(x) = \infty$ and $\lim_{x \downarrow \mathfrak{l}}f(x)/\boldsymbol{\varphi}_r(x) = 0$ may be used in the proof of [Reference Alvarez2, Corollary 3.2] instead of the condition $|f(0+)| < \infty$ .

2.2. The stopping problem

In this section we set up the optimal stopping problem. We start by recalling the Poisson stopping problem discussed in [Reference Lempa23]. Let N be a Poisson process with rate $\lambda > 0$ and assume that the filtration $\mathbb{F}$ carries information on both X and N. Define the discrete-time filtration $\mathbb{G}_0 = (\mathcal{G}_n )_{n \geq 0}$ as $\mathcal{G}_n = \mathcal{F}_{T_n}$ , where $T_0 = 0$ and $(T_n - T_{n-1}) \sim \textrm{Exp}(\lambda)$ for all $n \in \mathbb{N}$ . We call the times $T_n$ arrival times because they model the arrival of new stopping opportunities. Define the admissible stopping times $\mathcal{S}_0$ for a Poisson stopping problem as mappings $\tau\colon \Omega \rightarrow \{ T_n\colon n \geq 0 \}$ such that

  1. (i) for all $x \in I, \mathbb{E}_x[\tau ] < \infty;$

  2. (ii) for all $n \geq 0, \{\tau = T_n \} \in \mathcal{G}_n.$

Then the value function of the Poisson stopping problem is defined as

(7) \begin{align}V_0^{\lambda}(x) &= \sup\nolimits_{\tau \in \mathcal{S}_0} \mathbb{E}_x\left[ \textrm{e}^{-r\tau}g(X_{\tau}) \right],\end{align}

where the function g is the payoff. This Poisson stopping problem can be solved under mild assumptions and the value function (7) can be written in a semi-explicit form in terms of the minimal r-excessive functions; see [Reference Lempa23].

We extend the idea above to accommodate rate controls. Let $0 < \lambda_1 < \lambda_2$ and denote by $N^a$ , where $a=1,2$ , a Poisson process with rate $\lambda_i$ . Denote the arrival times of $N^a$ as $T^a_n$ , $a=1,2$ . We assume that the filtration $\mathbb{F}$ is rich enough to carry the processes X, $N^1$ and $N^2$ , and that X, $N^1$ and $N^2$ are independent. Now the agent has two arrival processes at her disposal and can switch between these over time. The switching decisions are expressed as sequences of random variables $A=(A_n)_{n \geq 0}$ ; we call such A attention sequences. We define attention sequences inductively as follows. Let $A_0$ be $\mathcal{F}_0$ -measurable such that $A_0(\omega)\in\{ \lambda_1, \lambda_2 \}$ for all $\omega \in \Omega$ . The value of $A_0$ indicates which arrival process the agent is following at time 0. Let $T^A_{0}=0$ and define

\begin{align*} T^A_{n} = T^A_{{n-1}} + U^{A_{n-1}}, \ (U^{A_{n-1}}| A_{n-1} &= \lambda_i ) \sim {\textrm{Exp}}(\lambda_i) \quad \text{for }n \geq 1.\end{align*}

The variables $U^{A_{m}}$ model the time remaining until the next arrival, conditional on the agent’s mth choice. Using this, we define $A_n$ as a $\mathcal{F}_{T^A_{n}-}$ -measurable random variable taking values in $\{ \lambda_1,\lambda_2 \}$ for $n \geq 1$ . These variables indicate which arrival process the agent chooses to follow at the nth time of choosing. If $A_n = \lambda_a$ for a given n, we say that the agent in maintaining $\lambda_a$ attention, $a=1,2$ .

Define the discrete time filtration $\mathbb{G}^A_0 = (\mathcal{G}^A_n)_{n \geq 0}$ recursively by

\begin{align*}\mathcal{G}^A_n = \mathcal{G}_{n-1}^A \bigvee \mathcal{F}_{T_{n-1}^A + U^{A_{n-1}}},\qquad \mathcal{G}^A_0 = \mathcal{F}_0.\end{align*}

For a fixed sequence A, we define the admissible stopping times as mappings $\tau\colon \Omega \rightarrow \{ T^A_n\colon n \geq 0 \}$ such that

  1. (i) for all $x \in I,\mathbb{E}_x[\tau ] < \infty$ ;

  2. (ii) for all $n \geq 0, \{\tau = T^A_n \} \in \mathcal{G}_n^A.$

Let $\mathcal{S}_0^A$ be the class of stopping times that are admissible with respect to the attention sequence A. We define an auxiliary stopping problem as

\begin{align*} V_0^A(x) = \sup\nolimits_{\tau \in \mathcal{S}_0^A} \mathbb{E}_x\left[ \textrm{e}^{-r\tau}g(X_\tau) - \sum_{n=0}^\infty \textrm{e}^{-rT^A_n}c(X_{T^A_n})\textbf{1}( A_n = \lambda_2) \textbf{1}( T^A_n < \tau ) \right],\end{align*}

where g is the exercise payoff and c is the period cost function. The function c is the cost the agent must pay in order to wait for a single arrival with the higher rate $\lambda_2$ . Now our main stopping problem reads as

(8) \begin{align} V_0(x) = \sup\nolimits_A V_0^A(x).\end{align}

We also set up constrained stopping problems, where immediate stopping is not possible; this corresponds to the situation where immediate observation of X is not available but the future observations occur according to our model. To this end, we define two sets of stopping times as follows. Let $\mathcal{S}^A_a$ , where $a=1,2$ , be the set of mappings $\tau\colon \Omega\rightarrow \{ T^A_n\colon n \geq 0 \}$ such that

  1. (i) for all $x \in I, \mathbb{E}_x[\tau ] < \infty;$

  2. (ii) for all $n \geq 1, \{\tau = T^A_n \} \in \mathcal{G}_n^A;$

  3. (iii) $\mathbb{P}(\tau = T^A_0) = 0,\: A_0 = \lambda_a.$

We define, for a fixed A with $A_0= \lambda_a$ , the auxiliary stopping problem

\begin{align*} V_a^A(x) = \sup\nolimits_{\tau \in \mathcal{S}_a^A} \mathbb{E}_x\left[ \textrm{e}^{-r\tau}g(X_\tau) - \sum_{n=0}^\infty \textrm{e}^{-rT^A_n}c(X_{T^A_n})\textbf{1}( A_n = \lambda_2 ) \textbf{1}( T^A_n < \tau ) \right].\end{align*}

Thus, the version of the main stopping problem where immediate stopping is not allowed reads as

(9) \begin{align} V_a(x) = \sup\nolimits_A V_a^A(x), \ a \in \{1,2\}.\end{align}

The function (9) can be interpreted as the value of the future stopping potentiality contingent on the level of attention and the state of diffusion X. This function is related to the value of the stopping problem in between the arrival times. In our proposed model we assume that the state of X is not observed between the arrivals. However, this does not prohibit us from devising this function and using it in our analysis. In the following, we refer to this function as the value of future stopping potentiality.

We study the problems (8) and (9) under the following assumptions.

Assumption 1. Let

\begin{align*} & p(x) = \lambda_2(\textbf{R}_rc)(x) + c(x), \quad P(x) = g(x) + p(x), \\ & y^{\lambda_2,c} = \inf \bigg\{ y \in I\colon \dfrac{P(y)}{\boldsymbol{\psi}(y)} > \dfrac{(\boldsymbol{\Phi}_{r+\lambda_2}P)(y)}{(\boldsymbol{\Phi}_{r+\lambda_2} \boldsymbol{\psi}_r)(y)} \bigg\},\\ &l_g = \inf \{ x \in I\colon g(x) > 0 \}.\end{align*}

We assume the following assertions hold.

  1. (i) The period cost c is non-negative, non-decreasing and r excessive. Furthermore, $c \in L^1_r(I) \cap C^2(I)$ , $(r - \mathcal{A})c \in L^1_r(I)$ , $\lim_{x \downarrow \mathfrak{l}}p(x)/\boldsymbol{\varphi}_r(x) = \lim_{x \uparrow \mathfrak{r}}p(x)/\boldsymbol{\psi}_r(x) = 0$ and $p/\boldsymbol{\psi_r}$ is strictly decreasing.

  2. (ii) The payoff g is non-negative, non-decreasing and piecewise $C^2$ . The left and right derivatives $d^\pm g(x)/dx$ , $d^{2\pm} g(x)/dx^2$ are assumed to be finite for all $x \in I$ . Furthermore, $g \in L^1_{r}(I) \cap C^0(I)$ , $(r - \mathcal{A})g \in L^1_r(I)$ , $\lim_{x \downarrow \mathfrak{l}}g(x)/\boldsymbol{\varphi}_r(x) = \lim_{x \uparrow \mathfrak{r}}g(x)/\boldsymbol{\psi}_r(x) = 0$ and there exists a unique state

    $$y^{\infty}= \textrm{argmax}_{x \in (l_g, \mathfrak{r})}\left\{ g(x)/\boldsymbol{\psi}_r(x) \right\}$$
    such that $g/\boldsymbol{\psi}_r$ is non-decreasing on $(l_g , y^{\infty})$ and non-increasing on $(y^{\infty}, \mathfrak{r})$ .
  3. (iii) The function ${(g - \lambda_2(\textbf{R}_{r+\lambda_2}g) + c)}/{\boldsymbol{\varphi}_{r+\lambda_2}}$ is strictly increasing on $(y^{\lambda_2,c}, \mathfrak{r})$ . Moreover, the function ${(g - \lambda_1(\textbf{R}_{r+\lambda_1}g))}/{\boldsymbol{\varphi}_{r+\lambda_1}}$ is strictly increasing on $(y^{\lambda_1}, \mathfrak{r})$ and

    $$\frac{P(x)}{\boldsymbol{\psi}_r(x)} < \frac{(\boldsymbol{\Phi}_{r+\lambda_2}P)(x)}{(\boldsymbol{\Phi}_{r+\lambda_2} \boldsymbol{\psi}_r)(x)} \quad{for\ }x \in (l_g, y^{\lambda_2,c}).$$

Remark 1. Assumption 1(i) and 1(ii) contain relevant regularity and monotonicity assumptions on the period cost and the payoff. It is interesting to note that r excessivity of the cost function c follows from the other assumptions for the examples we consider in Section 6. In particular, positive constants are trivially r excessive and if $c(x) = kx$ for some $k>0$ , the diffusion is a GBM and the payoff is $g(x) = x^\theta - \eta$ with $0<\theta < 1 < \eta$ , then r excessivity of c follows from $(r - \mathcal{A})g \in L^1_r(I)$ . The specific reason we require the cost to be r excessive is that in certain points of the proofs in Appendix A we encounter integrals that need to be of a certain sign. These signs behave as they should if $(\mathcal{A} - r)c(x) \leq 0$ for all $x \in I$ , i.e. if c is r excessive.

On the other hand, in Assumption 1(iii) the monotonicity assumptions are related to the classical stopping problem studied, for example, in [Reference Alvarez1]. These are typical assumptions that guarantee that a unique point splits the state space to the continuation region and the stopping region. Together with the limit assumptions for $g/\boldsymbol{\psi_r}$ , these correspond to the assumptions made in [Reference Lempa23] for the Poisson stopping problem.

Assumption 1 (iii) is connected to two Poisson stopping problems: one with the attention rate $\lambda_1$ and one with the attention rate $\lambda_2$ and period cost c. From a technical point of view it is needed to prove uniqueness of the optimal solution in our case. We also observe that Assumption 1(iii) is closely connected to typical superharmonicity assumptions for twice continuously differentiable payoffs. More precisely, assume for a moment that $g \in C^2(I)$ and that there exists a unique $\hat{x} \in I$ such that $((\mathcal{A} - r)g)(x) \lesseqqgtr 0$ , when $x \gtreqqless \hat{x}$ (see [Reference Alvarez3, Reference Lempa22]). These type of conditions imply the monotonicity conditions in Assumption 1(ii) and are easily verified. In our case, making use of (5) and (6), we find that

\begin{align*}&\frac{P(x) - \lambda_2(\textbf{R}_{r+\lambda_2}P)(x)}{\boldsymbol{\varphi}_{r+\lambda_2}(x)} \\&\qquad = B_{r+\lambda_2}^{-1}\left( (\boldsymbol{\Psi}_{r+\lambda_2}(r - \mathcal{A})P)(x) + \frac{\boldsymbol{\varphi}_{r+\lambda_2}(x)}{\boldsymbol{\psi}_{r+\lambda_2}(x)}(\boldsymbol{\Phi}_{r+\lambda_2}(r - \mathcal{A})P)(x) \right),\end{align*}

which is strictly increasing if and only if $(\boldsymbol{\Phi}_{r+\lambda_2}(\mathcal{A} - r)P)(x) < 0$ . Also,

$$\dfrac{P(y)}{\boldsymbol{\psi}_r(y)} < \dfrac{(\boldsymbol{\Phi}_{r+\lambda_2}P)(y)}{(\boldsymbol{\Phi}_{r+\lambda_2}\boldsymbol{\psi}_r)(y)}$$

if and only if

\begin{align*} \frac{\boldsymbol{\psi}_r(y)}{\boldsymbol{\varphi}_{r+\lambda_2}(y)}(\boldsymbol{\Phi}_{r+\lambda_2}(\mathcal{A} - r)P)(y) + (\boldsymbol{\Psi}_r(\mathcal{A} - r)P)(y) &> 0.\end{align*}

Hence, it is evident that, for twice continuously differentiable payoffs, Assumption 1(iii) is closely connected to the sign of $(\mathcal{A} - r)g$ .

We use the standard Poisson stopping problem (7) as a point of reference in our study. Indeed, we know from [Reference Lempa23] and Remark 1 that under Assumption 1, the value function in (7) reads as

(10) \begin{equation} V_0^{\lambda}(x) = \begin{cases} g(x), & x \geq y^{\lambda}, \\ \displaystyle\frac{g(y^{\lambda})}{\boldsymbol{\psi}_r(y^{\lambda})}\boldsymbol{\psi}_r(x), & x \leq y^{\lambda}, \end{cases}\end{equation}

where the level $y^{\lambda} < y^{\infty}$ is characterized uniquely by the condition

\begin{align*} g(y^{\lambda})(\boldsymbol{\Phi}_{r+\lambda}\boldsymbol{\psi}_r)(y^{\lambda}) = \boldsymbol{\psi}_r(y^{\lambda})(\boldsymbol{\Phi}_{r+\lambda}g)(y^{\lambda}).\end{align*}

We close this section with certain technical calculations that will be useful in the later sections.

Lemma 1. Let $\lambda >0$ , and $x \in I$ . Then

  1. 1.

    \begin{align*}\\[-28pt]& \frac{1}{B_r} \left( (\boldsymbol{\Psi}_{r+\lambda}\boldsymbol{\varphi}_{r})(x)(\boldsymbol{\Psi}_r(r - \mathcal{A})p)(x) + (\boldsymbol{\Psi}_{r+\lambda}\boldsymbol{\psi}_{r})(x)(\boldsymbol{\Phi}_r(r - \mathcal{A})p)(x) \right)\\&\qquad= (\boldsymbol{\Psi}_{r+\lambda}p)(x) + \lambda^{-1}(\boldsymbol{\Psi}_{r+\lambda}(r - \mathcal{A})p)(x),\\& \frac{1}{B_r} \left( (\boldsymbol{\Phi}_{r+\lambda}\boldsymbol{\varphi}_{r})(x)(\boldsymbol{\Psi}_r(r - \mathcal{A})p)(x) + (\boldsymbol{\Phi}_{r+\lambda}\boldsymbol{\psi}_{r})(x)(\boldsymbol{\Phi}_r(r - \mathcal{A})p)(x) \right)\\&\qquad= (\boldsymbol{\Phi}_{r+\lambda}p)(x) + \lambda^{-1}(\boldsymbol{\Phi}_{r+\lambda}(r - \mathcal{A})p)(x);\end{align*}
  2. 2.

    \begin{equation*}\: \left\lvert \begin{matrix}p &\boldsymbol{\psi}_r &\boldsymbol{\varphi}_r\\(\boldsymbol{\Phi}_{r+\lambda}p) &(\boldsymbol{\Phi}_{r+\lambda}\boldsymbol{\psi}_r) &(\boldsymbol{\Phi}_{r+\lambda}\boldsymbol{\varphi}_r)\\(\boldsymbol{\Psi}_{r+\lambda}p) &(\boldsymbol{\Psi}_{r+\lambda}\boldsymbol{\psi}_r) &(\boldsymbol{\Psi}_{r+\lambda}\boldsymbol{\varphi}_r)\end{matrix} \right\rvert (x) = \frac{B_rB_{r+\lambda}}{\lambda^2}\left( c(x) + (\lambda_2 - \lambda)(\textbf{R}_rc)(x) \right).\end{equation*}

Proof. See Appendix A.

3. Necessary conditions

For brevity, we use from here on the following notation for the relevant resolvents, fundamental solutions and functionals:

\begin{align*} &R_a := \textbf{R}_{r+\lambda_a}, \quad R_0 := \textbf{R}_r, \quad \psi_a := \boldsymbol{\psi}_{r+\lambda_a}, \quad \varphi_a := \boldsymbol{\varphi}_{r+\lambda_a}, \quad \psi_0 := \boldsymbol{\psi}_r, \quad \varphi_0 := \boldsymbol{\varphi}_r, \\ &\Psi_a := \boldsymbol{\Psi}_{r+\lambda_a}, \quad \Phi_a := \boldsymbol{\Phi}_{r+\lambda_a}, \quad \Psi_0 := \boldsymbol{\Psi}_r, \quad \Phi_0 := \boldsymbol{\Phi}_r,\end{align*}

for $a=1,2$ .

The purpose of this section is to derive a solution candidate for the problem (8) under Assumption 1. Since the diffusion X is time homogeneous and the functions g and c are non-decreasing, we make the working assumption that the optimal policy is a threshold rule with two constant thresholds $x^*$ and $y^*$ . The policy is as follows.

(A1) Assume that the agent is maintaining $\lambda_1$ attention. Let $\hat{T}^1$ be the next $N^1$ -arrival time. If $X_{\hat{T}^1}< x^*$ then the agent maintains the $\lambda_1$ rate. If $X_{\hat{T}^1} \geq x^* $ then the agent either switches to rate $\lambda_2$ or stops. Stopping occurs if $X_{\hat{T}^1} \geq y^*$ .

(A2) Assume that the agent is maintaining $\lambda_2$ attention. Let $\hat{T}^2$ be the next $N^2$ -arrival time. If $X_{\hat{T}^2} < x^*$ then the agent switches to the $\lambda_1$ rate. If $X_{\hat{T}^2}\geq x^*$ then the agent either maintains the $\lambda_2$ rate or stops. Stopping occurs if $X_{\hat{T}^2} \geq y^*$ .

If $x^*>y^*$ then the main problem (8) reduces to the Poisson stopping problem (7). We assume in the following that $x^*<y^*$ unless otherwise stated.

Next we construct the value function candidate associated with the strategy described in (A1) and (A2). Denote as $G_1$ the associated candidate value of future stopping potentiality while maintaining $\lambda_1$ attention; function $G_2$ has an analogous definition. Moreover, we denote as $G_0$ the associated candidate value at the arrival times. We let $A^* = (A^*_n)_{n \geq 0}$ and $\tau^*$ denote the attention sequence and the stopping time associated with the strategy described in (A1) and (A2) when the problem is started at an arrival time. Note that $A^*_n$ and $\tau^*$ are measurable since their values are determined by collections of sets of the form $\{X_{T_m} \in B_m\} \subseteq \Omega$ , where $m \geq 0$ and $B_m$ are measurable subsets of the state space I. Looking at the right-hand side of (8), we get an explicit expression for $G_0$ :

(11) \begin{align} G_0(x) &= \mathbb{E}_x\left[ \textrm{e}^{-r\tau^*}g(X_{\tau^*}) - \sum_{n=0}^\infty \textrm{e}^{-rT^{A^*}_n}c(X_{T^{A^*}_n})\textbf{1}( A^*_n = \lambda_2 ) \textbf{1}( T^{A^*}_n < \tau^* ) \right].\end{align}

Expressions for $G_1$ and $G_2$ may be obtained analogously. Supposing further that the dynamic programming principle (DPP) holds for $G_0$ and making use of the memoryless property of the exponential distribution, we get

(12) \begin{equation}\begin{split} G_0(x) &= \max\{ g(x),G_1(x),G_2(x) \},\\ G_1(x) &= \mathbb{E}_x\left[\textrm{e}^{-rU_1}G_0(X_{U_{1}}) \right] ,\\ G_2(x) &= \mathbb{E}_x\left[\textrm{e}^{-rU_2}G_0(X_{U_{2}}) \right] - c(x),\end{split}\end{equation}

where independent random variables $U_a \sim {\textrm{Exp}}(\lambda_a)$ for $a = 1,2$ . It should be noted that in this section the DPP (12) is used to derive certain necessary conditions (described by the pair of equations (22)), that an optimal strategy conforming to (A1) and (A2) must satisfy. In Proposition 1 we prove that there exist three uniquely determined functions satisfying the DPP (12) and the mentioned necessary conditions. In Section 4 we prove that these three functions coincide with the three value functions defined in (8) and (9), verifying the optimality of the strategy outlined in (A1) and (A2).

We may write the candidates $G_1,G_2$ alternatively as

(13) \begin{equation} \begin{split} G_1(x) &= G_0(x) + \mathbb{E}_x\left[\textrm{e}^{-rU_1}G_0(X_{U_1}) \right] - G_0(x),\\ G_2(x) &= G_0(x) + \mathbb{E}_x\left[\textrm{e}^{-rU_2}G_0(X_{U_2}) \right] - c(x) - G_0(x). \end{split}\end{equation}

Equations (13) are mathematically trivial in relation to (12), but their significance is to highlight the relations between the candidates conceptually rather than technically. Indeed, keeping in mind the DPP of (12), condition (13) states that, for $i=1,2$ , the candidate $G_i$ is obtained from $G_0$ by subtracting the loss incurred from being forced to initially wait for an arrival with attention rate $\lambda_i$ . We also have the following equivalence result for the smoothness properties of $G_0$ , $G_1$ and $G_2$ .

Lemma 2. Suppose that Assumption 1 holds and the infinitesimal parameter $\sigma^2$ satisfies $\sigma^2 \in C^0(I)$ . Then the conditions $G_1 \in C^2(I)$ and $G_2 \in C^2(I)$ are equivalent to $G_0 \in C^0(I)$ .

Proof. Recall that $c \in C^2(I)$ by our assumptions and note that it follows from (1) and (12) that

\begin{align*} G_1(x) &= \mathbb{E}_x\left[\textrm{e}^{-rU_1}G_0(X_{U_{1}}) \right] = \lambda_1 (R_1G_0)(x),\\ G_2(x) &= \mathbb{E}_x\left[\textrm{e}^{-rU_2}G_0(X_{U_{2}}) \right] - c(x) = \lambda_2 (R_2G_0)(x) - c(x).\end{align*}

Representation (2) implies that the mappings $(R_iG_0), i\in \{ 1, 2 \}$ , are continuously differentiable so that $G_1,G_2 \in C^1(I)$ . Calculating the second derivatives gives

\begin{align*} G^{\prime\prime}_1(x) &= \dfrac{\lambda_1}{B_{1}}\left( \varphi^{\prime\prime}_1(x)\left( \Psi_{1}G_0 \right) (x) + \psi^{\prime\prime}_1(x)\left( \Phi_{1}G_0 \right) (x) \right) - \dfrac{2\lambda_1}{\sigma^2(x)}G_0(x),\\ G^{\prime\prime}_2(x) &= \dfrac{\lambda_2}{B_{2}}\left( \varphi^{\prime\prime}_2(x)\left( \Psi_{2}G_0 \right) (x) + \psi^{\prime\prime}_2(x)\left( \Phi_{2}G_0 \right) (x) \right) - \dfrac{2\lambda_2}{\sigma^2(x)}G_0(x) - c''(x), \end{align*}

proving the claim.

The rest of this section is devoted to finding an analytical rather than probabilistic expression for $G_0$ and deriving necessary optimality conditions for $x^*$ and $y^*$ . By (A1) and (12), we find that $G_0(x)=G_1(x) = \lambda_1(R_{1}G_0)(x)$ for $x < x^*$ . Now [Reference Lempa23, Lemma 2.1] implies that $G_1$ is r-harmonic for $x < x^*$ . We find similarly that the function $x \mapsto G_0(x)+p(x)$ is r harmonic on $[x^*,y^*)$ . Since we are looking for a candidate that is finite at the lower boundary $\mathfrak{l}$ , we find that $G_0(x) = C_1\psi_0(x)$ for $x < x^*$ and some fixed $C_1 \in \mathbb{R}_+$ . By (12) we have

\begin{align*} \lambda_1C_1(R_{1}\psi_0)(x) = C_1\psi_0(x) &= \lambda_1C_1(R_{1}G_0)(x)\quad \text{for }x < x^*.\end{align*}

Hence,

(14) \begin{align} C_1(\Phi_1\psi_0)(x^*) &= (\Phi_1G_0)(x^*). \end{align}

Let $x \in [x^*,y^*)$ . Then $G_0(x) + p(x) = C_2\psi_0(x) + C_3\varphi_0(x)$ for some $C_2, C_3 \in \mathbb{R}_+$ . We can now expand condition (14) as

\begin{align*} C_1(\Phi_1\psi_0)(x^*) =& \int_{x^*}^{y^*}\varphi_{1}(z)(C_2\psi_0(z) + C_3\varphi_0(z) - p(z))m'(z)\,\textrm{d}z + (\Phi_1g)(y^*).\end{align*}

Conversely, the condition

\begin{align*} \lambda_2(R_{2}(C_2\psi_0 + C_3\varphi_0))(x) &= C_2\psi_0(x) + C_3\varphi_0(x) \\ & = G_0(x) + p(x) \\ & = \lambda_2(R_{2}(G_0 + p))(x)\end{align*}

holds and it can be rewritten as

\begin{gather*} \varphi_{2}(x)(\Psi_2(C_2\psi_0 + C_3\varphi_0))(x^*)+ \psi_{2}(x)(\Phi_2(C_2\psi_0 + C_3\varphi_0))(y^*) \\ = \varphi_{2}(x)(\Psi_2(C_1\psi_0 + p))(x^*) + \psi_{2}(x)(\Phi_2(P))(y^*).\end{gather*}

Since the functions $\psi_2$ and $\varphi_2$ are linearly independent on open intervals, the previous equality yields the pair of equations

(15) \begin{gather} (\Phi_2(C_2\psi_0 + C_3\varphi_0)(y^*) = (\Phi_2P)(y^*),\nonumber\\ (\Psi_2(C_2\psi_0 + C_3\varphi_0)(x^*) = (\Psi_2(C_1\psi_0 + p))(x^*).\end{gather}

We see from the considerations above that the candidate $G_0$ contains a total of five unknown variables ( $C_1,C_2,C_3,x^*,y^*$ ) that need to be determined. Conditions (14) and (15) contain three equations and the last two are given by the boundary conditions imposed by the continuity of $G_0$ :

(16) \begin{gather} C_2\psi_0(x^*) + C_3\varphi_0(x^*) - p(x^*) = C_1\psi_0(x^*),\nonumber\\ C_2\psi_0(y^*) + C_3\varphi_0(y^*) = P(y^*).\end{gather}

The system of equations consisting of (14), (15) and (16) can be rewritten as

(17a) \begin{align} & C_{1} (\Phi_{1}\psi_{0})(x^{*}) + C_{2} ((\Phi_{1}\psi_{0})(y^{*}) - (\Phi_{1}\psi_{0})(x^{*}))\nonumber\\&\qquad +C_{3}((\Phi_{1}\varphi_{0})(y^{*}) - (\Phi_{1}\varphi_{0})(x^{*})) = (\Phi_{1}P)(y^{*}) - (\Phi_{1}p)(x^{*}),\end{align}
(17b) \begin{align} &C_{2} (\Phi_2\psi_{0})(y^{*}) + C_{3} (\Phi_{2}\varphi_{0})(y^{*}) = (\Phi_{2P})(y^{*}), \end{align}
(17c) \begin{align} &(C_{2} - C_{1})(\Phi_{2}\psi_{0})(x^{*}) + C_{3} (\Phi_{2}\varphi_{0})(x^{*}) = (\Phi_{2}p)(x^{*}),\end{align}
(17d) \begin{align} &C_{2}\psi_{0}(x^{*}) +C_{3}\varphi_{0}(x^{*}) - p(x^{*}) = C_{1}\psi_{0}(x^{*}),\end{align}
(17e) \begin{align} &C_{2}\psi_{0}(y^{*}) +C_{3}\varphi_{0}(y^{*}) = P(y^{*}).\end{align}

Next we reduce the system (17) to a pair of equations where the only variables are $x^*$ and $y^*$ . The constants $C_2$ and $C_3$ can be solved from (17b) and (17e), yielding

(18) \begin{align}\begin{split} C_3 &= \dfrac{(\Phi_2\psi_0)(y^*)P(y^*) - (\Phi_2P)(y^*) \psi_0(y^*)}{(\Phi_2\psi_0)(y^*) \varphi_0(y^*) - (\Phi_2\varphi_0)(y^*)\psi_0(y^*)}, \\ C_2 &= \dfrac{(\Phi_2P)(y^*) - (\Phi_2\varphi_0)(y^*)C_3}{(\Phi_2\psi_0)(y^*)}. \end{split}\end{align}

Solving for $C_1$ in (17d) gives

(19) \begin{align} C_1 &= C_2+\frac{\varphi_0(x^*)}{\psi_0(x^*)} C_3 - \frac{p(x^*)}{\psi_0(x^*)},\end{align}

and lastly solving $C_3$ from (17a) and (17c) yields the condition

(20) \begin{equation} H_1(x^*) = H_2(y^*) = f(x^*, y^*),\end{equation}

where

\begin{gather*} H_1(x) = \frac{p(x)(\Psi_2\psi_0)(x) - \psi_0(x)(\Psi_2p)(x)}{\varphi_0(x)(\Psi_2\psi_0)(x) - \psi_0(x)(\Psi_2\varphi_0)(x)}, \\[5pt] H_2(x) = \frac{P(x)(\Phi_2\psi_0)(x) - \psi_0(x)(\Phi_2P)(x)}{\varphi_0(x)(\Phi_2\psi_0)(x) - \psi_0(x)(\Phi_2\varphi_0)(x)}, \\[5pt] f(x,y) = \frac{({(\Phi_2P)(y)}/{(\Phi_2\psi_0)(y)})(\Phi_1\psi_0)(y) \!- \!(\Phi_1 P)(y) \!- \!({p(x)}/{\psi_0(x)})(\Phi_1\psi_0)(x) \!+ \!(\Phi_1p)(x)}{({(\Phi_2\varphi_0)(y)}/{(\Phi_2\psi_0)(y)})(\Phi_1\psi_0)(y)\! -\! (\Phi_1\varphi_0)(y)\! -\! ({\varphi_0(x)}/{\psi_0(x)})(\Phi_1\psi_0)(x)\! + \!(\Phi_1\varphi_0)(x)}.\end{gather*}

Even though the pair of equations (20) is nonlinear, it can be solved under our assumptions. We begin by deriving an additional constraint on the pair (20) that rules out degenerate cases, where the candidate $G_0$ reduces to the value function of the Poisson stopping problem (10).

Lemma 3. If $H_1(y^{\lambda_1}) \leq H_2(y^{\lambda_1})$ then $G_0 = V_0^{y^{\lambda_1}}$ and if $H_1(y^{\lambda_1}) > H_2(y^{\lambda_1})$ then the thresholds $x^*$ and $y^*$ satisfy $x^* < y^{\lambda_1} < y^*$ .

Proof. See Appendix A.

In light of Lemma 3, we now assume that

(21) \begin{equation} H_1(y^{\lambda_1}) > H_2(y^{\lambda_1})\end{equation}

for the remainder of this section. Despite its technical appearance, condition (21) is closely connected to the structure of the optimal strategy. We prove in Theorem 2 that choosing to wait for the next arrival time with $\lambda_2$ attention is suboptimal for all initial states $x \in I$ precisely when $H_1(y^{\lambda_1}) \leq H_2(y^{\lambda_1})$ . This is in line with the implication of Lemma 3 stating that the value function candidate $G_0$ coincides with the value function of the Poisson stopping problem (10) precisely when $H_1(y^{\lambda_1}) \leq H_2(y^{\lambda_1})$ .

To further simplify the pair (20), we can use the condition $H_1(x^*) = H_2(y^*)$ to separate the parts depending on $x^*$ and $y^*$ in the other equation. Then we end up with the equivalent necessary conditions

(22) \begin{align} & H_1(x^*) = H_2(y^*), \nonumber\\ & K_1(x^*) = K_2(y^*),\end{align}

where

\begin{align*} K_1(x) & = \frac{p(x)}{\psi_0(x)} (\Phi_1\psi_0)(x)-(\Phi_1p)(x) - H_1(x) \bigg(\frac{\varphi_0(x)}{\psi_0(x)} (\Phi_1\psi_0)(x)-(\Phi_1\varphi_0)(x) \bigg), \\ K_2(x) & = \frac{(\Phi_2P)(x)}{(\Phi_2\psi_0)(x)}(\Phi_1\psi_0)(x) -(\Phi_1 P)(x) \\ &\hskip11pt - H_2(x) \bigg( \frac{(\Phi_2\varphi_0)(x)}{(\Phi_2\psi_0)(x)}(\Phi_1\psi_0)(x) - (\Phi_1\varphi_0)(x) \bigg). \end{align*}

Remark 2. If $g \in C^2(I)$ , we can simplify the functions $H_i,K_i$ further by expressing them in terms of the generator $\mathcal{A}$ . Defining

\begin{equation*} F_2(x) = \frac{\varphi_0(x)(\Phi_2P)(x) - P(x)(\Phi_2\varphi_0)(x)}{\varphi_0(x)(\Phi_2\psi_0)(x) - \psi_0(x)(\Phi_2\varphi_0)(x)}\end{equation*}

we have the relations

\begin{align*} P(x) &= H_2(x)\varphi_0(x) + F_2(x)\psi_0(x),\\ (\Phi_2P)(x) &= H_2(x)(\Phi_2\varphi_0)(x) + F_2(x),(\Phi_2\psi_0)(x),\end{align*}

and, thus,

\begin{equation*} K_2(x) = \frac{P(x)}{\psi_0(x)}(\Phi_1\psi_0)(x) -(\Phi_1 P)(x) - H_2(x) \bigg( \frac{\varphi_0(x)}{\psi_0(x)}(\Phi_1\psi_0)(x) - (\Phi_1\varphi_0)(x) \bigg).\end{equation*}

Now we can use (5) and (6) to obtain

\begin{align*} H_1(x) &= \frac{1}{B_0}\left(- \frac{\psi_0(x)}{\psi_{2}(x)}(\Psi_2(r - \mathcal{A})p)(x) + (\Psi_0(r - \mathcal{A})p)(x) \right),\\ H_2(x) &= \frac{1}{B_0}\left( \frac{\psi_0(x)}{\varphi_{2}(x)}(\Phi_2(r - \mathcal{A})P)(x) + (\Psi_0(r - \mathcal{A})P)(x) \right),\\ K_1(x) &= \frac{1}{\lambda_1} \left( \frac{\varphi_1(x)}{\psi_2(x)}(\Psi_2(r - \mathcal{A})p)(x) + (\Phi_1(r - \mathcal{A})p)(x) \right), \\ K_2(x) &= \frac{1}{\lambda_1} \left( - \frac{\varphi_{1}(x)}{\varphi_{2}(x)}(\Phi_2(r - \mathcal{A})P)(x) + (\Phi_1(r - \mathcal{A})P)(x) \right).\end{align*}

Recalling that $p(x) = c(x) + \lambda_2(R_rc)(x)$ , we see that in this case the monotonicity properties of $H_i$ and $K_i$ can be easily determined by studying the behaviour of the functions $(r - \mathcal{A})p(x) = (r + \lambda_2 - \mathcal{A})c(x)$ and $(r - \mathcal{A})g(x)$ .

We now prove some auxiliary results regarding the functions $H_1$ , $H_2$ , $K_1$ and $K_2$ to study the solvability of the equation pair (22).

Lemma 4. The functions $H_i$ and $K_i$ have opposite signs of derivative everywhere for $i=1,2$ and the curves $H_1(x)$ and $H_2(x)$ intersect at least once in the interval $(y^{\lambda_1} \vee y^{\lambda_2,c}, y^{\lambda_2})$ . We denote the smallest such intersection point by $\hat{y}$ . Furthermore,

  1. (i) $H_1(x) > 0$ , $H_1'(x) > 0$ , $\lim_{x \downarrow \mathfrak{l}} H_1(x) = 0$ ;

  2. (ii) $K_1(x) > 0$ , $K_1'(x) < 0$ , $\lim_{x \downarrow \mathfrak{l}}K_1(x) = \infty$ ;

  3. (iii) $H_2$ is positive and strictly increasing on $(y^{\lambda_2,c}, \mathfrak{r})$ ;

  4. (iv) $K_2$ is positive and strictly decreasing on $(y^{\lambda_2,c}, \hat{y})$ ;

  5. (v) $K_2(\hat{y}) > K_1(\hat{y})$ .

Proof. See Appendix A.

The reader is advised to refer to Figure 1 in order to discern how the functions $H_i,K_i$ will look under our assumptions in a typical example.

Figure 1. A prototype plot of the functions $H_1$ , $H_2$ , $K_1$ , $K_2$ and the thresholds $x^*$ , $y^*$ , $y^{\lambda_1}$ , $y^{\lambda_2}$ .

We can now prove our main proposition about the existence and uniqueness of solutions to the pair of equations (22).

Proposition 1. Let Assumption 1 hold and suppose that $H_1(y^{\lambda_1}) > H_2(y^{\lambda_1})$ . Then the pair (22) has a unique solution $(x^*,y^*)$ such that $x^* \in (\mathfrak{l}, \hat{y}), y^* \in (y^{\lambda_2,c}, \hat{y})$ , and $x^* < y^{\lambda_1} < y^*$ .

Proof. See Appendix A.

To close the section, we collect our findings on the solution candidate to the next proposition. The purpose of Proposition 2 is to show that there exist three uniquely determined functions that satisfy the necessary conditions described by the equation pair (22) and a DPP of the form (12). In Section 4 we prove that the three functions given by Proposition 2 coincide with the three value functions defined in (8) and (9).

Proposition 2. Let Assumption 1 hold and suppose that $H_1(y^{\lambda_1}) > H_2(y^{\lambda_1})$ . Then the functions

(23) \begin{align}\begin{split} \mathcal{K}_0(x) &= \begin{cases} g(x), & x \geq y^*, \\ C_2\psi_0(x) + C_3\varphi_0(x) - c(x) - \lambda_2 (R_0c)(x), & x^* \leq x < y^*, \\ C_1\psi_0(x), & x < x^*, \end{cases}\\ \mathcal{K}_1(x) &= \lambda_1(R_1\mathcal{K}_0)(x),\\ \mathcal{K}_2(x) &= \lambda_2(R_2\mathcal{K}_0)(x) - c(x), \end{split}\end{align}

where the constants $C_2,C_3$ are given by (18), the constant $C_1$ is given by (19), and the thresholds $x^*$ and $y^*$ are given by Proposition 1, satisfy the DPP

(24) \begin{equation}\begin{split} \mathcal{K}_0(x) &= \max\{ g(x),\mathcal{K}_1(x),\mathcal{K}_2(x) \},\\ \mathcal{K}_1(x) &= \mathbb{E}_x\left[\textrm{e}^{-rU_1}\mathcal{K}_0(X_{U_{1}}) \right] ,\\ \mathcal{K}_2(x) &= \mathbb{E}_x\left[\textrm{e}^{-rU_2}\mathcal{K}_0(X_{U_{2}}) \right] - c(x),\end{split}\end{equation}

where independent random variables $U_a \sim {\textrm{Exp}}(\lambda_a)$ for $a = 1,2$ . Moreover, the function $\mathcal{K}_0$ is continuous and the functions $\mathcal{K}_1$ and $\mathcal{K}_2$ are in $C^2(I)$ if and only if $\sigma^2 \in C^0(I)$ .

Proof. See Appendix A.

4. Sufficient conditions

The purpose of this section is to prove the following theorem, which is our main result on the stopping problem (8).

Theorem 1. Let Assumption 1 hold and suppose that $H_1(y^{\lambda_1}) > H_2(y^{\lambda_1})$ . Let $(x^*,y^*)$ be the solution to (22) given by Proposition 1. Then the values (8) and (9) read as $V_0 = \mathcal{K}_0$ , $V_1=\mathcal{K}_1$ and $V_2=\mathcal{K}_2$ , where the functions $\mathcal{K}_0$ , $\mathcal{K}_1$ and $\mathcal{K}_2$ are given in Proposition 2. The optimal policy can be described as follows.

  1. (a) Assume that the agent is maintaining $\lambda_1$ attention If the process is in the region $[y^*,\mathfrak{r})$ at the next $\lambda_1$ arrival, stop. If the process is in the region $[x^*,y^*)$ at the next $\lambda_1$ arrival, switch to $\lambda_2$ attention. If the process is in the region $(\mathfrak{l},x^*)$ at the next $\lambda_1$ arrival, maintain $\lambda_1$ -attention.

  2. (b) Assume that the agent is maintaining $\lambda_2$ attention. If the process is in the region $[y^*,\mathfrak{r})$ at the next $\lambda_2$ arrival, stop. If the process is in the region $[x^*,y^*)$ at the next $\lambda_2$ arrival, maintain $\lambda_2$ attention. If the process is in the region $(\mathfrak{l},x^*)$ at the next $\lambda_2$ arrival, switch to $\lambda_1$ attention.

Remark 3. Recall that $V_0^{\lambda_2}$ is the value of the Poisson stopping problem (7) with arrival rate $\lambda_2$ and the representation (10) holds under Assumption 1. The cost c is a nonnegative function so it follows from the first paragraph in the proof of Lemma 3 that (using the proof’s notation)

\begin{align*} \mathcal{K}_0(x) &= V_0(x) \leq \sup\nolimits_{A} \sup\nolimits_{\tau \in \mathcal{S}_0^A} \mathbb{E}_x\left[ e^{-r\tau}g(X_\tau) \right] = G_{0,0}(x) = V_0^{\lambda_2}(x).\end{align*}

We need the following result on uniform integrability to prove Theorem 1.

Lemma 5. For a fixed attention sequence A, the process

\begin{align*} S^A &= (S_n^A; \ \mathcal{G}^A_n)_{n\geq0},\\ S^A_n &= \textrm{e}^{-rT_n^A}\mathcal{K}_0(X_{T_n^A})-\sum_{n=0}^\infty \textrm{e}^{-rT^A_n}c(X_{T^A_n})\textbf{1}(A_n = \lambda_2)\textbf{1}(T^A_n < \tau)\end{align*}

is a uniformly integrable supermartingale.

Proof. Fix the attention sequence A and let $U_a \sim \textrm{Exp}(\lambda_a)$ , $a \in \{ 1, 2 \}$ . Then

\begin{align*} &\mathcal{K}_0(x) \geq \max\{ \mathcal{K}_1(x),\mathcal{K}_2(x) \} \\ &\qquad= \mathcal{K}_1(x)\textbf{1}(A_0=1) + \mathcal{K}_2(x)\textbf{1}(A_0=2) \\ &\qquad = \mathbb{E}_x\left[ \textrm{e}^{-rU_{1}}\mathcal{K}_0(X_{U_{1}})\textbf{1}(A_0=1) + \left(\textrm{e}^{-rU_{2}}\mathcal{K}_0(X_{U_{2}}) - c(x) \right)\textbf{1}(A_0=2) \right].\end{align*}

Hence, $S^A$ has the claimed supermartingale property.

To prove uniform integrability, define the process

\begin{align*} L^A &= \left(L_n^A;\mathcal{G}_n^A \right)_{n\geq0},\qquad L_n^A = \textrm{e}^{-rT_n^A} \frac{\psi_0(X_{T_n^A})}{\psi_0(x)}\end{align*}

Since the boundary $\mathfrak{r}$ is natural, we find that

\begin{align*} \mathbb{E}_x[L^A_1] = \frac{1}{\psi_0(x)}\mathbb{E}_x\left[ \textrm{e}^{-rU_{2}}\psi_0(X_{U_{2}})\textbf{1}(A_0=2)+\textrm{e}^{-rU_{1}}\psi_0(X_{U_{1}})\textbf{1}(A_0=1)\right] =1.\end{align*}

This implies that $\mathbb{E}_x[L^A_n] = 1$ for all n. Define now the measure

\begin{align*} \mathbb{P}^A_{x,n}(B) = \mathbb{E}_x[L_n^A\textbf{1}(B)] \quad\text{for }B \in \mathcal{F}.\end{align*}

By Remark 3 we find that

(25) \begin{align} 0 &\leq \frac{\mathbb{E}_x[S_n^A\textbf{1}(B)]}{\psi_0(x)} \nonumber\\ &\leq \mathbb{E}_x\left[ \frac{\mathcal{K}_0(X_{T_n^A})}{\psi_0(X_{T_n^A})}L_n^A\textbf{1}(B)\right] \nonumber\\ &\leq \mathbb{E}_x\left[ \frac{g(y^{\lambda_2})}{\psi_0(y^{\lambda_2})} L_n^A \textbf{1}(B)\textbf{1}(X_{T_n^A}\leq y^{\lambda_2}) \right] + \mathbb{E}_x\left[ \frac{g(X_{T_n^A})}{\psi_0(X_{T_n^A})} L_n^A \textbf{1}(B)\textbf{1}(X_{T_n^A}\geq y^{\lambda_2}) \right] \nonumber\\ & \leq \frac{g(y^{\infty})}{\psi_0(y^\infty)}\mathbb{P}^A_{x,n}(B)\end{align}

for all $B \in \mathcal{F}$ . By taking $B = \Omega$ , we find that $\sup_n \mathbb{E}_x[S_n^A] < \infty$ . On the other hand, we observe that if $\mathbb{P}_x(B) \rightarrow 0$ then $\mathbb{P}^A_{x,n}(B) \rightarrow 0$ for all $n \geq 0$ . Thus, inequality (25) implies that $\sup_n \mathbb{E}_x[S_n^A\textbf{1}(B)] \rightarrow 0$ as $\mathbb{P}_x(B) \rightarrow 0$ . This yields the claimed uniform integrability.

Proof of Theorem 1. Fix the attention sequence A and let $U_a \sim \textrm{Exp}(\lambda_a)$ , $a \in \{ 1, 2 \}$ . Then optional sampling implies that

\begin{align*} \mathcal{K}_0(x) &\geq \mathbb{E}_x\left[ \textrm{e}^{-r\tau} \mathcal{K}_0(X_{\tau}) - \sum_{n=0}^\infty \textrm{e}^{-rT^A_n}c(X_{T^A_n})\textbf{1}(A_n = \lambda_2)\textbf{1}(T^A_n < \tau) \right] \\ & \geq \mathbb{E}_x\left[ \textrm{e}^{-r\tau} g(X_{\tau}) - \sum_{n=0}^\infty \textrm{e}^{-rT^A_n}c(X_{T^A_n})\textbf{1}(A_n = \lambda_2)\textbf{1}(T^A_n < \tau) \right] \quad\text{for all }\tau \in S^A_0.\end{align*}

This implies that $\mathcal{K}_0 \geq V^A_0$ for all A and, consequently, $\mathcal{K}_0 \geq V_0$ .

To prove the opposite inequality, we consider the admissible stopping/switching rule that produced the candidate $\mathcal{K}_0$ . Denote the associated (random) attention sequence as $A^*$ and the stopping time as $\tau^* \in \mathcal{S}_0^{A^*}$ . To prove the results, we show that the stopped process

\begin{align*} Q &= (Q_n ; \ \mathcal{G}^{A^*}_{n} )_{n\geq0}, \\ Q_n &= \textrm{e}^{-r(\tau^*\land T_n)}\mathcal{K}_0(X_{\tau^*\land T_n})-\sum_{n=0}^\infty \textrm{e}^{-r T^{A^*}_n}c(X_{T^{A^*}_n})\textbf{1}(A^*_n = \lambda_2)\textbf{1}(T^{A^*}_n < \tau^*\land T_n)\end{align*}

is a martingale. To this end, let $n\geq1$ . Then

\begin{align*} \mathbb{E}_x\left[ Q_n \ | \ \mathcal{G}_{n-1}^{A^*} \right] = \mathbb{E}_x\left[ S_n^{A^*}\textbf{1}(\tau^* \geq T^{A^*}_n) \ | \ \mathcal{G}_{n-1}^{A^*} \right] + \sum_{k=0}^{n-1}S_k^{A^*}\textbf{1}(\tau^* = T^{A^*}_k).\end{align*}

Consider the first term on right-hand side: by the strong Markov property and the dynamic programming equation (24), we find that

\begin{align*} \mathbb{E}&_x\left[ S_n^{A^*}\textbf{1}(\tau^*\geq T^{A^*}_n) \ | \ \mathcal{G}_{n-1}^{A^*} \right] \\ &= \left\{\textrm{e}^{-rT_{n-1}^{A^*}}\mathbb{E}_{X_{T_{n-1}^{A^*}}}\left[\vphantom{\int} \textrm{e}^{-rU_{1}}\mathcal{K}_0(X_{U_{1}})\textbf{1}(A^*_{n-1}=1) \right. \right. \\ & \qquad + \left. \left(\textrm{e}^{-rU_{2}}\mathcal{K}_0(X_{U_{2}}) - c(X_{T_{n-1}^{A^*}})\right)\textbf{1}(A^*_{n-1}=2)\right] \\ & \qquad \qquad - \left. \sum_{k=0}^{n-1} \textrm{e}^{-rT^{A^*}_k}c(X_{T^{A^*}_k})\textbf{1}(A^*_k = \lambda_2) \right\}\textbf{1}(\tau^* \geq T_n^{A^*}) \\ &= \left\{ \textrm{e}^{-rT_{n-1}^{A^*}} \left(\vphantom{\int}\mathcal{K}_1(X_{T_{n-1}^{A^*}})\textbf{1}(A^*_{n-1}=1) + \mathcal{K}_2(X_{T_{n-1}^{A^*}})\textbf{1}(A^*_{n-1}=2)\right) \right. \\ & \qquad - \left. \sum_{k=0}^{n-1} \textrm{e}^{-rT^{A^*}_k}c(X_{T^{A^*}_k})\textbf{1}(A^*_k = \lambda_2) \right\}\textbf{1}(\tau^* \geq T_n^{A^*}).\end{align*}

Thus, we obtain

\begin{align*} \mathbb{E}_x[Q_n|\mathcal{G}_{n-1}^{A^*}] = S_{n-1}^{A^*} \textbf{1}(\tau^* \geq T_n^{A^*}) + \sum_{k=0}^{n-1} S_k^{A^*}\textbf{1}(\tau^* = T^{A^*}_k) = Q_{n-1}\end{align*}

and optional sampling on Q with respect to $\tau^*$ gives

\begin{align*} \mathcal{K}_0(x) &= \mathbb{E}_x\left[ \textrm{e}^{-r\tau^*} \mathcal{K}_0(X_{\tau^*}) - \sum_{n=0}^\infty \textrm{e}^{-rT^{A^*}_n}c(X_{T^{A^*}_n})\textbf{1}(A^*_n = \lambda_2)\textbf{1}(T^{A^*}_n < \tau^*) \right] \\ & = \mathbb{E}_x\left[ \textrm{e}^{-r\tau^*} g(X_{\tau^*}) - \sum_{n=0}^\infty \textrm{e}^{-rT^{A^*}_n}c(X_{T^{A^*}_n})\textbf{1}(A^*_n = \lambda_2)\textbf{1}(T^{A^*}_n < \tau^*) \right].\end{align*}

On the other hand, $A^*$ is an admissible attention sequence and $\tau^* \in \mathcal{S}_0^{A^*}$ so we have the inequality $\mathcal{K}_0 \leq \sup\nolimits_{A} V_0^A = V_0$ , proving that $\mathcal{K}_0 = V_0$ .

Next we prove $\mathcal{K}_1 = V_1$ and $\mathcal{K}_2 = V_2$ . Conditions (24) and optional sampling imply that, for a fixed A,

\begin{align*} \mathcal{K}_1(x) &= \lambda_1(R_{1}\mathcal{K}_0)(x)\\ &\geq \mathbb{E}_x\left[\textrm{e}^{-r\tau} g(X_{\tau}) - \sum_{n=0}^\infty \textrm{e}^{-rT^{A}_n}c(X_{T^{A}_n})\textbf{1}(A_n = \lambda_2)\textbf{1}(T^A_n < \tau) \right]\quad\text{for all }\tau \in \mathcal{S}^A_1.\end{align*}

Consequently $\mathcal{K}_1 \geq V_1$ . To prove the opposite inequality, let $\tau^*$ and $A^*$ be as before. Recall that if $\tau$ and A are an admissible stopping time and an attention sequence for $V_0$ , then $U_1 + \tau \circ \theta_{U_1}$ and (1, A) are admissible for $V_1$ . Here $\theta$ denotes the usual shift operator that acts on the process X as $\theta_{U_1} (X_t) = X_{U_1 + t}$ . Let $\tau^1 = U_1 + \tau^* \circ \theta_{U_1}$ and $A^1 = (1,A^*)$ . Conditioning on the first arrival time and using the strong Markov property yields

\begin{align*} \mathcal{K}_1(x) &= \mathbb{E}_x\left[ \textrm{e}^{-rU_1} \mathbb{E}_{X_{U_1}} \left[ \textrm{e}^{-r\tau^*}g(X_{\tau^*})-\sum_{n=0}^\infty \textrm{e}^{-rT^{A^*}_n}c(X_{T^{A^*}_n})\textbf{1}(A^*_n = \lambda_2)\textbf{1}(T^{A^*}_n < \tau^*) \right] \right] \\ &= \mathbb{E}_x \left[ \textrm{e}^{-r\tau^1} g(X_{\tau^1}) - \sum_{n=0}^\infty \textrm{e}^{-rT^{A^1}_n}c(X_{T^{A^1}_n})\textbf{1}(A^1_n = \lambda_2)\textbf{1}(T^{A^1}_n < \tau^1) \right],\end{align*}

and since the stopping time and attention sequence associated with $\mathcal{K}_1$ are admissible for $V_1$ , we have $\mathcal{K}_1 \leq \sup\nolimits_{A} V_1^A = V_1$ and, hence, $\mathcal{K}_1 = V_1$ . The proof of $\mathcal{K}_2 = V_2$ is completely analogous.

The following theorem presents an easily verifiable condition determining whether the main problem (8) reduces to the Poisson stopping problem (10). It also gives further clarification for the assumption $H_1(y^{\lambda_1}) > H_2(y^{\lambda_1})$ made in Section 3. There the condition was shown to be connected to the order of the thresholds $x^*$ and $y^*$ . Here the proof of Theorem 2 indicates that $H_1(y^{\lambda_1}) \leq H_2(y^{\lambda_1})$ if and only if waiting with the higher attention rate is suboptimal everywhere in the state space I.

Theorem 2. The function $V_0 = V_0^{\lambda_1}$ if and only if $\lambda_2(R_{2}(V_0^{\lambda_1} + p))(y^{\lambda_1}) \leq V_0^{\lambda_1}(y^{\lambda_1}) + p(y^{\lambda_1})$ and the inequality is equivalent to $H_1(y^{\lambda_1}) \leq H_2(y^{\lambda_1})$ , where $H_i$ are given by (22).

Proof. Let $V_0 = V_0^{\lambda_1}$ . Waiting with $\lambda_2$ attention is always suboptimal so

\begin{align*} \lambda_2(R_{2}(V_0^{\lambda_1} + p))(x) &\leq V_0^{\lambda_1}(x) + p(x) \quad\text{for all }x \in I. \end{align*}

Now suppose that the inequality holds for $x = y^{\lambda_1}$ . Noting that $a_2(x) \varphi_{2}(x) = - b(x)\psi_{2}(x)$ for all $x \in I$ , we get

\begin{align*} & \lambda_2(R_{2}(V_0^{\lambda_1} + p))(y^{\lambda_1}) \leq V_0^{\lambda_1}(y^{\lambda_1}) + p(y^{\lambda_1}) \\&\qquad\Leftrightarrow \dfrac{\lambda_2}{B_{2}} \left( \varphi_{2} \left( \Psi_{2}\left( \dfrac{g(y^{\lambda_1})}{\psi_0(y^{\lambda_1})} +p \right) \right) + \psi_{2}(\Phi_{2}P) \right) (y^{\lambda_1}) \leq P(y^{\lambda_1}) \\ &\qquad\Leftrightarrow \left( \varphi_{2} (\Psi_{2}p) + \psi_{2}(\Phi_{2}P) \right) (y^{\lambda_1}) \leq \left( \dfrac{B_{2}}{\lambda_2}P - \dfrac{g}{\psi_0}(\Psi_{2}\psi_0)\varphi_{2} \right) (y^{\lambda_1}) \\ &\qquad\Leftrightarrow \left( \dfrac{p}{\psi_0} (\Psi_2\psi_r) - (\Psi_2p) \right) (y^{\lambda_1}) \varphi_{2}(y^{\lambda_1}) \geq \left( (\Phi_{2}P) - \dfrac{P}{\psi_{0}} (\Phi_{2}\psi_0) \right) (y^{\lambda_1}) \psi_{2}(y^{\lambda_1}) \\ &\qquad\Leftrightarrow H_1(y^{\lambda_1}) \leq H_2(y^{\lambda_1}). \end{align*}

Now Lemma 3 implies that $V_0^{\lambda_1}$ satisfies the DPP

\begin{align*} V_0^{\lambda_1} &= \max \{g, \lambda_1(R_{1}V_0^{\lambda_1}), \lambda_2(R_{2}(V_0^{\lambda_1} + p)) - p\}. \end{align*}

We can employ the same arguments as in the proof of Lemma 5 and Theorem 1 to show that $V_0^{\lambda_1} \geq V_0^A$ for any fixed attention sequence A and, consequently, that $V_0^{\lambda_1} \geq V_0$ . But the optimal stopping rule corresponding to $V_0^{\lambda_1}$ is admissible for $V_0$ as well, so $V_0^{\lambda_1} \leq V_0$ and, thus, $V_0^{\lambda_1} = V_0$ .

5. A note on asymptotics

In this section we briefly discuss various asymptotic properties of the stopping problem (8). We recall the proof of Lemma 3, where it was stated that if c is the cost function of the main problem (8) and $\alpha \geq 0$ , then $\alpha c$ is the cost function of another similar problem. It was also proven that in the vanishing period cost limit, i.e. when $\alpha = 0$ (or c is identically zero), we have $V_0 = V_0^{\lambda_2}$ meaning that the problem (8) reduces to the usual Poisson stopping problem (7) with attention rate $\lambda_2$ . On the other hand, if c is not identically zero then there exists a critical value $\alpha_c$ for the scaling coefficient $\alpha$ , such that the problem (8) reduces to the Poisson stopping problem (7) with attention rate $\lambda_1$ . A natural interpretation for these reduction results is that if choosing the attention rate $\lambda_2$ results in a nonzero period cost, then it is optimal to choose the smaller but free attention rate $\lambda_1$ far away from the stopping region. In this case removing the period cost outweighs the opportunity cost that results from choosing the smaller attention rate. On the other hand, if the period cost is sufficiently large, the cost of choosing the higher attention rate $\lambda_2$ will outweigh its benefits.

It is also interesting to study the asymptotic properties of the problem with respect to the attention rates. Recalling the analysis of Sections 3 and 4, we have by monotone convergence $\lim_{\lambda_1 \downarrow 0}\mathcal{K}_1 = 0$ and $\lim_{\lambda_2 \downarrow 0}\mathcal{K}_2 = -c$ . Thus, $\lim_{\lambda_1 \downarrow 0}V_0 = V_0^{\lambda_2,c}$ and $\lim_{\lambda_2 \downarrow 0}V_0 = V_0^{\lambda_1}$ , where $V_0^{\lambda_2,c} = \max \{ g, \mathcal{K}_2 \}$ . Here $V_0^{\lambda_2,c}$ may also be shown to be the value function of a Poisson stopping problem with an attention rate $\lambda_2$ and a cost function c by utilizing the verification arguments of Section 4. We also have $\lim_{\lambda_1 \uparrow \lambda_2}V_0 = V_0^{\lambda_2}$ since at the limit the agent can choose to stop or to continue with either a free or a costly attention rate $\lambda_2$ , the last option being always suboptimal.

In order to determine the limit $\lim_{\lambda_1 \uparrow \infty} V_0$ we have to remove the assumption $\lambda_1 < \lambda_2$ . Let $\tilde{V}$ be the value function of the optimal stopping problem in continuous time, i.e.

(26) \begin{align} \tilde{V}(x) &= \sup\nolimits_{\tau \in \mathcal{T}}\mathbb{E}_x\left[\textrm{e}^{-r\tau}g(X_{\tau})1_{\{ \tau < \infty \} } \right],\end{align}

where $x \in I$ and $\mathcal{T}$ is the set of stopping times of X. Here, $V_0 \geq \max \{ V_0^{\lambda_1}, V_0^{\lambda_2,c} \}$ and Assumption 1, Remark 1 and Proposition 2.6 in [Reference Lempa23] imply that $\lim_{\lambda_1 \uparrow \infty} V_0^{\lambda_1} = \tilde{V}$ and $V_0^{\lambda_2,c} \leq V_0^{\lambda_2} \leq \tilde{V}$ , so $\lim_{\lambda_1 \uparrow \infty} V_0 \geq \tilde{V}$ . But $\tilde{V}$ is an r-excessive majorant of g, so $V_0 \leq \tilde{V}$ for all positive $\lambda_1$ . Thus, $\lim_{\lambda_1 \uparrow \infty} V_0 = \tilde{V}$ . The result is intuitively clear, since letting $\lambda_1 \uparrow \infty$ allows the agent to get rid of the period cost and the Poisson constraint at the same time.

Determining the last limit $\lim_{\lambda_2 \uparrow \infty}V_0$ is not so straightforward. At the limit, an agent may either choose to stop, continue with free Poisson arrivals having an attention rate $\lambda_1$ , or continue in continuous time while paying a running cost c. This is a completely different type of problem since the optimal stopping strategy may be a mixture of Poisson and continuous-time stopping strategies. As such, the problem lies beyond the scope of the present paper and is therefore left for future research.

6. Illustration

We illustrate the general results obtained in the previous sections with two examples. In the first example the underlying is taken to be a GBM and the period cost is assumed to be proportional to the state of the process. In the second example we study a logistic diffusion process with a fixed period cost.

6.1. Geometric Brownian motion and proportional period cost

In this example we assume that the underlying is a GBM, the payoff is $g(x) = x^{\theta}-\eta$ , where $0 < \theta < 1 < \eta$ , and the period cost is $c(x) = kx$ where $k>0$ . The infinitesimal generator of the diffusion reads as

\begin{align*} \mathcal{A} = \frac{1}{2}\sigma^2 x^2\frac{d^2}{dx^2} + \mu x \frac{d}{dx},\end{align*}

where $\mu \in \mathbb{R}_+$ and $\sigma \in \mathbb{R}_+$ are given constants. we make the additional assumption that $r > \mu$ . Furthermore, the scale density and the density of the speed measure read as

$$S'(x) = x^{-{2 \mu}/{\sigma^2}}, \qquad m'(x) = \frac{2}{\sigma^2} x^{({2 \mu}/{\sigma^2})-2}.$$

Denote

\begin{align*}& \beta_{i} = \frac{1}{2}-\frac{\mu}{\sigma^2}+\sqrt{\bigg(\frac{1}{2}-\frac{\mu}{\sigma^2} \bigg)^2+\frac{2(r+\lambda_i)}{\sigma^2}}, \\& \alpha_{i} = \frac{1}{2}-\frac{\mu}{\sigma^2} - \sqrt{\bigg(\frac{1}{2}-\frac{\mu}{\sigma^2} \bigg)^2+\frac{2(r+\lambda_i)}{\sigma^2}}.\end{align*}

Then the minimal r-excessive functions for X read as

$$\psi_{i}(x) = x^{\beta_i}, \qquad \varphi_{i}(x) = x^{\alpha_i}.$$

It is worth emphasizing that now $\beta_i > 1 > \theta > 0 > \alpha_i$ and it is a straightforward exercise to show that the conditions of Assumption 1 are satisfied.

Since the payoff is twice continuously differentiable everywhere, we can write the pair of equations as in Remark 2. The auxiliary functionals

\begin{align*}& (\Phi_i (r+\lambda_2 - \mathcal{A})c)(x) = \frac{2k(r+\lambda_2 - \mu)}{\sigma^2 (\beta_i - 1)} x^{1 - \beta_i},\\& (\Psi_i (r+\lambda_2 - \mathcal{A})c)(x) = \frac{2k(r+\lambda_2 - \mu)}{\sigma^2 (1 - \alpha_{i})} x^{1 - \alpha_{i}}, \\& (\Phi_i(r - \mathcal{A})g)(x) = \frac{2}{\sigma^2} \left( \frac{r - \mu \theta - \frac{\sigma^2}{2}\theta (\theta - 1)}{\beta_i - \theta}x^{\theta - \beta_i} - \frac{r\eta}{\beta_i - 1}x^{-\beta_i} \right), \\& (\Psi_i(r - \mathcal{A})g)(x) = \frac{2}{\sigma^2} \left( \frac{r - \mu \theta - \frac{\sigma^2}{2}\theta (\theta - 1)}{\theta - \alpha}x^{\theta - \alpha_i} - \frac{r\eta}{1 - \alpha_i}x^{-\alpha_i} \right)\end{align*}

can be used to write an explicit form for the equation pair.

Next we investigate the condition $H_1(y^{\lambda_1}) > H_2(y^{\lambda_1})$ that guarantees that our solution does not reduce to a Poisson stopping problem with rate $\lambda_1$ . We know from Theorem 2 that this condition is equivalent to $\lambda_2 (R_{2} (V_0^{\lambda_1} + p)(y^{\lambda_1})\geq g(y^{\lambda_1}) + p(y^{\lambda_1})$ . Expanding the resolvent with the help of (2) and using the resolvent identity (4) on the period cost terms, we see that the stated condition is equivalent to

\begin{align*} \frac{\lambda_2}{B_2}\psi_0(y^{\lambda_1})\left( \Phi_2\left( g - \frac{g(y^{\lambda_1})}{\psi_0(y^{\lambda_1})}\psi_0 \right) \right) (y^{\lambda_1}) \geq c(y^{\lambda_1}),\end{align*}

where $y^{\lambda_1}$ is the unique root of the equation $\psi_0(y^{\lambda_1})(\Phi_1g)(y^{\lambda_1}) = g(y^{\lambda_1})(\Phi_1\psi_0)(y^{\lambda_1})$ and it can be calculated to be

$$y^{\lambda_1} = \left( \dfrac{\beta_0(\beta_1-\theta)}{\beta_1(\beta_0-\theta)} \eta \right) ^{1/\theta}.$$

If the proportional period cost parameter k exceeds the critical value

\begin{align*} k^* &= \frac{\lambda_2}{B_2y^{\lambda_1}}\psi_0(y^{\lambda_1})\left( \Phi_2\left( g - \frac{g(y^{\lambda_1})}{\psi_0(y^{\lambda_1})}\psi_0 \right) \right) (y^{\lambda_1})\\ &= \frac{2\lambda_2 \eta \beta_0(\beta_2 - \beta_1)}{B_2\sigma^2 y^{\lambda_1}\beta_1 \beta_2 (\beta_2 - \beta_0)(\beta_2 - \theta)},\end{align*}

then $V_0 = V_0^{\lambda_1}$ and the problem reduces to the standard Poisson stopping problem. The value $k^*$ has a natural interpretation in terms of the cost parameters $\eta$ and k. Since $\theta \in (0,1)$ , $\eta > 1$ and $k^*$ is always positive when $\lambda_1 < \lambda_2$ , we find that increasing $\eta$ or k enough will reduce the problem to Poisson stopping when other parameters are fixed. Indeed, increasing either cost parameter results in a smaller realized payoff when the process is stopped.

Figure 2. Optimal thresholds as a function of attention rate $\lambda_1$ . The other parameters are chosen to be $\mu = 0.02$ , $\sigma = \sqrt{0.3}$ , $r = 0.05$ , $k = 0.001$ , $\lambda_2 = 4.0$ , $\theta = 0.67$ , $\eta = 1.5$ .

Regarding the attention rates, we find that if $\lambda_1 = \lambda_2$ then $k^*=0$ and the problem reduces to the standard Poisson stopping problem. Likewise, because $y^{\lambda_1}$ is increasing in $\lambda_1$ , we see that increasing $\lambda_1$ will decrease $k^*$ and make the agent less likely to swap to $\lambda_2$ . This is again in line with the intuition that it is never optimal to swap to $\lambda_2$ and pay the cost c(x) if the rate $\lambda_1$ is high enough. These observations are further illustrated in Figures 2 and 3. In Figure 2 we see that there exists a critical threshold $\hat{\lambda}_1$ for the attention rate $\lambda_1$ , where eventually the optimal thresholds $x^*$ , $y^*$ and $y^{\lambda}_1$ are the same (in this case about 2.6). When $\lambda_1 < \hat{\lambda}_1$ , the optimal strategy is to switch to higher attention rate, but when $\lambda_1 > \hat{\lambda}_1$ , switching is always suboptimal. In Figure 3 we see a rather similar situation. In this case if $\lambda_2$ is below the critical threshold $\hat{\lambda}_2$ (about 1.15 in the figure), it is suboptimal to switch, but when $\lambda_2$ is above the threshold, we see that switching is optimal.

Figure 3. Optimal thresholds as a function of attention rate $\lambda_2$ . The other parameters are chosen to be $\mu = 0.02$ , $\sigma = \sqrt{0.3}$ , $r = 0.05$ , $k = 0.001$ , $\lambda_1 = 1.0$ , $\theta = 0.67$ , $\eta = 1.5$ .

The effects of r, $\mu$ and $\sigma$ are not as direct and straightforward to see, but differentiations show that their effects are not linear, and instead usually for high and low values of these parameters the problem tends to reduce to Poisson stopping with attention rate $\lambda_1$ .

6.2. Logistic diffusion

Let the underlying follow a standard logistic diffusion driven by the stochastic differential equation

$$dX_t=\mu X_t (1- \gamma X_t)dt +\sigma X_t dW_t,\qquad X_0=x \in \mathbb{R_+},$$

where $\mu, \gamma, \sigma > 0$ . The scale density and the density of the speed measure read as

$$S'(x)= x^{-{2\mu}/{\sigma^2}} \textrm{e}^{{2\mu\gamma}x/{\sigma^2}} , \qquad m'(x)=\frac{2}{\sigma^2} x^{({2\mu}/{\sigma^2})-2} \textrm{e}^{-{2\mu\gamma}x/{\sigma^2}}.$$

The minimal r-excessive functions are

$$\varphi_{r}(x) = x^\alpha U \big( \alpha, 1+\alpha-\beta, \frac{2\mu\gamma}{\sigma^2}x \big),\qquad\psi_{r}(x) = x^\alpha M \big( \alpha, 1+\alpha-\beta, \frac{2\mu\gamma}{\sigma^2}x \big),$$

where M and U denote, respectively, the confluent hypergeometric functions of first and second kind and

\begin{align*}& \alpha = \frac{1}{2}-\frac{\mu}{\sigma^2} + \sqrt{\bigg(\frac{1}{2}-\frac{\mu}{\sigma^2} \bigg)^2+\frac{2r}{\sigma^2}}, \\& \beta = \frac{1}{2}-\frac{\mu}{\sigma^2}-\sqrt{\bigg(\frac{1}{2}-\frac{\mu}{\sigma^2} \bigg)^2+\frac{2r}{\sigma^2}}.\end{align*}

We choose the payoff to be linear $g(x) = x-\eta, \, \eta \in \mathbb{R}^+$ and the cost to be constant $c(x) = k \in \mathbb{R}^+$ .

Unfortunately, due to the complicated nature of the minimal r-excessive functions in this example, the integral functionals such as $(\Psi_{r+\lambda_l} \psi_0)(x)$ and $(\Phi_{r+\lambda_l} P)(x)$ cannot be calculated explicitly. Consequently, the pair of equations for the optimal thresholds cannot be simplified from their original integral forms in any meaningful way, and are thus left unstated. Hence, we demonstrate the results numerically.

In Table 1 we demonstrate how the optimal thresholds behave when the attention rates $\lambda_1$ and $\lambda_2$ are changed. We observe that increasing the rate $\lambda_2$ increases the stopping threshold $y^*$ , but interestingly the switching threshold $x^*$ can increase or decrease. In Table 2 the effect of increasing the constant cost of higher attention rate is studied. As is intuitively clear, increasing the cost postpones the switching decision by increasing threshold $x^*$ .

Table 1. Optimal thresholds with various attention rates. The other parameters are chosen to be $\mu = 0.01$ , $\sigma = \sqrt{0.1}$ , $r = 0.05$ , $\gamma = 0.5$ , $k = 0.01$ and $\eta = 2.0$ .

Table 2. Optimal thresholds with various costs for higher attention rate. The other parameters are chosen to be $\mu = 0.01$ , $\sigma = \sqrt{0.1}$ , $r = 0.05$ , $\gamma = 0.5$ , $\lambda_1 = 0.1$ , $\lambda_2 = 1.0$ and $\eta = 2.0$ .

Appendix A. Proofs of technical results

A.1. Proof of Lemma 1

  1. (i) We present a proof for the first condition, the second one is completely analogous. Using (5) and (6) yields

    \begin{align*} &\frac{1}{B_r} \left( (\boldsymbol{\Psi}_{r+\lambda}\boldsymbol{\varphi}_{r})(x)(\boldsymbol{\Psi}_r(r - \mathcal{A})p)(x) + (\boldsymbol{\Psi}_{r+\lambda}\boldsymbol{\psi}_{r})(x)(\boldsymbol{\Phi}_r(r - \mathcal{A})p)(x) \right)\\ &\qquad= \frac{1}{\lambda B_r S'(x)^2}(\boldsymbol{\psi}^{\prime}_{r}(x)\boldsymbol{\varphi}_r(x) - \boldsymbol{\psi}_r(x)\boldsymbol{\varphi}^{\prime}_{r}(x))(\boldsymbol{\psi}_{r+ \lambda}'(x)p(x) - \boldsymbol{\psi}_{r+ \lambda}(x)p'(x))\\ &\qquad= (\boldsymbol{\Psi}_{r+\lambda}p)(x) + \lambda^{-1}(\boldsymbol{\Psi}_{r+\lambda}(r - \mathcal{A})p)(x).\end{align*}
  2. (ii) A straightforward calculation yields

    \begin{align*} & \left\lvert \begin{matrix} p &\boldsymbol{\psi}_r &\boldsymbol{\varphi}_r\\ (\boldsymbol{\Phi}_{r+\lambda}p) &(\boldsymbol{\Phi}_{r+\lambda}\boldsymbol{\psi}_r) &(\boldsymbol{\Phi}_{r+\lambda}\boldsymbol{\varphi}_r)\\ (\boldsymbol{\Psi}_{r+\lambda}p) &(\boldsymbol{\Psi}_{r+\lambda}\boldsymbol{\psi}_r) &(\boldsymbol{\Psi}_{r+\lambda}\boldsymbol{\varphi}_r) \end{matrix} \right\rvert (x) \\ &\qquad= \frac{p(x)}{\lambda^2S'(x)^2}\left( (\boldsymbol{\psi}^{\prime}_{r}\boldsymbol{\varphi}_r - \boldsymbol{\psi}_r\boldsymbol{\varphi}^{\prime}_{r})(\boldsymbol{\psi}_{r+\lambda}' \boldsymbol{\varphi}_{r+\lambda} - \boldsymbol{\psi}_{r+\lambda}\boldsymbol{\varphi}_{r+\lambda}') \right)(x) \\ &\hskip29pt- \left(\left( \frac{(\boldsymbol{\Phi}_{r+\lambda}p)}{\lambda S'} \boldsymbol{\psi}_{r+\lambda} + \frac{(\boldsymbol{\Psi}_{r+\lambda}p)}{\lambda S'} \boldsymbol{\psi}_{r+\lambda} \right) (\boldsymbol{\psi}^{\prime}_{r}\boldsymbol{\varphi}_r - \boldsymbol{\psi}_r\boldsymbol{\varphi}^{\prime}_{r})\right)(x) \\ &\qquad= \frac{B_rB_{r+\lambda}}{\lambda^2}\left( p(x) - \lambda(\textbf{R}_{r+\lambda}p)(x) \right) \\ &\qquad= \frac{B_rB_{r+\lambda}}{\lambda^2}\left( c(x) + \lambda_2(\textbf{R}_{r}c)(x) - \lambda(\textbf{R}_{r+\lambda}c)(x) \right) \\ &\qquad- \frac{B_rB_{r+\lambda}}{\lambda^2} \lambda \lambda_2(\textbf{R}_{r+\lambda}\textbf{R}_rc)(x) \\ &\qquad= \frac{B_rB_{r+\lambda}}{\lambda^2}\left( c(x) + (\lambda_2 - \lambda) (\textbf{R}_{r}c)(x) \right), \end{align*}

    where the first equality follows by using (5) and (6), and the second equality is obtained by recalling the representations (2) and (3) for the resolvent and the Wronskian. The third equality follows from the definition of p and the last step is achieved by applying the resolvent equation (4). In particular, we may use (5) and (6) on $(\boldsymbol{\Psi}_{r+\lambda}\boldsymbol{\psi}_r)$ and $(\boldsymbol{\Phi}_{r+\lambda}\boldsymbol{\varphi}_r)$ because

    $$\lim_{x \downarrow \mathfrak{l}} \frac{\boldsymbol{\varphi}_r(x)}{\boldsymbol{\varphi}_{r+\lambda}(x)} = \lim_{x \uparrow \mathfrak{r}} \frac{\boldsymbol{\psi}_r(x)}{\boldsymbol{\psi}_{r+\lambda}(x)} = 0$$
    by the proof of Lemma 2.1 in [Reference Lempa23]. The said proof contains the result $\lim_{x \uparrow \mathfrak{r}} {\boldsymbol{\psi}_r(x)}/{\boldsymbol{\psi}_{r+\lambda}(x)} = 0$ for the choice $\mathfrak{r} = \infty$ , but proving the condition for some other $\mathfrak{r} \in \mathbb{R}$ or the condition $\lim_{x \downarrow \mathfrak{l}} {\boldsymbol{\varphi}_r(x)}/{\boldsymbol{\varphi}_{r+\lambda}(x)} = 0$ is completely analogous.

A.2. Proof of Lemma 3

We introduce a nonnegative scaling coefficient on the cost function in order to prove the result. Let $\alpha \geq 0$ . Since $c \in L^1_r(I)$ , we also have that $\alpha c \in L^1_r(I)$ . Let

\begin{align*} G_{2,\alpha}(x) &= \mathbb{E}_x\left[\textrm{e}^{-rU^{\lambda_2}}G_{0, \alpha}(X_{U^{\lambda_2}}) \right] - \alpha c(x), \\ G_{0,\alpha}(x) &= \max\{ g(x),G_1(x),G_{2,\alpha}(x) \}, \end{align*}

and let $x^*_{\alpha}, y^*_{\alpha}$ be the corresponding thresholds. Here $G_{0, \alpha}$ is continuous with respect to $\alpha$ so $x^*_{\alpha}$ and $y^*_{\alpha}$ are continuous functions of $\alpha$ as well. We naturally have

\begin{align*} G_{0,0}(x) &= \max \{ \max \{g(x), \lambda_1(R_1G_{0,0})(x) \}, \max \{g(x), \lambda_2(R_2G_{0,0})(x) \} \}\quad\text{for all }x \in I. \end{align*}

Under Assumption 1, the Poisson stopping problem (10) satisfies the DPP $V_0^{\lambda_1} = \max \{g, \lambda_1(R_1G_{0,0}) \}$ (see Remark 1 and the proof of [Reference Lempa23, Theorem 1.1]). Thus, $G_{0,0} = \max \{V_0^{\lambda_1}, V_0^{\lambda_2} \}$ . On the other hand, we assumed that $\lambda_1 < \lambda_2$ so that $g(y^{\lambda_1})/\psi_0(y^{\lambda_1}) < g(y^{\lambda_2})/\psi_0(y^{\lambda_2})$ and, consequently, $V_0^{\lambda_1} \leq V_0^{\lambda_2}$ . Thus, $G_{0,0} = V_0^{\lambda_2}$ , meaning that $\lim_{\alpha \downarrow 0}x^*_\alpha =l_g$ and $y^*_0 = y^{\lambda_2}$ .

Next we show that there exists a critical value $\alpha_c > 0$ for the scale such that $G_{0,\alpha} = V_0^{\lambda_1}$ for $\alpha \geq \alpha_c$ and $x^*_{\alpha} < y^{\lambda_1} < y^*_{\alpha}$ for $\alpha < \alpha_c$ . To see this, note that, for all $x \in I$ , $G_{2,\alpha}(x)$ is strictly decreasing with respect to $\alpha$ with $\lim_{\alpha \uparrow \infty}G_{2,\alpha}(x) = - \infty$ . Consequently, $x^*_{\alpha}$ is strictly increasing and $y^*_{\alpha}$ is strictly decreasing with respect to $\alpha$ and $\lim_{\alpha \uparrow \infty}G_{0,\alpha} = V_0^{\lambda_1}$ .

Consider now a critically scaled (i.e. the cost function is $\alpha_cc$ ) version of the pair (20). We know that $x^*_{\alpha_c}=y^{\lambda_1}=y^*_{\alpha_c}$ and it is an easy exercise to show that $H_2(y^{\lambda_1}) = f(y^{\lambda_1},y^{\lambda_1})$ . The remaining equation $H_1(y^{\lambda_1}) = H_2(y^{\lambda_1})$ yields an expression for the critical scale,

\begin{align*} \alpha_c &= \dfrac{ \dfrac{g(y^{\lambda_1})(\Phi_2\psi_0)(y^{\lambda_1}) - \psi_0(y^{\lambda_1})(\Phi_2g)(y^{\lambda_1})}{\varphi_0(y^{\lambda_1})(\Phi_2\psi_0)(y^{\lambda_1}) - \psi_0(y^{\lambda_1})(\Phi_2\varphi_0)(y^{\lambda_1})} }{ H_1(y^{\lambda_1}) - \dfrac{p(y^{\lambda_1})(\Phi_2\psi_0)(y^{\lambda_1}) - \psi_0(y^{\lambda_1})(\Phi_2p)(y^{\lambda_1})}{\varphi_0(y^{\lambda_1})(\Phi_2\psi_0)(y^{\lambda_1}) - \psi_0(y^{\lambda_1})(\Phi_2\varphi_0)(y^{\lambda_1})}}, \end{align*}

from which we can conclude that

\begin{align*} \alpha_c &\gtrless 1 \Leftrightarrow H_1(y^{\lambda_1}) \gtrless H_2(y^{\lambda_1}). \end{align*}

The result follows since $G_0 = G_{0,1}, x^* = x^*_1$ and $y^* = y^*_1$ .

A.3. Proof of Lemma 4

We may write the condition (21) as $H_1(y^{\lambda_1}) > H_2(y^{\lambda_1})$ and if $y^{\lambda_1} < y^{\lambda_2,c}$ then $H_2(y^{\lambda_2,c}) = 0 < H_1(y^{\lambda_2,c})$ . On the other hand, a direct calculation shows that $H_1(y^{\lambda_2}) - H_2(y^{\lambda_2}) < 0$ if and only if

(27) \begin{equation} \left\lvert \begin{matrix} p &\psi_0 &\varphi_0\\ (\Phi_{2}p) &(\Phi_{2}\psi_0) &(\Phi_{2}\varphi_0)\\ (\Psi_{2}p) &(\Psi_{2}\psi_0) &(\Psi_{2}\varphi_0) \end{matrix} \right\rvert (y^{\lambda_2}) > 0,\end{equation}

which follows from Lemma 1(ii). Recalling that $y^{\lambda_2,c} \leq y^{\lambda_2}$ , we see that $H_i$ ( $i=1,2$ ) have opposite order at $y^{\lambda_1} \vee y^{\lambda_2,c}$ and $y^{\lambda_2}$ . Consequently, by continuity the curves $H_1(x)$ and $H_2(x)$ must intersect at least once in the interval $(y^{\lambda_1} \vee y^{\lambda_2,c}, y^{\lambda_2})$ . Next we prove parts (i)–(v) of Lemma 4.

Part (i): $H_1$ is positive everywhere in I since $p/\psi_0$ and $\varphi_0/\psi_0$ are strictly decreasing. Standard differentiation and conditions $\lambda_2(R_{2}\psi_0)(x)=\psi_0(x)$ and $\lambda_2(R_{2}\varphi_0)(x)=\varphi_0(x)$ imply that the inequality $H_1'(x) >0$ follows by the first equation of Lemma 1(i). At the lower boundary of I we get

\begin{align*} \lim_{x \downarrow \mathfrak{l}} H_1(x) &= \lim_{x \downarrow \mathfrak{l}} \frac{\left( {p}/{\psi_0} \right)'(x)}{\left( {\varphi_0}/{\psi_0} \right)'(x)} = \lim_{x \downarrow \mathfrak{l}} B_0^{-1}(\Psi_0(r + \lambda_2 - \mathcal{A})c)(x) = 0,\end{align*}

where the last equality follows since $c,(r - \mathcal{A})c \in L^1_r(I)$ by Assumption 1.

Part (ii): conditions (5), (6) and Lemma 1 imply that the inequality $K_1(x) >0$ is equivalent to

\begin{align*}\frac{c(x)}{\psi_2(x)}(\Psi_2\varphi_1)(x) + (\Phi_1c)(x) &>0,\end{align*}

which holds by the nonnegativity of c. We also have

\begin{align*} K_1'(x) &= -\bigg(\frac{(\Psi_2\varphi_0)(x)}{(\Psi_2\psi_0)(x)} (\Phi_1\psi_0)(x)-(\Phi_1\varphi_0)(x) \bigg) H_1'(x),\end{align*}

so that $K_1'(x) < 0$ follows from part (i). Moreover, defining a function $F_1\colon I \rightarrow \mathbb{R}_+$ by

\begin{align*} F_1(x) &= \frac{\varphi_0(x)(\Psi_2p)(x) - p(x)(\Psi_2\varphi_0)(x)}{\varphi_0(x)(\Psi_2\psi_0)(x) - \psi_0(x)(\Psi_2\varphi_0)(x)},\end{align*}

we can write $K_1$ as

\begin{align*} K_1(x) &= H_1(x)(\Phi_1\varphi_0)(x) + F_1(x)(\Phi_1\psi_0)(x) - (\Phi_1p)(x).\end{align*}

Consequently, the limit of $K_1$ at the lower boundary of the state space can be seen to be

\begin{align*} \lim_{x \downarrow \mathfrak{l}}K_1(x) =& \lim_{x \downarrow \mathfrak{l}}( (\Phi_1\varphi_0)(x)B_0^{-1}(\Psi_0(r - \mathcal{A})p)(x) + (\Phi_1\psi_0)(x)B_0^{-1}(\Phi_0(r - \mathcal{A})p)(x) \\ &- (\Phi_1(r - \mathcal{A})p)(x) )\\ =& \lim_{x \downarrow \mathfrak{l}} \frac{1}{\lambda_1B_0}(\Phi_1(r - \mathcal{A})p)(x)\\ =& \lim_{x \downarrow \mathfrak{l}} \frac{1}{\lambda_1B_0}\left( \frac{c'(x) \varphi_1(x) - c(x) \varphi_1'(x)}{S'(x)} + (\lambda_2 - \lambda_1)(\Phi_1c)(x) \right)\\ =& \infty.\end{align*}

In the above calculation, the second equality follows from Lemma 1 and the last equality is true because c is nondecreasing, $\mathfrak{l}$ is a natural boundary (which implies that $\lim_{x \downarrow \mathfrak{l}} \varphi_1'(x)/S'(x) = \infty$ ), and $\lambda_1 < \lambda_2$ .

Part (iii): by Assumption 1 we know that $H_2 >0$ at least on the interval $(y^{\lambda_2,c}, \mathfrak{r})$ . Assumption 1(iii) implies that $H_2$ is strictly increasing on this interval, because $H_2$ can be written as

\begin{align*} H_2(x) &= \frac{(\Phi_{2}\psi_0)(x)(\Psi_{2}P)(x) - (\Phi_{2}P)(x)(\Psi_{2}\psi_0)(x) + (\Phi_{2}\psi_0)(x)({(P(x) - \lambda_2(R_{2}P)(x))}/{\varphi_{2}(x)})}{(\Phi_{2}\psi_0)(x)(\Psi_{2}\varphi_0)(x) - (\Phi_{2}\varphi_0)(x)(\Psi_{2}\psi_0)(x)},\end{align*}

and thus,

\begin{align*} &\frac{d}{dx} ((\Phi_{2}\psi_0)(x)(\Psi_{2}\varphi_0)(x) - (\Phi_{2}\varphi_0)(x)(\Psi_{2}\psi_0)(x))= 0,\\ &\frac{d}{dx} ((\Phi_{2}\psi_0)(x)(\Psi_{2}P)(x) - (\Phi_{2}P)(x)(\Psi_{2}\psi_0)(x)) \\ &\qquad=- \frac{P(x) - \lambda_2(R_{2}P)(x)}{\varphi_{2}(x)}\frac{d}{dx}(\Phi_{2}\psi_0)(x),\end{align*}

so that the monotonicity of $H_2$ is determined by the monotonicity of $(P - \lambda_2(R_{2}P))/\varphi_{2}$ . Using the resolvent identity (4) on p gives the form used in Assumption 1(iii).

Parts (iv) and (v): we have

\begin{align*} \lim_{\varepsilon \downarrow 0} \frac{K_2(x+\varepsilon) - K_2(x)}{\varepsilon} &= -\bigg(\frac{(\Phi_2 \varphi_0)(x)}{(\Phi_2 \psi_0)(x)}(\Phi_1 \psi_0)(x)-(\Phi_1 \varphi_0)(x) \bigg) \lim_{\varepsilon \downarrow 0} \frac{H_2(x+\varepsilon) - H_2(x)}{\varepsilon}\end{align*}

and similarly for $\lim_{\varepsilon \downarrow 0} \varepsilon^{-1}(K_2(x) - K_2(x-\varepsilon))$ . The monotonicity of $K_2$ follows by part (iii).

Finally, we may prove part (v) and the positivity of $K_2$ on $(y^{\lambda_2,c}, \hat{y})$ simultaneously with the help of parts (ii) and (iv). Indeed, $\hat{y} \geq y^{\lambda_1}$ implies that

\begin{align*} K_2(\hat{y}) - K_1(\hat{y}) & = H_2(\hat{y}) (\Phi_1 \psi_0)(\hat{y}) \bigg( \frac{\varphi_0(\hat{y})}{\psi_0(\hat{y})} -\frac{(\Phi_2\varphi_0)(\hat{y})}{(\Phi_2\psi_0)(\hat{y})} \bigg) \\ & \hskip11pt+ (\Phi_1 \psi_0)(\hat{y}) \bigg(\frac{(\Phi_2p)(\hat{y})}{(\Phi_2\psi_0)(\hat{y})}-\frac{p(\hat{y})}{\psi_0(\hat{y})} \bigg) - (\Phi_1 g)(\hat{y}) + \frac{g(\hat{y})}{\psi_0(\hat{y})} (\Psi_1 \psi_0)(\hat{y}) \\ & = (\Phi_1 \psi_0)(\hat{y}) \bigg( \frac{g(\hat{y})}{\psi_0(\hat{y})} - \frac{(\Phi_1 g)(\hat{y})}{(\Phi_1 \psi_0)(\hat{y})} \bigg) \\ &> 0,\end{align*}

and we know that $K_1(\hat{y})>0$ by part (ii). Now the monotonicity of $K_2$ on $(y^{\lambda_2,c}, \hat{y})$ implies that $K_2$ is positive on this interval as well.

A.4. Proof of Proposition 1

The proof is divided into four parts that we label (i), (ii), (iii) and (iv). In part (i) we construct a certain auxiliary function $\Gamma$ and show that there is a correspondence between solutions to (22) and fixed points of $\Gamma$ . In part (ii) we show that $\Gamma$ has a fixed point and in part (iii) the fixed point is found to be unique. We prove that the obtained solution $(x^*,y^*)$ to (22) satisfies $x^* < y^{\lambda_1} < y^*$ in part (iv).

  1. 1. The function $H_1$ is strictly increasing on $(\mathfrak{l}, \hat{y})$ and, by Assumption 1, $H_2$ is strictly increasing on $(y^{\lambda_2,c}, \hat{y})$ . We also have $H_1(\mathfrak{l}) = 0 = H_2(y^{\lambda_2,c})$ and $H_1(\hat{y}) = H_2(\hat{y})$ , so that $H_2^{-1}$ is continuous and strictly increasing on $(0, H_1(\hat{y}))$ and $(H_2^{-1} \circ H_1):\: [\mathfrak{l},\hat{y}] \rightarrow [y^{\lambda_2,c}, \hat{y}]$ is a well-defined, continuous and strictly increasing function. The monotonicity properties of $K_1$ imply that $K_1^{-1}$ is positive, continuous and strictly decreasing everywhere on I. The ordering $K_1(\hat{y}) < K_2(\hat{y}) < K_2(y^{\lambda_2,c}) < K_1(\mathfrak{l})$ now implies that the restriction of $(K_1^{-1} \circ K_2)$ to $[y^{\lambda_2,c}, \hat{y}]$ is a well-defined, continuous, strictly increasing function with the codomain $[\mathfrak{l}, \hat{y}]$ . Thus, the mapping $\Gamma$ defined by

    \begin{equation*} \Gamma(x) = (K_1^{-1} \circ K_2 \circ H_2^{-1} \circ H_1)(x)\end{equation*}
    is a well-defined, continuous, strictly increasing function $\Gamma\colon [\mathfrak{l}, \hat{y}] \to [\mathfrak{l}, \hat{y}]$ . Differentiating $\Gamma$ yields
    \begin{align*} \lim_{\varepsilon \downarrow 0} \frac{\Gamma (x+\varepsilon) - \Gamma(x)}{\varepsilon} &= \frac{(({(\Phi_2 \varphi_0)}/{(\Phi_2 \psi_0)})(\Phi_1 \psi_0)-(\Phi_1 \varphi_0) ) ((H_2^{-1}\circ H_1)(x))}{(({(\Psi_2\varphi_0)}/{(\Psi_2\psi_0)}) (\Phi_1\psi_0)-(\Phi_1\varphi_0) ) (\Gamma(x))} >0,\end{align*}
    and $\lim_{\varepsilon \downarrow 0} \varepsilon^{-1}(\Gamma (x) - \Gamma(x-\varepsilon))$ gives the same result. Thus, $\Gamma \in C^1(I)$ even though $H_2,K_2 \in C^0(I)$ . It is also evident that $\Gamma'(x) >0 $ for all $x \in I$ . We now see that the pair (22) has a solution $(x^*,y^*)$ with $x^* \in [\mathfrak{l}, \hat{y}), y^* \in ([y^{\lambda_2,c}, \hat{y})$ if and only if $x^*$ is a fixed point of the mapping $\Gamma$ and $y^* = (H_2^{-1} \circ H_1)(x^*)$ . Thus, it suffices to prove that $\Gamma$ has a unique fixed point $x^* \in (\mathfrak{l}, \hat{y})$ .
  2. 2. First note that $\Gamma (\hat{y}) < \hat{y}$ since $H_1(\hat{y}) = H_2(\hat{y})$ , $K_1(\hat{y}) < K_2(\hat{y})$ and $K_1$ is strictly decreasing. We also have $\Gamma (\mathfrak{l}) > \mathfrak{l}$ because $(H_2^{-1} \circ H_1)(\mathfrak{l}) = y^{\lambda_2,c}$ and $K_2(y^{\lambda_2,c}) < K_1(\mathfrak{l})$ . Combining these observations with the monotonicity of $\Gamma$ implies that $\Gamma$ must have a fixed point $x^* \in (\mathfrak{l}, \hat{y})$ . Moreover, it is given by $\lim_{n \uparrow \infty} \Gamma^n(\hat{y})$ .

  3. 3. We continue by proving $x^* < y^* = (H_2^{-1} \circ H_1)(x^*)$ as an intermediate step. Note that if $y^{\lambda_1} \leq y^{\lambda_2,c}$ then $H_1(x) > H_2(x)$ for $x \in [y^{\lambda_2,c}, \hat{y})$ and, consequently, $x^* < (H_2^{-1} \circ H_1)(x^*)$ . Now suppose that $y^{\lambda_1} > y^{\lambda_2,c}$ . We have $H_2(y^{\lambda_2,c}) = 0 < H_1(y^{\lambda_2,c})$ . If $H_1$ and $H_2$ do not intersect on $(y^{\lambda_2,c}, y^{\lambda_1})$ , then again $H_1(x) > H_2(x)$ for $x \in [y^{\lambda_2,c}, \hat{y})$ and $x^* < (H_2^{-1} \circ H_1)(x^*)$ . Now suppose that $H_1$ and $H_2$ do intersect on $(y^{\lambda_2,c}, y^{\lambda_1})$ and denote the largest such intersection point as $\breve{y}$ . We have

    \begin{align*} K_2(\breve{y}) - K_1(\breve{y}) &= (\Phi_1 \psi_0)(\breve{y}) \bigg( \frac{g(\breve{y})}{\psi_0(\breve{y})} - \frac{(\Phi_1 g)(\breve{y})}{(\Phi_1 \psi_0)(\breve{y})} \bigg) < 0\end{align*}
    since $\breve{y} < y^{\lambda_1}$ . It follows that $\Gamma(\breve{y}) > \breve{y}$ and, consequently, $x^* > \breve{y}$ . But the definition of $\breve{y}$ implies that $H_1(x) > H_2(x)$ for $x \in [\breve{y}, \hat{y})$ and so $x^* < (H_2^{-1} \circ H_1)(x^*)$ . Thus, the inequality $x^* < (H_2^{-1} \circ H_1)(x^*)$ holds for all fixed points $x^*$ of the function $\Gamma$ .

Now suppose that $x' \in (\mathfrak{l}, \hat{y})$ is a fixed point of $\Gamma$ . Then $y' = (H_2^{-1} \circ H_1)(x') > x'$ and we have

\begin{equation*}\begin{split}\Gamma'(x') & = \dfrac{({(\Phi_2 \varphi_0)(y')}/{(\Phi_2 \psi_0)(y')})(\Phi_1 \psi_0)(y')-(\Phi_1 \varphi_0)(y')}{({(\Psi_2\varphi_0)(x')}/{(\Psi_2\psi_0)(x')}) (\Phi_1\psi_0)(x')-(\Phi_1\varphi_0)(x')} \\& = \dfrac{(\Phi_2 \varphi_0)(y')(\Phi_1 \psi_0)(y')-(\Phi_1 \varphi_0)(y')(\Phi_2 \psi_0)(y')}{(\Psi_2\varphi_0)(x')(\Phi_1\psi_0)(x')-(\Phi_1\varphi_0)(x')(\Psi_2\psi_0)(x') } \frac{(\Psi_2\psi_0)(x') }{(\Phi_2 \psi_0)(y')} \\& < \dfrac{\varphi_0(y')(\Phi_1 \psi_0)(y')-\psi_0(y') (\Phi_1 \varphi_0)(y') }{\varphi_0(x')(\Phi_1\psi_0)(x')-\psi_0(x')(\Phi_1\varphi_0)(x') } \frac{(\Psi_2\psi_0)(x') }{(\Phi_2 \psi_0)(y')} \frac{(\Phi_2 1)(y') }{(\Psi_2 1)(x')} \\& < \dfrac{\varphi_0(y')(\Phi_1 \psi_0)(y')-\psi_0(y') (\Phi_1 \varphi_0)(y') }{\varphi_0(x')(\Phi_1\psi_0)(x')-\psi_0(x') (\Phi_1\varphi_0)(x') } \frac{\psi_0(x') }{\psi_0(y')} \\& < 1.\end{split}\end{equation*}

In the last inequality we used the fact that the function $x \mapsto \varphi_0(x)(\Phi_1 \psi_0)(x)-\psi_0(x) (\Phi_1 \varphi_0)(x)$ is decreasing for all x. Thus, by continuity, whenever the function $\Gamma$ intersects the diagonal of $\mathbb{R}^2$ on $(\mathfrak{l}, \hat{y})$ , the intersection must be from above. This implies that the fixed point $x^*$ is unique in $(\mathfrak{l}, \hat{y})$ .

  1. 4. Next we prove that $x^*$ and $y^*$ satisfy $x^* < y^{\lambda_1} < y^*$ . In order to do this, we first define $\mathcal{T}\colon (y^{\lambda_2,c}, \hat{y}) \rightarrow (y^{\lambda_2,c}, \hat{y})$ by $\mathcal{T}(x) = (H_2^{-1}\circ H_1 \circ K_1^{-1} \circ K_2)(x)$ . Using similar arguments as with $\Gamma$ , it can be shown that $\mathcal{T}$ is well-defined, continuous and strictly increasing and $y^*$ is the unique fixed point of $\mathcal{T}$ . We also have $\mathcal{T}'(y^*) < 1$ .

Now we reintroduce the cost scaling parameter $\alpha > 0$ as in the proof of Lemma 3. Owing to the uniqueness of $x^*$ and $y^*$ and $\Gamma'(x^*), \mathcal{T}'(y^*) < 1$ , it suffices to show that $d\Gamma_\alpha(y^{\lambda_1})/d\alpha >0$ and $d\mathcal{T}_\alpha(y^{\lambda_1})/d\alpha <0$ for all $\alpha \in (0, \alpha_c)$ . Expanding the first derivative gives

\begin{align*} \frac{d}{d\alpha}\Gamma_\alpha(y^{\lambda_1}) =& \left( \frac{-\alpha^{-2}K_{2,\alpha}((H_{2,\alpha}^{-1} \circ H_{1,\alpha})(y^{\lambda_1})) + \alpha^{-1}({d}/{d\alpha})K_{2,\alpha}((H_{2,\alpha}^{-1} \circ H_{1,\alpha})(y^{\lambda_1}))}{K_1'( K_1^{-1} ({(K_{2,\alpha} \circ H_{2,\alpha}^{-1} \circ H_{1,\alpha})(y^{\lambda_1})}/{\alpha} ) )} \right. \\ &\left. + \frac{K_{2,\alpha}'((H_{2,\alpha}^{-1} \circ H_{1,\alpha})(y^{\lambda_1}))}{K_1'( K_1^{-1}( {(K_{2,\alpha} \circ H_{2,\alpha}^{-1} \circ H_{1,\alpha})(y^{\lambda_1})}/{\alpha} ) )} \right) \frac{d}{d\alpha}(H_{2,\alpha}^{-1} \circ H_{1,\alpha})(y^{\lambda_1})\end{align*}

By Assumption 1, Lemma 4, and $\alpha < \alpha_c$ , we know that in the above expression the denominator and the numerator on the second line are negative. The numerator in the first line is negative as well. To see this, note that

\begin{align*} &-\alpha^{-2}K_{2,\alpha}((H_{2,\alpha}^{-1} \circ H_{1,\alpha})(y^{\lambda_1})) + \alpha^{-1}\frac{d}{d\alpha}K_{2,\alpha}((H_{2,\alpha}^{-1} \circ H_{1,\alpha})(y^{\lambda_1})) \\ &\qquad= - \alpha^{-2} \left( \frac{g}{\psi_0}(\Phi_1\psi_0) - (\Phi_1g) - \frac{g(\Phi_2\psi_0) - \psi_0(\Phi_2g)}{\varphi_0(\Phi_2\psi_0) - \psi_0(\Phi_2\varphi_0)} \right. \\ &\hskip33pt \times \left( \frac{\varphi_0}{\psi_0}(\Phi_1\psi_0) - (\Phi_1\varphi_0) \right) ) ((H_{2,\alpha}^{-1} \circ H_{1,\alpha})(y^{\lambda_1})).\end{align*}

$\varphi_0/\psi_0$ is strictly decreasing so

\begin{align*} \varphi_0(\Phi_2\psi_0)(x) - \psi_0(\Phi_2\varphi_0)(x) >0,\qquad \varphi_0(\Phi_1\psi_0)(x) - \psi_0(\Phi_1\varphi_0)(x) > 0.\end{align*}

On the other hand, $\alpha < \alpha_c$ is equivalent to $y^{\lambda_1} < (H_{2,\alpha}^{-1} \circ H_{1,\alpha})(y^{\lambda_1})$ and we know that $(H_{2,\alpha}^{-1} \circ H_{1,\alpha})(y^{\lambda_1}) \in (y^{\lambda_2,c}, \hat{y})$ and $\hat{y} < y^{\lambda_2}$ so

\begin{align*}(g(x)(\Phi_1\psi_0)(x) - \psi_0(x)(\Phi_1 g)(x))((H_{2,\alpha}^{-1} \circ H_{1,\alpha})(y^{\lambda_1})) &< 0,\\ (g(\Phi_2\psi_0) - \psi_0(\Phi_2g))((H_{2,\alpha}^{-1} \circ H_{1,\alpha})(y^{\lambda_1})) &> 0.\end{align*}

Next we show that $d(H_{2,\alpha}^{-1} \circ H_{1,\alpha})(y^{\lambda_1})/d\alpha < 0$ . Indeed, for $x \in (y^{\lambda_2,c}, \hat{y})$ , we get

\begin{align*} \frac{d}{d\alpha}(H_{1,\alpha}^{-1} \circ H_{2,\alpha})(x) &= \frac{-\alpha^{-2}H_{2,\alpha}(x) + \alpha^{-1}({d}/{d\alpha} ) H_{2,\alpha}(x) }{H_1'\left( H_1^{-1}\left( {H_{2,\alpha}(x)}/{\alpha} \right) \right) }\\ &= - \frac{ \frac{g(x)(\Phi_2\psi_0)(x) - \psi_0(x)(\Phi_2g)(x)}{\varphi_0(x)(\Phi_2\psi_0)(x) - \psi_0(x)(\Phi_2\varphi_0)(x)} }{\alpha^{2} H_1'\left( H_1^{-1}\left( {H_{2,\alpha}(x)}/{\alpha} \right) \right) }\\ &>0\end{align*}

and $H_{1,\alpha}^{-1} \circ H_{2,\alpha}$ is strictly increasing on $(y^{\lambda_2,c}, \hat{y})$ , so for all $x \in (\mathfrak{l}, \hat{y})$ , we have

\begin{align*} \frac{d}{d\alpha}(H_{2,\alpha}^{-1} \circ H_{1,\alpha})(x) &= \frac{d}{d\alpha}\left( (H_{1,\alpha}^{-1} \circ H_{2,\alpha})^{-1}(x)\right) <0\end{align*}

so that $d(H_{2,\alpha}^{-1} \circ H_{1,\alpha})(y^{\lambda_1})/d\alpha < 0$ . Combining the observations yields $d\Gamma_\alpha(y^{\lambda_1})/d\alpha >0$ for $\alpha \in (0, \alpha_c)$ and, consequently, $y^{\lambda_1} = \Gamma_{\alpha_c}(y^{\lambda_1}) > \Gamma_\alpha(y^{\lambda_1})$ . But the unique fixed point $x^*_\alpha$ of $\Gamma_\alpha$ satisfies $\Gamma_\alpha'(x^*_\alpha) < 1$ so it must hold that $x^*_\alpha < y^{\lambda_1}$ .

We may expand the derivative of $\mathcal{T}_\alpha(y^{\lambda_1})$ as

\begin{align*} \frac{d}{d\alpha}\mathcal{T}_\alpha(y^{\lambda_1}) =& \frac{d}{d\alpha} (H_{2,\alpha}^{-1} \circ H_{1,\alpha})((K_{1,\alpha}^{-1} \circ K_{2,\alpha})(y^{\lambda_1})) + (H_{2,\alpha}^{-1} \circ H_{1,\alpha})'((K_{1,\alpha}^{-1} \circ K_{2,\alpha})(y^{\lambda_1}))\\ &\times \frac{-\alpha^{-2}K_{2,\alpha}(y^{\lambda_1}) + \alpha^{-1}({d}/{d\alpha})K_{2,\alpha}(y^{\lambda_1})}{K_1'\left( K_1^{-1}\left( {K_{2,\alpha}(y^{\lambda_1})}/{\alpha} \right) \right)} (y^{\lambda_1}).\end{align*}

We have already seen that the first term is negative, the first half of the second term is positive and the second half is negative. Hence, $d\mathcal{T}_\alpha(y^{\lambda_1})/d\alpha < 0$ for $\alpha \in (0, \alpha_c)$ and, consequently, $y^{\lambda_1} = \Gamma_{\alpha_c}(y^{\lambda_1}) < \Gamma_\alpha(y^{\lambda_1})$ . But the unique fixed point $y^*_\alpha$ of $\mathcal{T}_\alpha$ satisfies $\mathcal{T}_\alpha'(y^*_\alpha) < 1$ so it must hold that $y^*_\alpha > y^{\lambda_1}$ . The proof is now complete.

A.5. Proof of Proposition 2

First note that the properties of the constants $C_1, C_2, C_3$ and the thresholds $x^*, y^*$ imply that $\mathcal{K}_0 \in C^0(I)$ . The rest of the proof is divided into four parts that we label (i), (ii), (iii) and (iv). We begin by showing in part (i) that $\mathcal{K}_0$ satisfies the DPP (24) on $[y^\infty, \mathfrak{r})$ . Parts (ii), (iii) and (iv) contain the corresponding proofs for the intervals $[y^*, y^\infty)$ , $[x^*, y^*)$ and $(\mathfrak{l}, x^*)$ , respectively.

  1. 1. Let $x \geq y^\infty$ . Then $\mathcal{K}_0(x) = g(x) = \tilde{V}(x)$ , where $\tilde{V}$ is the value function of the optimal stopping problem in continuous time (26). Here $\tilde{V}$ is r excessive and $p(x) - \lambda_2(R_2p)(x) = c(x)$ by (4) so $\mathcal{K}_0(x) = \max \{g(x), \mathcal{K}_1(x), \mathcal{K}_2(x)\}$ . Thus, the DPP holds on $[y^\infty, \mathfrak{r})$ .

  2. 2. Let $y^* \leq x < y^\infty$ . Now $\mathcal{K}_0(x) = g(x)$ and making use of the strong Markov property of linear diffusions, memoryless property of the exponential distribution, resolvent equation (4), and the assumption that $(P - \lambda_2(R_2P))/\varphi_2$ is strictly increasing on $(y^{\lambda_2,c}, y^\infty)$ , we get

    \begin{align*} g(x) - \mathcal{K}_2(x) =& g(x) - \mathbb{E}_x\left[ \textrm{e}^{-rU_2}\mathcal{K}_0(X_{U_2})(1(U_2 < \tau_{y^*}) + 1(U_2 \geq \tau_{y^*})) \right] + c(x)\\ =& g(x) - \mathbb{E}_x\left[ \textrm{e}^{-rU_2}g(X_{U_2}) \right] + c(x)\\ &+ \mathbb{E}_x\left[ \textrm{e}^{-(r+\lambda_2)\tau_{y^*}} \right] \mathbb{E}_{y^*}\left[ \textrm{e}^{-rU_2}(g(X_{U_2}) - \mathcal{K}_0(X_{U_2})) \right]\\ =& g(x) - \lambda_2(R_2g)(x) + c(x) - \frac{g(y^*) -\lambda_2(R_2g)(y^*) + c(y^*)}{\varphi_2(y^*)}\varphi_2(x)\\ \geq & 0.\end{align*}
    By similar arguments we obtain
    \begin{align*} g(x) - \mathcal{K}_1(x) &= g(x) - \lambda_1(R_1g)(x) - \frac{\mathcal{K}_1(y^*) - \lambda_1(R_1g)(y^*)}{\varphi_1(y^*)} \varphi_1(x)\\ &\geq \frac{g(y^*) - \mathcal{K}_1(y^*)}{\varphi_1(y^*)} \varphi_1(x).\end{align*}
    Recalling that $K_1(x^*) = K_2(y^*)$ and by Remark 2 and the proof of Lemma 4(ii), $K_1$ and $K_2$ can be written as
    \begin{align*} K_1(x) &= F_1(x)(\phi_1\psi_0)(x) + H_1(x)(\phi_1\varphi_0)(x) - (\phi_1p)(x),\\ K_2(x) &= F_2(x)(\phi_1\psi_0)(x) + H_2(x)(\phi_1\varphi_0)(x) - (\phi_1P)(x).\end{align*}
    We find that expanding the resolvent $\mathcal{K}_1(y^*) = \lambda_1(R_1\mathcal{K}_0)(y^*)$ according to (2) yields
    \begin{align*} g(y^*) - \mathcal{K}_1(y^*) =& g(y^*) - F_2(y^*)\psi_0(y^*) - H_2(y^*)\varphi_0(y^*) + \lambda_1(R_1p)(y^*)\\ &+ \frac{F_1(x^*)\psi_0(x^*) + H_1(x^*)\varphi_1(x^*) - \lambda_1(R_1p)(x^*)}{\varphi_1(x^*)}\varphi_1(y^*)\\ &+ \frac{\lambda_1}{B_1}K_1(x^*)\left( \psi_1(y^*) - \frac{\psi_1(x^*)}{\varphi_1(x^*)}\varphi_1(y^*) \right)\\ =& \lambda_1(R_1p)(y^*) - p(y^*) + \frac{p(x^*) - \lambda_1(R_1p)(x^*)}{\varphi_1(x^*)}\varphi_1(y^*)\\ &+ \frac{\lambda_1}{B_1}K_1(x^*)\left( \psi_1(y^*) - \frac{\psi_1(x^*)}{\varphi_1(x^*)}\varphi_1(y^*) \right).\end{align*}
    Letting $f(x) = c(x) + (\lambda_2 - \lambda_1)(R_1c)(x)$ and invoking again the resolvent identity (4), we see that the above expression is nonnegative if and only if
    \begin{align*} \frac{{f(y^*)}/{\varphi_1(y^*)} - {f(x^*)}/{\varphi_1(x^*)}}{{\psi_1(y^*)}/{\varphi_1(y^*)} - {\psi_1(x^*)}/{\varphi_1(x^*)}} &\leq \frac{\lambda_1}{B_1}K_1(x^*),\end{align*}
    The left-hand side is continuously differentiable and $K_1(x^*)$ may be written according to Remark 2 as
    \begin{align*} K_1(x^*) &= \frac{\lambda_2 - \lambda_1}{B_1}\left( \frac{c(x^*)}{\psi_2(x^*)}(\Psi_2\varphi_1)(x^*) + (\Phi_1c)(x^*) \right).\end{align*}
    The Cauchy mean value theorem then implies that
    \begin{align*} &\frac{{f(y^*)}/{\varphi_1(y^*)} - {f(x^*)}/{\varphi_1(x^*)}}{{\psi_1(y^*)}/{\varphi_1(y^*)} - {\psi_1(x^*)}/{\varphi_1(x^*)}} \\ &\qquad\leq \sup\nolimits_{x \in [x^*, y^*]} \frac{\left( {f(x)}/{\varphi_1(x)} \right) '}{\left( {\psi_1(x)}/{\varphi_1(x)} \right) '} \\ &\qquad= \sup\nolimits_{x \in [x^*, y^*]} \frac{1}{B_1} \left( (\Phi_1(r+\lambda_1 - \mathcal{A})c)(x) + (\lambda_2 - \lambda_1)(\Phi_1c)(x) \right)\\ &\qquad= \frac{1}{B_1}(\Phi_1(r+\lambda_2 - \mathcal{A})c)(x^*).\end{align*}
    But
    \begin{align*}(\Phi_1(r+\lambda_2 - \mathcal{A})c)(x^*) &\leq \frac{\lambda_2 - \lambda_1}{B_1}\left( \frac{c(x^*)}{\psi_2(x^*)}(\Psi_2\varphi_1)(x^*) + (\Phi_1c)(x^*) \right)\end{align*}
    is equivalent to
    \begin{align*} c'(x^*)\psi_2(x^*) - c(x^*)\psi_2'(x^*) &\leq 0,\end{align*}
    which is true since $\psi_0/\psi_2$ is strictly decreasing and $c/\psi_0$ is strictly decreasing by Assumption 1. This proves that $g(y^*) \geq \mathcal{K}_1(y^*)$ and, hence, $\mathcal{K}_0(x) \geq \mathcal{K}_1(x)$ for $x \in [y^*, y^{\infty})$ .
  3. 3. Let $x^* \leq x < y^*$ . Now $\mathcal{K}_0(x) = \mathcal{K}_2(x)$ . Denote by $\hat{\tau}$ the first exit time of X from the interval $(x^*, y^*)$ . We recall from [2, p.272] that

    \begin{align*} \mathbb{E}_x\left[ \textrm{e}^{-(r+\lambda_1)\hat{\tau}} 1(\tau_{x^*} < \tau^{y^*}) \right] &= \frac{\psi_1(y^*)\varphi_1(x) - \psi_1(x)\varphi_1(y^*)}{\psi_1(y^*)\varphi_1(x^*) - \psi_1(x^*)\varphi_1(y^*)},\\ \mathbb{E}_x\left[ \textrm{e}^{-(r+\lambda_1)\hat{\tau}} 1(\tau_{x^*} > \tau^{y^*}) \right] &= \frac{\psi_1(x)\varphi_1(x^*) - \psi_1(x^*)\varphi_1(x)}{\psi_1(y^*)\varphi_1(x^*) - \psi_1(x^*)\varphi_1(y^*)}.\end{align*}
    Applying similar arguments as before, we see that $\mathcal{K}_2(x) \geq \mathcal{K}_1(x)$ is equivalent to
    \begin{align*} &-f(x) + f(x^*)\frac{\psi_1(y^*)\varphi_1(x) - \psi_1(x)\varphi_1(y^*)}{\psi_1(y^*)\varphi_1(x^*) - \psi_1(x^*)\varphi_1(y^*)}\\ &+ (f(y^*) + g(y^*) - G_1(y^*))\frac{\psi_1(x)\varphi_1(x^*) - \psi_1(x^*)\varphi_1(x)}{\psi_1(y^*)\varphi_1(x^*) - \psi_1(x^*)\varphi_1(y^*)} \geq 0,\end{align*}
    where again $f(x) = c(x) + (\lambda_2 - \lambda_1)(R_1c)(x)$ . Expanding $\mathcal{K}_1(y^*)$ as in part (ii), the above condition is found to be equivalent to
    \begin{align*} \frac{{f(x)}/{\varphi_1(x)} - {f(x^*)}/{\varphi_1(x^*)}}{{\psi_1(x)}/{\varphi_1(x)} - {\psi_1(x^*)}/{\varphi_1(x^*)}} &\leq \frac{\lambda_1}{B_1}K_1(x^*),\end{align*}
    which can be proven as in part (ii). Thus, $\mathcal{K}_0(x) \geq \mathcal{K}_1(x)$ for $x \in [x^*, y^*)$ .

In order to prove $\mathcal{K}_2(x) \geq g(x)$ , we first note that the condition is equivalent to

\begin{align*} \frac{F_2(x) - F_2(y^*)}{H_2(y^*) - H_2(x)} &\leq \frac{\varphi_0(x)}{\psi_0(x)}.\end{align*}

On the other hand, $\varphi_0/\psi_0 > (\Phi_2\varphi_0)/(\Phi_2\psi_0)$ so it suffices to show that

\begin{align*} \frac{F_2(x) - F_2(y^*)}{H_2(y^*) - H_2(x)} &\leq \frac{(\Phi_2\varphi_0)(x)}{(\Phi_2\psi_0)(x)},\end{align*}

which in turn is equivalent to

\begin{align*} \frac{{(\Phi_2P)(x)}/{(\Phi_2\psi_0)(x)} - {(\Phi_2P)(y^*)}/{(\Phi_2\psi_0)(y^*)}}{{(\Phi_2\varphi_0)(x)}/{(\Phi_2\psi_0)(x)} - {(\Phi_2\varphi_0)(y^*)}/{(\Phi_2\psi_0)(y^*)}} &< H_2(y^*).\end{align*}

We know that $l_g < x^* < x < y^*$ , $H_2(y) \leq 0$ for $y \in (l_g, y^{\lambda_2,c})$ and $H_2$ is strictly increasing on $(y^{\lambda_2,c}, y^*)$ , so the Cauchy mean value theorem implies that

\begin{align*} &\frac{{(\Phi_2P)(x)}/{(\Phi_2\psi_0)(x)} - {(\Phi_2P)(y^*)}/{(\Phi_2\psi_0)(y^*)}}{{(\Phi_2\varphi_0)(x)}/{(\Phi_2\psi_0)(x)} - {(\Phi_2\varphi_0)(y^*)}/{(\Phi_2\psi_0)(y^*)}} \\ &\qquad\leq \sup\nolimits_{x \in [x^*, y^*)} \frac{\left( {(\Phi_2P)(x)}/{(\Phi_2\psi_0)(x)} \right)'}{\left( {(\Phi_2\varphi_0)(x)}/{(\Phi_2\psi_0)(x)} \right)'}\\ &\qquad= \sup\nolimits_{x \in [x^*, y^*)} H_2(x)\\ &\qquad= H_2(y^*),\end{align*}

proving the condition $\mathcal{K}_2(x) \geq g(x)$ for $x \in [x^*, y^*)$ .

  1. 4. Let $x < x^*$ . Now $\mathcal{K}_0(x) = \mathcal{K}_1(x)$ and $\mathcal{K}_1(x) \geq g(x)$ is equivalent to

    \begin{align*} \frac{g(x)}{\psi_0(x)} &\leq \frac{\mathcal{K}_1(x^*)}{\psi_0(x^*)}.\end{align*}
    The above inequality is true because $l_g < x^* < y^\infty$ so $g(x)/\psi_0(x) \leq g(x^*)/\psi_0(x^*)$ by Assumption 1 and $g(x^*) \leq \mathcal{K}_1(x^*)$ by part (iii). On the other hand,
    \begin{align*} \mathcal{K}_1(x) - \mathcal{K}_2(x) &= c(x) - \frac{c(x^*)}{\psi_2(x^*)}\psi_2(x) \geq 0,\end{align*}
    where the inequality follows by Assumption 1, since $\psi_0/\psi_2$ is strictly decreasing.

Combining the observations made in parts (i)–(iv), we see that $\mathcal{K}_0 = \max \{g, \mathcal{K}_1, \mathcal{K}_2\}$ so that $\mathcal{K}_0$ satisfies the DPP (24). Proving the smoothness properties of $\mathcal{K}_1$ and $\mathcal{K}_2$ is now completely analogous to the proof of Lemma 2.

Funding information

The Foundation for Economic Education (Liikesivistysrahasto) and OP Research Foundation (grant number 20240114) are acknowledged for funding. Emmy.network is acknowledged for continued support.

Competing interests

There were no competing interests to declare that arose during the preparation or publication process of this paper.

References

Alvarez, L.H.R.. Singular stochastic control, linear diffusions, and optimal stopping: A class of solvable problems. SIAM Journal on Control and Optimization, 39:16971710, 2001.Google Scholar
Alvarez, L.H.R.. A class of solvable impulse control problems. Applied Mathematics and Optimization, 49:265295, 2004.Google Scholar
Alvarez, L.H.R.. A class of solvable stopping games. Applied Mathematics and Optimization, 58:291314, 2008.Google Scholar
Alvarez, Luis HR. Stochastic forest stand value and optimal timber harvesting. SIAM Journal on Control and Optimization, 42(6):19721993, 2004.Google Scholar
Alvarez, Luis HR and Shepp, Larry A. Optimal harvesting of stochastically fluctuating populations. Journal of Mathematical Biology, 37:155177, 1998.Google Scholar
Arai, T. and Takenaka, M.. Constrained optimal stopping under a regime-switching model. arXiv:2204.07914, 2022.Google Scholar
Beneš, Václav E, Gaitsgori, Georgy, and Karatzas, Ioannis. Drift control with discretionary stopping for a diffusion process. arXiv preprint arXiv:2401.10043, 2024.Google Scholar
Borodin, A. N. and Salminen, P.. Handbook of Brownian Motion - Facts and Formulae. Springer, Birkhäuser, Basel, 2 edition, 2015.Google Scholar
Davis, Mark HA and Zervos, Mihail. A problem of singular stochastic control with discretionary stopping. The Annals of Applied Probability, pages 226240, 1994.10.1214/aoap/1177005209CrossRefGoogle Scholar
Dupuis, P. and Wang, H.. Optimal stopping with random intervention times. Advances in Applied Probability, 34:141157, 2002.Google Scholar
Gapeev, Pavel V.. Discounted nonzero-sum optimal stopping games under Poisson random intervention times. Stochastics, 96:18621892, 2024.Google Scholar
Guo, X. and Liu, J.. Stopping at the maximum of geometric Brownian motion when signals are received. Journal of Applied Probability, 42:826838, 2005.Google Scholar
Hobson, D., Liang, G., and Sun, H.. Callable convertible bonds under liquidity constraints. arXiv:2111.02554, 2021.Google Scholar
Hobson, D. and Zeng, M.. Randomised rules for stopping problems. Journal of Applied Probability, 57:9811004, 2020.10.1017/jpr.2020.43CrossRefGoogle Scholar
Hobson, D. and Zeng, M.. Constrained optimal stopping, liquidity and effort. Stochastic Processes and their Applications, 150:819843, 2022.Google Scholar
Hobson, David, Liang, Gechun, and Wang, Edward. Zero-sum Dynkin games under common and independent Poisson constraints. arXiv preprint arXiv:2411.07134, 2024.Google Scholar
Huang, Lixin and Liu, Hong. Rational inattention and portfolio selection. The Journal of Finance, 62(4):19992040, 2007.10.1111/j.1540-6261.2007.01263.xCrossRefGoogle Scholar
Karatzas, Ioannis and Sudderth, William D. Control and stopping of a diffusion process on an interval. Annals of Applied Probability, pages 188196, 1999.10.1214/aoap/1029962601CrossRefGoogle Scholar
Karatzas, Ioannis and Zamfirescu, Ingrid-Mona. Game approach to the optimal stopping problem. Stochastics, 77(5):401435, 2005.10.1080/17442500500219885CrossRefGoogle Scholar
Karatzas, Ioannis and Zamfirescu, Ingrid-Mona. Martingale approach to stochastic control with discretionary stopping. Applied Mathematics and Optimization, 53:163184, 2006.10.1007/s00245-005-0841-2CrossRefGoogle Scholar
Lange, R.-J., Ralph, D., and Støre, K. Real-option valuation in multiple dimensions using Poisson optional stopping times. Journal of Financial and Quantitative Analysis, 55:653677, 2020.10.1017/S0022109019000048CrossRefGoogle Scholar
Lempa, J.. A note on optimal stopping of diffusions with a two-sided optimal rule. Oper. Res. Lett., 38:1116, 2010.10.1016/j.orl.2009.10.002CrossRefGoogle Scholar
Lempa, J.. Optimal stopping with information constraint. Applied Mathematics and Optimization, 66:147173, 2012.Google Scholar
Lempa, J. and Saarinen, H.. A zero-sum Poisson stopping game with asymmetric signal rates. Applied Mathematics and Optimization, 87:35, 2023.10.1007/s00245-022-09945-1CrossRefGoogle Scholar
Liang, G. and Sun, H.. Dynkin games with Poisson random intervention times. SIAM Journal on Control and Optimization, 57:29622991, 2019.Google Scholar
Liang, G. and Sun, H.. Risk-sensitive Dynkin games with heterogeneous Poisson random intervention times. arXiv:2008.01787, 2020.Google Scholar
Liang, G. and Wei, W.. Optimal switching at Poisson random intervention times. Discrete and Continuous Dynamical Systems-Series B, 21:14831505, 2016.10.3934/dcdsb.2016008CrossRefGoogle Scholar
Liu, Hong, Kadan, Ohad, and Yang, Jun. Inattention, forced exercise, and the valuation of executive stock options. Available at SSRN 1106409, 2009.10.2139/ssrn.1106409CrossRefGoogle Scholar
Bartosz Maćkowiak, Filip Matĕjka, and Mirko Wiederholt. Rational inattention: A review. Journal of Economic Literature, 61(1):226–273, 2023.Google Scholar
McDonald, R and Siegel, D.. The value of waiting to invest. The Quarterly Journal of Economics, 101: 707728, 1986.10.2307/1884175CrossRefGoogle Scholar
Menaldi, J. L. and Robin, M.. On some optimal stopping problems with constraint. SIAM Journal on Control and Optimization, 54:26502671, 2016.10.1137/15M1040001CrossRefGoogle Scholar
Moscarini, Giuseppe. Limited information capacity as a source of inertia. Journal of Economic Dynamics and control, 28(10):20032035, 2004.10.1016/j.jedc.2003.08.002CrossRefGoogle Scholar
Reis, Ricardo. Inattentive consumers. Journal of monetary Economics, 53(8):17611800, 2006.10.1016/j.jmoneco.2006.03.001CrossRefGoogle Scholar
L-C-G Rogers and Omar Zane. A simple model of liquidity effects. In Advances in finance and stochastics: essays in honour of Dieter Sondermann, pages 161–176. Springer, 2002.10.1007/978-3-662-04790-3_9CrossRefGoogle Scholar
Saarinen, Harto. On Poisson Constrained Control of Linear Diffusions. PhD thesis, Annales Universitatis Turkuensis, Ser. AI, 689, Astronomica-Chemica-Physica-Mathematica, 2023.Google Scholar
Sims, Christopher A. Implications of rational inattention. Journal of monetary Economics, 50(3):665690, 2003.10.1016/S0304-3932(03)00029-1CrossRefGoogle Scholar
Sims, Christopher A. Rational inattention: Beyond the linear-quadratic case. American Economic Review, 96(2):158163, 2006.10.1257/000282806777212431CrossRefGoogle Scholar
Song, Qingshuo, Stockbridge, Richard H, and Zhu, Chao. On optimal harvesting problems in random environments. SIAM journal on control and optimization, 49(2):859889, 2011.Google Scholar
Figure 0

Figure 1. A prototype plot of the functions $H_1$, $H_2$, $K_1$, $K_2$ and the thresholds $x^*$, $y^*$, $y^{\lambda_1}$, $y^{\lambda_2}$.

Figure 1

Figure 2. Optimal thresholds as a function of attention rate $\lambda_1$. The other parameters are chosen to be $\mu = 0.02$, $\sigma = \sqrt{0.3}$, $r = 0.05$, $k = 0.001$, $\lambda_2 = 4.0$, $\theta = 0.67$, $\eta = 1.5$.

Figure 2

Figure 3. Optimal thresholds as a function of attention rate $\lambda_2$. The other parameters are chosen to be $\mu = 0.02$, $\sigma = \sqrt{0.3}$, $r = 0.05$, $k = 0.001$, $\lambda_1 = 1.0$, $\theta = 0.67$, $\eta = 1.5$.

Figure 3

Table 1. Optimal thresholds with various attention rates. The other parameters are chosen to be $\mu = 0.01$, $\sigma = \sqrt{0.1}$, $r = 0.05$, $\gamma = 0.5$, $k = 0.01$ and $\eta = 2.0$.

Figure 4

Table 2. Optimal thresholds with various costs for higher attention rate. The other parameters are chosen to be $\mu = 0.01$, $\sigma = \sqrt{0.1}$, $r = 0.05$, $\gamma = 0.5$, $\lambda_1 = 0.1$, $\lambda_2 = 1.0$ and $\eta = 2.0$.