1 Introduction
Consider a stochastic process ${\{{X_{n}^{\varepsilon }}\}_{n=0}^{\infty }}$ of autoregressive type, defined by
where f is a continuous mapping from $\mathbb{R}$ to itself with a fixed point at the origin, ${\{{\xi _{n}}\}_{n=1}^{\infty }}$ is a sequence of i.i.d. random variables and ε is a small positive parameter. It is a Markov chain, and under suitable assumptions on ${\{{\xi _{n}}\}_{n=1}^{\infty }}$ and f, it is interesting to investigate how long it takes for the process to leave a neighbourhood of the origin.
(1)
\[ {X_{n+1}^{\varepsilon }}=f\big({X_{n}^{\varepsilon }}\big)+\varepsilon {\xi _{n+1}},\hspace{1em}{X_{0}}=0,\]The original motivation behind our work is to study the time until extinction of a population. A stochastic process that models a population may be positive recurrent and stay at a certain level (or carrying capacity) for a very long time, and when extinction happens, the process first leaves a neighbourhood around that level. Populations can be modeled by, for example, branching processes (such as those treated in [5] and [6]) or models such as the Ricker model which we will use as an example at the end of the paper; it has been studied in [9].
In [12] and the updated version [13], Klebaner and Liptser used the large deviation principle to get an upper bound on the exit time from a set for a process. As an example, they considered the linear autoregressive process, defined as in (1) with $f(x)=ax$, $|a|\lt 1$, and normally distributed innovations. For this example, they showed that the exit time ${\tau _{\varepsilon }}$ from the interval $(-1,1)$ satisfies
The corresponding lower bound has been found by other methods [18]. This means that the upper bound on the right-hand side above is the best possible one.
(2)
\[ \underset{\varepsilon \to 0}{\limsup }{\varepsilon ^{2}}\log {E_{{x_{0}}}}({\tau _{\varepsilon }})\le \frac{1-{a^{2}}}{2}.\]We have studied a corresponding multivariate case [11], and the results were extended to the ARMA model in [15]. Exit times for autoregressive processes with other noise distributions have also been studied, in [8]. Different aspects of the exit time problem for linear autoregressive processes are treated in, for example, [2–4] and [17]. A related exit problem in a different setting was treated in [14]. A case with piecewise linear function $f(x)$ was studied by Anděl et al. in [1].
In this paper we extend the previous results from the linear case to some cases with piecewise linear functions. We will show that the large deviation principle gives explicit (asymptotic) upper bounds on the expectation of the exit time also in these cases. We also apply the methods used to other nonlinear functions, such as quadratic functions and the Ricker model.
In Section 2, we summarize how the large deviation method results in a sum that should be minimized for an upper bound of the exit time. This is based on the methods used in the proofs of Theorems 2.2 and 3.1 in [13]. In Section 3 we study how the minimization can be done over more restricted sets. In Section 4, which contains the main results of this paper, we get the explicit upper bounds in several piecewise linear cases. In Section 5 we explore some other nonlinear cases. In Section 6 we point out a connection between the results and the stationary distribution of the process.
2 Large deviation tools and bounds for exit times
In this section, which is based on work by Klebaner and Liptser in [12] and Jung in [11], we summarize how the large deviation principle (LDP) can be used to get an upper bound of the asymptotics of an exit time from a set for a process. The definition of the LDP used in [12] is as follows (this follows Varadhan’s definition in [19], with the addition that the rate of speed $q(\varepsilon )$ is a function of ε such that $q(\varepsilon )\to 0$ as $\varepsilon \to 0$).
Let $\{{P_{\varepsilon }}\}$ be a family of probability measures on the Borel subsets of a complete separable metric space Z. The family $\{{P_{\varepsilon }}\}$ satisfies the large deviation principle with a rate function I if there is a function I from Z into $[0,\infty ]$ that satisfies the following conditions: $0\le I(z)\le \infty \hspace{0.2778em}\forall z\in Z$, I is lower semicontinuous, the set $\{z:I(z)\le l\}$ is a compact set in Z for all $\hspace{0.2778em}l\lt \infty $ and
where g is a continuous function on ${\mathbb{R}^{m}}$, ${\{{\xi _{n}}\}_{n=m}^{\infty }}$ is a sequence of i.i.d. random variables and ${x_{0}},\dots ,{x_{m-1}}$ are given starting values. They gave conditions under which the LDP holds for the family $\varepsilon \xi $ (where ξ is a copy of ${\xi _{m}}$) and proved that when $\varepsilon \xi $ obeys an LDP with rate function $I(z)$, it follows that $({X_{n}^{\varepsilon }})$ obeys an LDP with a rate function $J(\bar{y})$ that can be written explicitly using $I(z)$:
and $J(\bar{y})=\infty $ otherwise.
\[\begin{array}{r@{\hskip10.0pt}c@{\hskip10.0pt}l}& & \displaystyle \underset{\varepsilon \to 0}{\limsup }q(\varepsilon )\log {P_{\varepsilon }}(C)\le -\underset{z\in C}{\inf }I(z)\hspace{0.2778em}\hspace{0.2778em}\hspace{2.5pt}\text{for every closed set}\hspace{2.5pt}C\subset Z\hspace{2.5pt}\text{and}\hspace{2.5pt}\\ {} & & \displaystyle \underset{\varepsilon \to 0}{\liminf }q(\varepsilon )\log {P_{\varepsilon }}(G)\ge -\underset{z\in G}{\inf }I(z)\hspace{0.2778em}\hspace{0.2778em}\hspace{2.5pt}\text{for every open set}\hspace{2.5pt}G\subset Z.\end{array}\]
In [12] and [13], Klebaner and Liptser considered a family of processes of the type
(3)
\[ {X_{n+1}^{\varepsilon }}=g\big({X_{n}^{\varepsilon }},\dots ,{X_{n-m+1}^{\varepsilon }},\varepsilon {\xi _{n+1}}\big),\](4)
\[ J(\bar{y})={\sum \limits_{k=m}^{\infty }}\underset{{v_{k}}:{y_{k}}=f({y_{k-1}},\dots {y_{k-m}},{v_{k}})}{\inf }I({y_{k}})\hspace{2.5pt}\text{when}\hspace{2.5pt}{u_{0}}={x_{0}},\dots ,{u_{m-1}}={x_{m-1}},\]Klebaner and Liptser also showed in [12] and [13] how the LDP can be used to get bounds of the asymptotics of the expected exit time of the process. Let the exit time ${\tau _{\varepsilon }}$ of the process be defined as
for a set Ω. For the expected exit time it holds that
if the right-hand side limit exists. Since
\[ {E_{{x_{0}},\dots ,{x_{m-1}}}}({\tau _{\varepsilon }})\le \frac{2M}{{\inf _{{x_{0}},\dots ,{x_{m-1}}\in \Omega }}{P_{{x_{0}},\dots ,{x_{m-1}}}}({\tau _{\varepsilon }}\le M)}\]
for any set of starting points ${x_{0}},\dots ,{x_{m-1}}\in \Omega $ and any integer $M\ge m$ (for details, see [11]). If the infimum in the denominator is attained for the starting points ${x_{0}^{\ast }},\dots ,{x_{m-1}^{\ast }}\in \Omega $, the inequality above implies that
(6)
\[ \underset{\varepsilon \to 0}{\limsup }q(\varepsilon )\log {E_{{x_{0}},\dots ,{x_{m-1}}}}({\tau _{\varepsilon }})\le -\underset{\varepsilon \to 0}{\lim }q(\varepsilon )\log {P_{{x_{0}^{\ast }},\dots ,{x_{m-1}^{\ast }}}}({\tau _{\varepsilon }}\le M),\]
\[ {P_{{x_{0}^{\ast }},\dots ,{x_{m-1}^{\ast }}}}({\tau _{\varepsilon }}\le M)={P_{{x_{0}^{\ast }},\dots ,{x_{m-1}^{\ast }}}}\big({X_{t}^{\varepsilon }}\notin \Omega \hspace{2.5pt}\text{for some}\hspace{2.5pt}t\in \{m,\dots ,M\}\big),\]
the limit on the right-hand side in (6) may be calculated if we have a large deviation principle for the family of probability measures induced by ${\{{X_{t}^{\varepsilon }}\}_{t\ge 0}}$ and if the function f and the set Ω are suitable.From this point onward in the paper, we consider a process of autoregressive type, where
and ${X_{0}}=0$. Here f is a continuous function on $\mathbb{R}$ and ${\{{\xi _{n}}\}_{n\ge 1}}$ is a sequence of i.i.d. standard normal random variables. Then $I(z)=\frac{{z^{2}}}{2}$ and the function g in (3) is reduced to
We consider exit times from the interval $(-1,1)$, so $\Omega =(-1,1)$.
(7)
\[ {X_{n+1}^{\varepsilon }}=f\big({X_{n}^{\varepsilon }}\big)+\varepsilon {\xi _{n+1}},\hspace{1em}n=0,1,2,\dots ,\]Klebaner and Liptser considered this case as an example in [12] and [13] (other examples were Poisson distributed noise, and sums of normally distributed and Poisson distributed random variables), and showed that this family of processes obeys the large deviation principle with $q(\varepsilon )={\varepsilon ^{2}}$ and
For the exit time
we then have
where the last equality holds because one can choose ${y_{n}}=f({y_{n-1}})$ for all $n\ge N+1$ and get the same infimum.
(8)
\[ I({y_{0}},{y_{1}},{y_{2}},\dots )=\frac{1}{2}{\sum \limits_{n=1}^{\infty }}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}.\](10)
\[\begin{aligned}{}& \underset{\varepsilon \to 0}{\limsup }{\varepsilon ^{2}}\log {E_{{x_{0}}}}{\tau _{\varepsilon }}\le -\underset{\varepsilon \to 0}{\lim }{\varepsilon ^{2}}\log {P_{{x_{0}^{\ast }}}}({\tau _{\varepsilon }}\le M)\\ {} & \hspace{1em}=\underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{{\max _{1\le n\le M}}|{y_{n}}|\ge 1}{{y_{0}}={x_{0}^{\ast }}}}{}}{\inf }I({y_{0}},{y_{1}},{y_{2}},\dots )=\underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{{\max _{1\le n\le M}}|{y_{n}}|\ge 1}{{y_{0}}={x_{0}^{\ast }}}}{}}{\inf }\frac{1}{2}{\sum \limits_{n=1}^{\infty }}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}\\ {} & \hspace{1em}=\underset{1\le N\le M}{\inf }\left(\underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{|{y_{N}}|\ge 1}{{y_{0}}={x_{0}^{\ast }}}}{|{y_{n}}|\lt 1,n=1,\dots ,N}}{\inf }\frac{1}{2}{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}\right),\end{aligned}\]For the autoregressive case with a linear function $f(x)=ax$, $|a|\lt 1$, and ${\tau _{\varepsilon }}$ as above, Klebaner and Liptser showed that
by minimizing the sum. This was done by considering the telescoping sum
when ${y_{0}}={x_{0}}=0$ and applying the Cauchy–Schwarz inequality to get
Here, $|{y_{N}}|=1$, and the result in (11) follows since M can be arbitrarily large. Note that if we instead were to study exits from a scaled interval $(-c,c)$, $c\gt 0$, so that ${\tau _{\varepsilon }^{c}}=\min \{n\ge 1:|{X_{n}^{\varepsilon }}|\ge c\}$, then $|{y_{N}}|=c$ and the upper bound would be
(11)
\[ \underset{\varepsilon \to 0}{\limsup }{\varepsilon ^{2}}\log {E_{{x_{0}}}}({\tau _{\varepsilon }})\le \frac{1-{a^{2}}}{2}\](12)
\[ {\sum \limits_{n=1}^{N}}{({y_{n}}-a{y_{n-1}})^{2}}\ge \frac{{y_{N}^{2}}}{{\textstyle\textstyle\sum _{n=1}^{N}}{a^{N-n}}}.\]3 Minimizing the sum
In the previous section we saw that the asymptotics for the exit times of processes of the type defined in (7) with $N(0,1)$-normal white noise is determined by the function f through the infimum of the sum of squares in (10). In this section we study some properties of these sums for particular classes of autoregression functions f. Since the infimum of the sum is attained for $|{y_{N}}|=1$, it does not matter how the function f is defined outside of the interval $[-1,1]$, and we only focus on how it is defined on $[-1,1]$.
Lemma 1.
If f is increasing on $[-1,1]$ and $f(0)=0$, the sum can be minimized separately over positive and negative values:
(14)
\[\begin{aligned}{}& \underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{|{y_{N}}|\ge 1}{{y_{0}}=0}}{}}{\inf }\frac{1}{2}{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}\end{aligned}\](15)
\[\begin{aligned}{}& \hspace{1em}=\min \left(\underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{{y_{N}}=1,{y_{0}}=0}{{y_{n}}\ge 0,n=1,\dots ,N-1}}{}}{\inf }\frac{1}{2}{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}},\right.\\ {} & \hspace{1em}\hspace{2em}\hspace{0.2778em}\hspace{0.2778em}\hspace{0.2778em}\hspace{0.2778em}\hspace{0.2778em}\hspace{0.2778em}\left.\underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{{y_{N}}=-1,{y_{0}}=0}{{y_{n}}\le 0,n=1,\dots ,N-1}}{}}{\inf }\frac{1}{2}{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}\right).\end{aligned}\]Proof.
Assume that the infimum of the sum is attained for ${\{{y_{n}^{\ast }}\}_{n=0}^{N}}$, where ${y_{0}^{\ast }}=0$ and ${y_{N}^{\ast }}=1$, and let
\[ {S^{\ast }}={\sum \limits_{n=1}^{N}}{\big({y_{n}^{\ast }}-f\big({y_{n-1}^{\ast }}\big)\big)^{2}}.\]
Then ${S^{\ast }}\le 1$. We will show by induction that ${y_{n}^{\ast }}\ge 0$ for $n=1,\dots ,N-1$. We show first that ${y_{N-1}^{\ast }}\ge 0$. If ${y_{N-1}^{\ast }}\lt 0$, then $f({y_{N-1}^{\ast }})\le 0$ and ${(1-f({y_{N-1}}))^{2}}\ge 1$. Also,
where $L=\min \{i|{y_{i}}\ne 0\}$. It follows that ${S^{\ast }}\gt 1$, which is a contradiction. Thus, ${y_{N-1}^{\ast }}\ge 0$.Now, assume that ${y_{N}^{\ast }},{y_{N-1}^{\ast }},\dots ,{y_{N-K+1}^{\ast }}\ge 0$ for some $K\lt N$. Make the contrary assumption that ${y_{N-K}^{\ast }}\lt 0$. Then $f({y_{N-K}^{\ast }})\le 0$ and
\[\begin{aligned}{}{S^{\ast }}& ={\big(1-f\big({y_{N-1}^{\ast }}\big)\big)^{2}}+\cdots +{\big({y_{N-K+2}^{\ast }}-f\big({y_{N-K+1}^{\ast }}\big)\big)^{2}}\\ {} & \hspace{1em}+{\big({y_{N-K+1}^{\ast }}-f\big({y_{N-K}^{\ast }}\big)\big)^{2}}+{\big({y_{N-K}^{\ast }}-f\big({y_{N-K-1}^{\ast }}\big)\big)^{2}}+\cdots +{y_{1}^{2}}\\ {} & \ge {\big(1-f\big({y_{N-1}^{\ast }}\big)\big)^{2}}+\cdots +{\big({y_{N-K+2}^{\ast }}-f\big({y_{N-K+1}^{\ast }}\big)\big)^{2}}\\ {} & \hspace{1em}+{\big({y_{N-K+1}^{\ast }}-0\big))^{2}}+0+\cdots +0.\end{aligned}\]
In fact, the inequality above is strict, since
\[\begin{aligned}{}{S^{\ast }}& \ge {\big(1-f\big({y_{N-1}^{\ast }}\big)\big)^{2}}+\cdots +{\big({y_{N-K+2}^{\ast }}-f\big({y_{N-K+1}^{\ast }}\big)\big)^{2}}\\ {} & \hspace{1em}+{\big({y_{N-K+1}^{\ast }}-0\big))^{2}}+{\big({y_{L}^{\ast }}\big)^{2}}\end{aligned}\]
where $L=\min \{i\le N-K|{y_{i}^{\ast }}\ne 0\}$. Thus, ${S^{\ast }}$ is not the minimal sum, which is a contradiction. It follows that ${y_{N-K}^{\ast }}\ge 0$.If we assume instead that the infimum on the left-hand side in (14) is attained for a sequence ${\{{y^{\prime }_{n}}\}_{n=0}^{N}}$, where ${y^{\prime }_{0}}=0$ and ${y^{\prime }_{N}}=-1$, one can show that ${y^{\prime }_{n}}\le 0$ for $n=1,\dots ,N$ in a similar way. □
Lemma 2.
If f is increasing on $[-1,1]$, $f(0)=0$ and f is odd, so that $f(-x)=-f(x)$ on $[-1,1]$, we can minimize over only positive values:
(16)
\[ \underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{|{y_{N}}|\ge 1}{{y_{0}}=0}}{}}{\inf }\frac{1}{2}{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}=\underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{{y_{N}}=1,{y_{0}}=0}{{y_{n}}\ge 0,n=1,\dots ,N-1}}{}}{\inf }\frac{1}{2}{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}.\]Lemma 3.
If f is increasing, $f(0)=0$ and $|f(x)|\lt |x|$ on $(-1,1)\setminus \{0\}$, the sum should be minimized separately over positive values and increasing sequences or negative values and decreasing sequences:
and
(17)
\[ \underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{{y_{N}}=1,{y_{0}}=0}{{y_{n}}\ge 0,n=1,\dots ,N-1}}{}}{\inf }\frac{1}{2}{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}=\underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{{y_{N}}=1,{y_{0}}=0}{{y_{0}}\le {y_{1}}\le {y_{2}}\le \cdots \le {y_{N}}}}{}}{\inf }\frac{1}{2}{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}\](18)
\[ \underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{{y_{N}}=-1,{y_{0}}=0}{{y_{n}}\le 0,n=1,\dots ,N-1}}{}}{\inf }\frac{1}{2}{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}=\underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{{y_{N}}=-1,{y_{0}}=0}{{y_{0}}\ge {y_{1}}\ge {y_{2}}\ge \cdots \ge {y_{N}}}}{}}{\inf }\frac{1}{2}{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}.\]Proof.
We prove equality (17). Assume that the sum ${\textstyle\sum _{n=1}^{N}}{({y_{n}}-f({y_{n-1}}))^{2}}$ is minimized by the sequence ${\{{y_{n}^{\ast }}\}_{n=0}^{N}}$, where ${y_{0}^{\ast }}=0$, ${y_{N}^{\ast }}=1$ and ${y_{n}^{\ast }}\in [0,1]$ for $n=1,\dots ,N-1$. We show by induction that ${y_{N}^{\ast }}\ge {y_{N-1}^{\ast }}\ge \cdots \ge {y_{1}^{\ast }}\ge {y_{0}^{\ast }}$. It is given that ${y_{N}^{\ast }}\ge {y_{N-1}^{\ast }}$. Assume that ${y_{N}^{\ast }}\ge {y_{N-1}^{\ast }}\ge \cdots \ge {y_{N-k+1}^{\ast }}$ for some k. We show that ${y_{N-k+1}^{\ast }}\ge {y_{N-k}^{\ast }}$.
If ${y_{N-k+1}^{\ast }}=0$, it is clear that the minimum of the sum is attained for ${y_{N-k}^{\ast }}={y_{N-k-1}^{\ast }}=\cdots ={y_{1}^{\ast }}={y_{0}^{\ast }}=0$. Then ${y_{N-k+1}^{\ast }}\ge {y_{N-k}^{\ast }}$.
If ${y_{N-k+1}^{\ast }}\gt 0$, make the contrary assumption that ${y_{N-k+1}^{\ast }}\lt {y_{N-k}^{\ast }}$. Then ${y_{N-k}^{\ast }}\in ({y_{N-m}^{\ast }},{y_{N-m+1}^{\ast }}]$ for some $m\in \{1,\dots ,k-1\}$. It follows that
because ${y_{N-m+1}^{\ast }}-f({y_{N-m}^{\ast }})\ge {y_{N-m+1}^{\ast }}-f({y_{N-k}^{\ast }})$. Equality is attained if
If this is true, f is constant on the interval $[{y_{N-m}^{\ast }},{y_{N-k}^{\ast }}]$, and $f({y_{N-m}^{\ast }})={y_{N-k+1}^{\ast }}$. Also,
(19)
\[\begin{aligned}{}{\sum \limits_{n=1}^{N}}{\big({y_{n}^{\ast }}-f\big({y_{n-1}^{\ast }}\big)\big)^{2}}& ={\big(1-f\big({y_{N-1}^{\ast }}\big)\big)^{2}}+\cdots +{\big({y_{N-m+2}^{\ast }}-f\big({y_{N-m+1}^{\ast }}\big)\big)^{2}}\\ {} & \hspace{1em}+{\big({y_{N-m+1}^{\ast }}-f\big({y_{N-m}^{\ast }}\big)\big)^{2}}+{\big({y_{N-m}^{\ast }}-f\big({y_{N-m-1}^{\ast }}\big)\big)^{2}}\\ {} & \hspace{1em}+\cdots +{\big({y_{N-k+1}^{\ast }}-f\big({y_{N-k}^{\ast }}\big)\big)^{2}}+{\big({y_{N-k}^{\ast }}-f\big({y_{N-k-1}^{\ast }}\big)\big)^{2}}\\ {} & \hspace{1em}+\cdots +{\big({y_{2}^{\ast }}-f\big({y_{1}^{\ast }}\big)\big)^{2}}+{\big({y_{1}^{\ast }}\big)^{2}}\\ {} & \ge {\big(1-f\big({y_{N-1}^{\ast }}\big)\big)^{2}}+\cdots +{\big({y_{N-m+2}^{\ast }}-f\big({y_{N-m+1}^{\ast }}\big)\big)^{2}}\\ {} & \hspace{1em}+{\big({y_{N-m+1}^{\ast }}-f\big({y_{N-k}^{\ast }}\big)\big)^{2}}+0+\cdots +0\\ {} & \hspace{1em}+{\big({y_{N-k}^{\ast }}-f\big({y_{N-k-1}^{\ast }}\big)\big)^{2}}+\cdots +{\big({y_{1}^{\ast }}\big)^{2}},\end{aligned}\](20)
\[ f\big({y_{N-m}^{\ast }}\big)=f\big({y_{N-k}^{\ast }}\big)\hspace{2.5pt}\text{and}\hspace{2.5pt}{y_{N-m}^{\ast }}=f\big({y_{N-m-1}^{\ast }}\big),\dots ,{y_{N-k+1}^{\ast }}=f\big({y_{N-k}^{\ast }}\big).\]
\[ f\big({y_{N-m}^{\ast }}\big)=f\big(f\big({y_{N-m-1}^{\ast }}\big)\big)=\cdots =f\big(f\big(\dots \big(f\big({y_{N-k+1}^{\ast }}\big)\big)\big)\big)\lt {y_{N-k+1}^{\ast }},\]
because ${y_{N-k+1}^{\ast }}\gt 0$. Thus, (20) does not hold, which implies that equality is not attained in (19). Then, the sum ${\textstyle\sum _{n=1}^{N}}{({y_{n}}-f({y_{n-1}}))^{2}}$ is not minimized by the sequence ${\{{y_{n}^{\ast }}\}_{n=0}^{N}}$, which is a contradiction. Thus, ${y_{N-k+1}^{\ast }}\ge {y_{N-k}^{\ast }}$. □Remark 1.
If f and g are as in Lemma 3 and $|f(x)|\le |g(x)|$ on $[-1,1]$, then
(21)
\[ \underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{|{y_{N}}|\ge 1}{{y_{0}}=0}}{}}{\inf }\frac{1}{2}{\sum \limits_{n=1}^{N}}{\big({y_{n}}-g({y_{n-1}})\big)^{2}}\le \underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{|{y_{N}}|\ge 1}{{y_{0}}=0}}{}}{\inf }\frac{1}{2}{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}.\]Proof.
By Lemma 3, the minimum on the right-hand side is attained for an increasing sequence ${\{{y_{n}}\}_{n=0}^{N}}$:
\[ \underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{{y_{N}}=1,{y_{0}}=0}{{y_{n}}\ge 0,n=1,\dots ,N-1}}{}}{\inf }\frac{1}{2}{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}=\underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{{y_{N}}=1,{y_{0}}=0}{{y_{0}}\le {y_{1}}\le \cdots \le {y_{N}}}}{}}{\inf }\frac{1}{2}{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}.\]
Now, for each $n=1,\dots ,N$,
and the same is true when f is replaced by g. Since $f(x)\le g(x)$ on $[0,1]$,
and it follows that
The statement of the lemma then follows. □4 Piecewise linear functions
Consider the process
where ${X_{0}}=0$, ${\{{\xi _{n}}\}_{n\ge 1}}$ is a sequence of i.i.d. standard normal random variables, ε is a small positive parameter and f is a continuous piecewise linear function. We consider exit times from the interval $(-1,1)$, so $\Omega =(-1,1)$ and
As in Section 3, the definition of f outside of $[-1,1]$ does not have an impact on the results in this section.
Proposition 1.
Let f be a function on $\mathbb{R}$ that satisfies
If $a=1$ and $0\le b\lt 1$,
for any $M\ge 1$.
\[ f(x)=\left\{\begin{array}{l@{\hskip10.0pt}l}a(x+b)\hspace{1em}& \mathit{if}\hspace{2.5pt}-1\le x\le -b,\\ {} 0\hspace{1em}& \mathit{if}\hspace{2.5pt}-b\lt x\lt b,\\ {} a(x-b)\hspace{1em}& \mathit{if}\hspace{2.5pt}b\le x\le 1,\end{array}\right.\]
where $0\le a\lt 1$ and $0\le b\lt 1$. Then, for any $M\ge 1$,
(24)
\[ \underset{\varepsilon \to 0}{\limsup }{\varepsilon ^{2}}\log {E_{{x_{0}}}}{\tau _{\varepsilon }}\le \frac{1}{2}\min \bigg(1,\underset{2\le N\le M}{\inf }\bigg(\big(1-{a^{2}}\big)\frac{{(1+\frac{a-{a^{N}}}{1-a}b)^{2}}}{1-{a^{2N}}}\bigg)\bigg).\](25)
\[ \underset{\varepsilon \to 0}{\limsup }{\varepsilon ^{2}}\log {E_{{x_{0}}}}{\tau _{\varepsilon }}\le \frac{1}{2}\underset{1\le N\le M}{\inf }\bigg(\frac{{(1+(N-1)b)^{2}}}{N}\bigg)\]Fig. 1.
The upper bound on the right-hand side in (24) illustrated for different values of a and b and $N=1$ (when the value is 1/2) and $N=2,\dots 8$. The dotted line shows the value $(1-{a^{2}})/2$ in each case
Proof.
We have
\[\begin{aligned}{}\underset{\genfrac{}{}{0.0pt}{}{|{y_{N}}|\ge 1}{{y_{0}}=0}}{\inf }{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}& =\underset{\genfrac{}{}{0.0pt}{}{|{y_{N}}|=1}{|{y_{0}}|=b}}{\inf }{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}\\ {} & =\underset{\genfrac{}{}{0.0pt}{}{{y_{N}}=1}{{y_{0}}=b}}{\inf }{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}},\end{aligned}\]
because $f(0)=f(b)$ and it is enough to take the infimum over positive values because of Lemma 2. If ${y_{N}}=1$, ${y_{0}}=b$ and ${c_{n}}={a^{N-n}}$ for $n=1,\dots ,N$, we have the following telescoping sum:
\[ {\sum \limits_{n=1}^{N}}{c_{n}}\big({y_{n}}-f({y_{n-1}})\big)=\left\{\begin{array}{l@{\hskip10.0pt}l}1\hspace{1em}& \text{if}\hspace{2.5pt}N=1,\\ {} 1+ab\frac{1-{a^{N-1}}}{1-a}\hspace{1em}& \text{if}\hspace{2.5pt}N\ge 2.\end{array}\right.\]
By the Cauchy–Schwarz inequality,
\[ {\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}\ge \frac{{\textstyle\textstyle\sum _{n=1}^{N}}{c_{n}}({y_{n}}-f({y_{n-1}}))}{({\textstyle\textstyle\sum _{n=1}^{N}}{c_{n}^{2}})},\]
where equality can be attained. Since ${\textstyle\sum _{n=1}^{N}}{c_{n}^{2}}=(1-{a^{2N}})/(1-{a^{2}})$, it follows that
\[ \underset{\varepsilon \to 0}{\limsup }{\varepsilon ^{2}}\log {E_{{x_{0}}}}{\tau _{\varepsilon }}\le \frac{1}{2}\min \bigg(1,\underset{2\le N\le M}{\inf }\bigg(\big(1-{a^{2}}\big)\frac{{(1+\frac{a-{a^{N}}}{1-a}b)^{2}}}{1-{a^{2N}}}\bigg)\bigg),\]
for any $M\ge 1$. The value of the infimum, as well as for which N it is attained, depends on the choices of a and b; in some cases the minimum is 1 and in some cases it it less than one.If we let $a=1$ and $0\lt b\lt 1$, we can use the same method as above with the telescoping sum. Then ${c_{n}}={a^{N-n}}=1$ for $n=1,\dots ,N$,
and the result is
Here, the infimum is 1 (which is attained for $N=1$) if $b\ge 1/3$. For $b\lt 1/3$, the optimal N is either $\big\lfloor \frac{1}{b}\big\rfloor -1$ or $\big\lfloor \frac{1}{b}\big\rfloor $. □
(26)
\[ \underset{\varepsilon \to 0}{\limsup }{\varepsilon ^{2}}\log {E_{{x_{0}}}}{\tau _{\varepsilon }}\le \frac{1}{2}\underset{1\le N\le M}{\inf }\frac{{(1+(N-1)b)^{2}}}{N}.\]Note that if $b=0$ in expression (24), the infimum is attained for $N=M$. The inequality holds for any $M\ge 1$ and also as $M\to \infty $. Then the right-hand side in (24) becomes $(1-{a^{2}})/2$, the result for the autoregressive process (11).
Proposition 2.
Let $a\in (0,1)$ and $c\in (0,1]$ and let f be a function that satisfies
\[ f(x)=\left\{\begin{array}{l@{\hskip10.0pt}l}-ac\hspace{1em}& \mathit{if}\hspace{2.5pt}-1\le x\lt -c,\\ {} ax\hspace{1em}& \mathit{if}\hspace{2.5pt}-c\le x\le c,\\ {} ac\hspace{1em}& \mathit{if}\hspace{2.5pt}c\lt x\le 1.\end{array}\right.\]
Then
(27)
\[ \underset{\varepsilon \to 0}{\limsup }{\varepsilon ^{2}}\log E{\tau _{\varepsilon }}\le \left\{\begin{array}{l@{\hskip10.0pt}l}\frac{1}{2}({(1-ac)^{2}}+(1-{a^{2}}){c^{2}}),\hspace{1em}& \textit{if}\hspace{2.5pt}c\le a,\\ {} \frac{1}{2}(1-{a^{2}}),\hspace{1em}& \textit{if}\hspace{2.5pt}c\ge a.\end{array}\right.\]Fig. 2.
The upper bound on the right-hand side in (27) drawn as a function of c for some chosen values of a. The dotted lines show the value $(1-{a^{2}})/2$ for each a
Proof.
By Lemma 2 and Lemma 3 we only need to minimize over positive and increasing sequences. We determine the infimum of
for ${y_{0}}=0$, ${y_{0}}\le {y_{1}}\le \cdots \le {y_{N-1}}\le {y_{N}}$ and ${y_{N}}=1$. The set of sequences which we minimize over can be split into two parts: Either ${y_{N-1}}\lt c$ or the d last elements in the sequence are larger than c: ${y_{N-d}}\ge c$. If ${y_{N-d}}\ge c$,
\[ {\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}\ge {(1-ac)^{2}}+(d-1){(c-ac)^{2}}+{\sum \limits_{n=1}^{N-d}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}},\]
and this lower bound is attained for ${y_{N-d+1}}=\cdots ={y_{N-1}}=c$. Minimizing the sum on the right-hand side above when $0\le {y_{1}}\le \cdots \le {y_{N-d-1}}\le c$ and ${y_{N-d}}\ge c$ is the same minimizing problem as in (13), since $f({y_{n-1}})=a{y_{n-1}}$ when ${y_{n-1}}\le c$. We get
\[ \underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{{y_{N}}=1,{y_{0}}=0,{y_{N-d}}\ge c}{{y_{0}}\le {y_{1}}\le \cdots \le {y_{N}}}}{}}{\inf }{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}={(1-ac)^{2}}+(d-1){(c-ac)^{2}}+\frac{(1-{a^{2}}){c^{2}}}{1-{a^{2(N-d)}}}.\]
This value is smallest if $d=1$, and it is then
On the other hand, if ${y_{N-1}}\lt c$,
\[\begin{aligned}{}& \underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{{y_{N}}=1,{y_{0}}=0,{y_{N-1}}\lt c}{{y_{0}}\le {y_{1}}\le \cdots \le {y_{N}}}}{}}{\inf }{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}\\ {} & \hspace{1em}=\underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{{y_{N}}=1,{y_{0}}=0,{y_{N-1}}\lt c}{{y_{0}}\le {y_{1}}\le \cdots \le {y_{N}}}}{}}{\inf }{\sum \limits_{n=1}^{N}}{({y_{n}}-a{y_{n-1}})^{2}}=\frac{1-{a^{2}}}{1-{a^{2N}}}\end{aligned}\]
if $a\lt c$, because the minimizing problem then coincides with the same problem for the autoregressive process. If $a\ge c$,
\[\begin{aligned}{}& \underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{{y_{N}}=1,{y_{0}}=0,{y_{N-1}}\lt c}{{y_{0}}\le {y_{1}}\le \cdots \le {y_{N}}}}{}}{\inf }{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}\\ {} & \hspace{1em}=\underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{{y_{N}}=1,{y_{0}}=0,{y_{N-1}}\lt c}{{y_{0}}\le {y_{1}}\le \cdots \le {y_{N}}}}{}}{\inf }\Bigg({(1-a{y_{n-1}})^{2}}+{\sum \limits_{n=1}^{N-1}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}\Bigg)\end{aligned}\]
where it is optimal to have ${y_{N-1}}$ close to c, and we get the same infimum as in (28). To summarize,
\[\begin{aligned}{}& \underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{{y_{N}}=1,{y_{0}}=0}{{y_{0}}\le {y_{1}}\le \cdots \le {y_{N}}}}{}}{\inf }{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}\\ {} & \hspace{1em}=\min \bigg\{{(1-ac)^{2}}+\frac{(1-{a^{2}}){c^{2}}}{1-{a^{2(N-1)}}},\frac{1-{a^{2}}}{1-{a^{2N}}}\bigg\}.\end{aligned}\]
Since
\[ \underset{\varepsilon \to 0}{\limsup }{\varepsilon ^{2}}\log {E_{{x_{0}}}}{\tau _{\varepsilon }}\le \underset{1\le N\le M}{\inf }\underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{{y_{N}}=1,{y_{0}}=0}{{y_{0}}\le {y_{1}}\le \cdots \le {y_{N}}}}{}}{\inf }{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}\]
for any positive integer M, we can let M be arbitrarily large. We get
\[ \underset{\varepsilon \to 0}{\limsup }{\varepsilon ^{2}}\log {E_{{x_{0}}}}{\tau _{\varepsilon }}\le \left\{\begin{array}{l@{\hskip10.0pt}l}\frac{1}{2}({(1-ac)^{2}}+(1-{a^{2}}){c^{2}}),\hspace{1em}& \text{if}\hspace{2.5pt}c\le a,\\ {} \frac{1}{2}(1-{a^{2}}),\hspace{1em}& \text{if}\hspace{2.5pt}c\ge a.\end{array}\right.\]
□Proposition 3.
Proof.
We study the sum
If ${y_{N}}=1$, then ${S_{N}}\ge 1$, since $f(x)\le 0$ for all $x\in [-1,1]$. Also, if ${y_{N}}=-1$ and ${y_{N-1}}\lt 0$, we have ${S_{N}}\ge 1$. In the case when ${y_{N}}=-1$ and ${y_{N-1}}\ge 0$, it is optimal to have ${y_{N-2}}=\cdots ={y_{1}}={y_{0}}=0$. The smallest value that the sum can take is the minimum of ${(-1+ax)^{2}}+{x^{2}}$ for $x\in [0,1]$, which is $1/(1+{a^{2}})$. Thus,
and it follows that
□
Note that if $a=0$ in Proposition 3, we have $f(x)=0$ on $[-1,1]$, which is a special case of the autoregressive process with $a=0$. The upper bound is then just $1/2$ which agrees with the result in the autoregressive case.
Proposition 4.
Proof.
Let
Consider the cases ${y_{N}}=1$ and ${y_{N}}=-1$ separately. First, let ${y_{N}}=-1$. Clearly, the sum S is smallest if the sequence ${\{{y_{n}}\}_{n=1}^{N}}$ has alternating signs: ${y_{N-1}}\gt 0,{y_{N-2}}\lt 0,\dots $ . Putting the derivative of S with respect to ${y_{n}}$ equal to zero gives
where ${f^{\prime }}({y_{n}})=-b$ if ${y_{n}}\lt 0$ and ${f^{\prime }}({y_{n}})=-a$ if ${y_{n}}\gt 0$. Let ${s_{n}}:={y_{n}}-f({y_{n-1}})$. Then, if $N=2M$ so that N is even, it is optimal to have ${y_{1}}\gt 0$. Then
\[\begin{aligned}{}{s_{2k}}& =-{a^{-k}}{b^{-k+1}}{s_{1}},\\ {} {s_{2k+1}}& ={a^{-k}}{b^{-k}}{s_{1}},\end{aligned}\]
and it follows that
\[\begin{aligned}{}{y_{N}}& =-{s_{1}}\big[{a^{M}}{b^{M-1}}+{a^{M-2}}{b^{M-1}}+{a^{M-2}}{b^{M-3}}+{a^{M-4}}{b^{M-3}}+\cdots \\ {} & \hspace{1em}+{a^{-M+2}}{b^{-M+1}}+{a^{-M}}{b^{-M+1}}\big].\end{aligned}\]
This is a sum of two geometric sums. Since ${y_{N}}=-1$,
\[ {s_{1}}=\frac{1}{(1+{a^{2}}){(ab)^{N}}{a^{-2}}{b^{-1}}}\cdot \frac{{(ab)^{-2}}-1}{{(ab)^{-2N}}-1}.\]
The value of the sum is
\[ S={\sum \limits_{n=1}^{N}}{s_{n}^{2}}=\frac{1-{(ab)^{2}}}{1+{a^{2}}}\cdot \frac{1}{1-{(ab)^{2N}}}.\]
If N is odd, it is optimal to have ${y_{1}}\lt 0$. Then the same calculations as above follow, if a is replaced by b and vice versa. It follows that the value of the sum is
\[ S={\sum \limits_{n=1}^{N}}{s_{n}^{2}}=\frac{1-{(ab)^{2}}}{1+{b^{2}}}\cdot \frac{1}{1-{(ab)^{2N}}}.\]
The cases when ${y_{N}}=1$ and N is odd or even, give the same values of the sum. The smallest values are attained when N is large, and the statement of the proposition follows. □We note that when $a=b$ in Proposition 4, we have the autoregressive process with $f(x)=-ax$ on $[-1,1]$. The upper bound in Proposition 4 reduces to $1/(2(1-{a^{2}}))$ which is the bound for the autoregressive process.
Proof.
If $0\le a\lt 1$,
and we have the same infimum as in the autoregressive case (which was treated in [12] and [13]). The case $-1\lt a\le 0$ is treated in a similar way. □
(32)
\[ \underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{|{y_{N}}|\ge 1}{{y_{0}}=0}}{}}{\inf }\frac{1}{2}{\sum \limits_{n=1}^{N}}{({y_{n}}-a|{y_{n-1}}|)^{2}}=\underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{{y_{N}}=1,{y_{0}}=0}{{y_{n}}\ge 0,n=1,\dots ,N-1}}{}}{\inf }\frac{1}{2}{\sum \limits_{n=1}^{N}}{({y_{n}}-a{y_{n-1}})^{2}},\]5 Other nonlinear functions
For a more general function f it is not always possible to minimize the sum and get explicit upper bounds. Numerical calculations may be needed. It is also possible to construct (nonstrict) upper bounds by simply evaluating the sum for some sequence of values instead of actually minimizing it. This is illustrated in the following case. Consider the case when f is quadratic, so we have the process
where ${X_{0}}=0$, ${\{{\xi _{n}}\}_{n\ge 1}}$ is a sequence of i.i.d. standard normal random variables, ε is a small positive parameter and $f(x)=a{x^{2}}$.
Proposition 5.
Proof.
Since $f(x)\ge 0$ and f is even, it is optimal to have ${y_{N}}=1$ and ${y_{n}}\ge 0$ $\forall n$. We have
\[\begin{aligned}{}{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n}})\big)^{2}}& ={\sum \limits_{n=1}^{N}}{\big({y_{n}}-a{y_{n-1}^{2}}\big)^{2}}\\ {} & =1+{y_{n-1}^{2}}(1-2a)+{y_{n-2}^{2}}(1-2a{y_{n-1}})+\cdots +{y_{1}^{2}}(1-2a{y_{2}})\\ {} & \hspace{1em}+{a^{2}}\big({y_{n-1}^{4}}+{y_{n-2}^{4}}+\cdots +{y_{2}^{2}}\big)\ge 1\end{aligned}\]
when $0\le a\le 0.5$, and equality is achieved by putting ${y_{1}}={y_{2}}=\cdots ={y_{N-1}}=0$.For $a\ge 0.5$,
\[ \underset{1\le N\le M}{\inf }\left(\underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{|{y_{N}}|\ge 1,{y_{0}}=0}{|{y_{n}}|\lt 1,n=1,2,\dots ,n-1}}{}}{\inf }\frac{1}{2}{\sum \limits_{n=1}^{N}}{\big({y_{n}}-a{y_{n-1}^{2}}\big)^{2}}\right)\le \underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{{y_{2}}=1,{y_{0}}=0}{|{y_{1}}|\lt 1,}}{}}{\inf }\frac{1}{2}{\sum \limits_{n=1}^{2}}{\big({y_{n}}-a{y_{n-1}^{2}}\big)^{2}},\]
where the infimum on right-hand side is $\frac{1}{2}(\frac{1}{a}-\frac{1}{4{a^{2}}})$ (it is attained for ${y_{1}}=\sqrt{\frac{1}{a}-\frac{1}{2{a^{2}}}}$). This gives the upper bound in (34). □This is not necessarily the best upper bound for all $a\ge 0.5$, since we have only calculated the infimum for $N=2$, but at least we get an upper bound. Numerical calculations suggest that the bound in (34) is the best one for a slightly larger than 0.5, and that the optimal choice of N increases as a increases.
For another example of a nonlinear case, consider the classical deterministic Ricker model defined by
where ${x_{t}}$ represents the size or density of a population at time t, $r\gt 0$ models the growth rate and $\gamma \gt 0$ is an environmental factor [9]. By suitably renorming the population, we may take $\gamma =1$. The model then has a fixed point at $x=r$ (and one at $x=0$). If we introduce stochasticity in the model by adding normally distributed white noise, and move the fixed point $x=r$ to the origin, we have the process
where $f(x)=(x+r){e^{-x}}-r$ and ${\{{\xi _{n}}\}_{n\ge 1}}$ is a sequence of i.i.d. $N(0,1)$ variables. We can examine the time until the process exits from a suitable neighbourhood of the origin.
Fig. 3.
The function $f(x)=x{e^{r-x}}$ for $r=1.5$. We have a fixed point at $x=r$. On the right, we see a part of the plot with the fixed point moved to the origin
If $r=1.5$, consider for example the time until exit from the interval $[-0.5,0.5]$. Numerical calculations of the infimum
\[ \underset{1\le N\le M}{\inf }\left(\underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{|{y_{N}}|\ge 0.5}{{y_{0}}=0}}{|{y_{n}}|\lt 0.5,n=1,\dots ,N}}{\inf }\frac{1}{2}{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}\right)\]
give the approximate value 0.09, so that
\[ \underset{\varepsilon \to 0}{\limsup }{\varepsilon ^{2}}\log E{\tau _{\varepsilon }}\lessapprox 0.09.\]
The derivative of f at the origin is $1-r$, so a suitable linear approximation of the function f is $l(x)=-0.5x$. By replacing f by l, we have an autoregressive process for which the upper bound of the exit time is
We note that a linear approximation might give good enough approximations of the upper bounds, in cases when the neighbourhood around the origin is chosen to be rather small.6 Connection to stationary distribution
Mark Kac proved in 1947 that the mean return time of a discrete Markov chain to a point x is the reciprocal of the invariant probability $\pi (x)$. In [7], we explored this idea by comparing the exit time for the process defined by the stochastic difference equation
and the return time to a certain set for the same process. We get an upper bound for the asymptotics of the exit time, and that bound is the reciprocal of the stationary distribution of the process, evaluated at the point where the level curve of the stationary distribution touches the boundary of the set (or interval in the univariate case) from which the process exits.
In the univariate case with $f(x)=ax$, $x\in \mathbb{R}$, (that is, for the autoregressive process), this method gives the same upper bound as the LDP method:
These methods only give an upper bound, but we know that the bound is sharp in the autoregressive case; the corresponding lower bound can be shown by other methods – see [18], where Novikov’s martingale method ([16] and [17]) was used, and [11].
In Section 4, we saw several examples of processes of the type defined in (35) where f was a piecewise linear function. For these examples, it is not straightforward to derive expressions for the stationary distributions of the processes. We also note that in Section 4, where the LDP method was used, the definition of f outside of the interval $[-1,1]$ could be ignored. However, when calculating stationary distributions, the definition of f outside of $[-1,1]$ matters a great deal.
We observe that there is another piecewise linear case for which the stationary distribution is known, and this is a TAR(1) process with a threshold of 0, studied by Anděl et al. [1], where
with $f(x)=-|ax|$ for $|a|\lt 1$. In [1], they gave the following formula for the stationary distribution of this process:
\[ \frac{1}{\varepsilon }{\bigg(\frac{2(1-{a^{2}})}{\pi }\bigg)^{1/2}}\exp \bigg(-\frac{1}{2}\big(1-{a^{2}}\big)\frac{{x^{2}}}{{\varepsilon ^{2}}}\bigg)\Phi \bigg(\frac{-ax}{\varepsilon }\bigg),\]
where Φ is the cumulative distribution function of a $N(0,1)$ random variable. We note that
\[ -{\varepsilon ^{2}}\log \bigg(\frac{1}{\varepsilon }{\bigg(\frac{2(1-{a^{2}})}{\pi }\bigg)^{1/2}}\exp \bigg(-\frac{1}{2}\big(1-{a^{2}}\big)\frac{{x^{2}}}{{\varepsilon ^{2}}}\bigg)\Phi \bigg(\frac{-ax}{\varepsilon }\bigg)\bigg)\to \frac{1-{a^{2}}}{2}\]
at the point $x=-1$. This means that in this particular case, the value of the limit of the stationary distribution, evaluated at the point where the process exits the interval, coincides with the upper bound achieved by use of the LDP method in Remark 2.A preprint of a previous version of this paper has been posted on ArXiv [10].