Modern Stochastics: Theory and Applications logo


  • Help
Login Register

  1. Home
  2. To appear
  3. Exit times for some nonlinear autoregres ...

Modern Stochastics: Theory and Applications

Submit your article Information Become a Peer-reviewer
  • Article info
  • Full article
  • More
    Article info Full article

Exit times for some nonlinear autoregressive processes
Göran Högnäs   Brita Jung ORCID icon link to view author Brita Jung details  

Authors

 
Placeholder
https://doi.org/10.15559/25-VMSTA277
Pub. online: 25 March 2025      Type: Research Article      Open accessOpen Access

Received
29 August 2024
Revised
10 January 2025
Accepted
9 March 2025
Published
25 March 2025

Abstract

The expected exit time from the interval $[-1,1]$ is investigated for an autoregressive process defined recursively by
\[ {X_{n+1}^{\varepsilon }}=f\big({X_{n}^{\varepsilon }}\big)+\varepsilon {\xi _{n+1}},\hspace{1em}n=0,1,2,\dots ,\hspace{2.5pt}{X_{0}}=0.\]
Here, ε is a small positive parameter, $f:\mathbb{R}\mapsto \mathbb{R}$ is usually a contractive function and ${\{{\xi _{n}}\}_{n\ge 1}}$ is a sequence of i.i.d. random variables. In this paper, previous results for a linear function $f(x)=ax$ are extended to more general cases, with the main focus on piecewise linear functions.

1 Introduction

Consider a stochastic process ${\{{X_{n}^{\varepsilon }}\}_{n=0}^{\infty }}$ of autoregressive type, defined by
(1)
\[ {X_{n+1}^{\varepsilon }}=f\big({X_{n}^{\varepsilon }}\big)+\varepsilon {\xi _{n+1}},\hspace{1em}{X_{0}}=0,\]
where f is a continuous mapping from $\mathbb{R}$ to itself with a fixed point at the origin, ${\{{\xi _{n}}\}_{n=1}^{\infty }}$ is a sequence of i.i.d. random variables and ε is a small positive parameter. It is a Markov chain, and under suitable assumptions on ${\{{\xi _{n}}\}_{n=1}^{\infty }}$ and f, it is interesting to investigate how long it takes for the process to leave a neighbourhood of the origin.
The original motivation behind our work is to study the time until extinction of a population. A stochastic process that models a population may be positive recurrent and stay at a certain level (or carrying capacity) for a very long time, and when extinction happens, the process first leaves a neighbourhood around that level. Populations can be modeled by, for example, branching processes (such as those treated in [5] and [6]) or models such as the Ricker model which we will use as an example at the end of the paper; it has been studied in [9].
In [12] and the updated version [13], Klebaner and Liptser used the large deviation principle to get an upper bound on the exit time from a set for a process. As an example, they considered the linear autoregressive process, defined as in (1) with $f(x)=ax$, $|a|\lt 1$, and normally distributed innovations. For this example, they showed that the exit time ${\tau _{\varepsilon }}$ from the interval $(-1,1)$ satisfies
(2)
\[ \underset{\varepsilon \to 0}{\limsup }{\varepsilon ^{2}}\log {E_{{x_{0}}}}({\tau _{\varepsilon }})\le \frac{1-{a^{2}}}{2}.\]
The corresponding lower bound has been found by other methods [18]. This means that the upper bound on the right-hand side above is the best possible one.
We have studied a corresponding multivariate case [11], and the results were extended to the ARMA model in [15]. Exit times for autoregressive processes with other noise distributions have also been studied, in [8]. Different aspects of the exit time problem for linear autoregressive processes are treated in, for example, [2–4] and [17]. A related exit problem in a different setting was treated in [14]. A case with piecewise linear function $f(x)$ was studied by Anděl et al. in [1].
In this paper we extend the previous results from the linear case to some cases with piecewise linear functions. We will show that the large deviation principle gives explicit (asymptotic) upper bounds on the expectation of the exit time also in these cases. We also apply the methods used to other nonlinear functions, such as quadratic functions and the Ricker model.
In Section 2, we summarize how the large deviation method results in a sum that should be minimized for an upper bound of the exit time. This is based on the methods used in the proofs of Theorems 2.2 and 3.1 in [13]. In Section 3 we study how the minimization can be done over more restricted sets. In Section 4, which contains the main results of this paper, we get the explicit upper bounds in several piecewise linear cases. In Section 5 we explore some other nonlinear cases. In Section 6 we point out a connection between the results and the stationary distribution of the process.

2 Large deviation tools and bounds for exit times

In this section, which is based on work by Klebaner and Liptser in [12] and Jung in [11], we summarize how the large deviation principle (LDP) can be used to get an upper bound of the asymptotics of an exit time from a set for a process. The definition of the LDP used in [12] is as follows (this follows Varadhan’s definition in [19], with the addition that the rate of speed $q(\varepsilon )$ is a function of ε such that $q(\varepsilon )\to 0$ as $\varepsilon \to 0$).
Let $\{{P_{\varepsilon }}\}$ be a family of probability measures on the Borel subsets of a complete separable metric space Z. The family $\{{P_{\varepsilon }}\}$ satisfies the large deviation principle with a rate function I if there is a function I from Z into $[0,\infty ]$ that satisfies the following conditions: $0\le I(z)\le \infty \hspace{0.2778em}\forall z\in Z$, I is lower semicontinuous, the set $\{z:I(z)\le l\}$ is a compact set in Z for all $\hspace{0.2778em}l\lt \infty $ and
\[\begin{array}{r@{\hskip10.0pt}c@{\hskip10.0pt}l}& & \displaystyle \underset{\varepsilon \to 0}{\limsup }q(\varepsilon )\log {P_{\varepsilon }}(C)\le -\underset{z\in C}{\inf }I(z)\hspace{0.2778em}\hspace{0.2778em}\hspace{2.5pt}\text{for every closed set}\hspace{2.5pt}C\subset Z\hspace{2.5pt}\text{and}\hspace{2.5pt}\\ {} & & \displaystyle \underset{\varepsilon \to 0}{\liminf }q(\varepsilon )\log {P_{\varepsilon }}(G)\ge -\underset{z\in G}{\inf }I(z)\hspace{0.2778em}\hspace{0.2778em}\hspace{2.5pt}\text{for every open set}\hspace{2.5pt}G\subset Z.\end{array}\]
In [12] and [13], Klebaner and Liptser considered a family of processes of the type
(3)
\[ {X_{n+1}^{\varepsilon }}=g\big({X_{n}^{\varepsilon }},\dots ,{X_{n-m+1}^{\varepsilon }},\varepsilon {\xi _{n+1}}\big),\]
where g is a continuous function on ${\mathbb{R}^{m}}$, ${\{{\xi _{n}}\}_{n=m}^{\infty }}$ is a sequence of i.i.d. random variables and ${x_{0}},\dots ,{x_{m-1}}$ are given starting values. They gave conditions under which the LDP holds for the family $\varepsilon \xi $ (where ξ is a copy of ${\xi _{m}}$) and proved that when $\varepsilon \xi $ obeys an LDP with rate function $I(z)$, it follows that $({X_{n}^{\varepsilon }})$ obeys an LDP with a rate function $J(\bar{y})$ that can be written explicitly using $I(z)$:
(4)
\[ J(\bar{y})={\sum \limits_{k=m}^{\infty }}\underset{{v_{k}}:{y_{k}}=f({y_{k-1}},\dots {y_{k-m}},{v_{k}})}{\inf }I({y_{k}})\hspace{2.5pt}\text{when}\hspace{2.5pt}{u_{0}}={x_{0}},\dots ,{u_{m-1}}={x_{m-1}},\]
and $J(\bar{y})=\infty $ otherwise.
Klebaner and Liptser also showed in [12] and [13] how the LDP can be used to get bounds of the asymptotics of the expected exit time of the process. Let the exit time ${\tau _{\varepsilon }}$ of the process be defined as
(5)
\[ {\tau _{\varepsilon }}:=\min \big\{t\ge m:{X_{n}^{\varepsilon }}\notin \Omega \big\}\]
for a set Ω. For the expected exit time it holds that
\[ {E_{{x_{0}},\dots ,{x_{m-1}}}}({\tau _{\varepsilon }})\le \frac{2M}{{\inf _{{x_{0}},\dots ,{x_{m-1}}\in \Omega }}{P_{{x_{0}},\dots ,{x_{m-1}}}}({\tau _{\varepsilon }}\le M)}\]
for any set of starting points ${x_{0}},\dots ,{x_{m-1}}\in \Omega $ and any integer $M\ge m$ (for details, see [11]). If the infimum in the denominator is attained for the starting points ${x_{0}^{\ast }},\dots ,{x_{m-1}^{\ast }}\in \Omega $, the inequality above implies that
(6)
\[ \underset{\varepsilon \to 0}{\limsup }q(\varepsilon )\log {E_{{x_{0}},\dots ,{x_{m-1}}}}({\tau _{\varepsilon }})\le -\underset{\varepsilon \to 0}{\lim }q(\varepsilon )\log {P_{{x_{0}^{\ast }},\dots ,{x_{m-1}^{\ast }}}}({\tau _{\varepsilon }}\le M),\]
if the right-hand side limit exists. Since
\[ {P_{{x_{0}^{\ast }},\dots ,{x_{m-1}^{\ast }}}}({\tau _{\varepsilon }}\le M)={P_{{x_{0}^{\ast }},\dots ,{x_{m-1}^{\ast }}}}\big({X_{t}^{\varepsilon }}\notin \Omega \hspace{2.5pt}\text{for some}\hspace{2.5pt}t\in \{m,\dots ,M\}\big),\]
the limit on the right-hand side in (6) may be calculated if we have a large deviation principle for the family of probability measures induced by ${\{{X_{t}^{\varepsilon }}\}_{t\ge 0}}$ and if the function f and the set Ω are suitable.
From this point onward in the paper, we consider a process of autoregressive type, where
(7)
\[ {X_{n+1}^{\varepsilon }}=f\big({X_{n}^{\varepsilon }}\big)+\varepsilon {\xi _{n+1}},\hspace{1em}n=0,1,2,\dots ,\]
and ${X_{0}}=0$. Here f is a continuous function on $\mathbb{R}$ and ${\{{\xi _{n}}\}_{n\ge 1}}$ is a sequence of i.i.d. standard normal random variables. Then $I(z)=\frac{{z^{2}}}{2}$ and the function g in (3) is reduced to
\[ g({y_{n-1}},\dots ,{y_{n-m+1}},{z_{n}})=f({y_{n-1}})+{z_{n}}.\]
We consider exit times from the interval $(-1,1)$, so $\Omega =(-1,1)$.
Klebaner and Liptser considered this case as an example in [12] and [13] (other examples were Poisson distributed noise, and sums of normally distributed and Poisson distributed random variables), and showed that this family of processes obeys the large deviation principle with $q(\varepsilon )={\varepsilon ^{2}}$ and
(8)
\[ I({y_{0}},{y_{1}},{y_{2}},\dots )=\frac{1}{2}{\sum \limits_{n=1}^{\infty }}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}.\]
For the exit time
(9)
\[ {\tau _{\varepsilon }}=\min \big\{n\ge 1:|{X_{n}^{\varepsilon }}|\ge 1\big\},\]
we then have
(10)
\[\begin{aligned}{}& \underset{\varepsilon \to 0}{\limsup }{\varepsilon ^{2}}\log {E_{{x_{0}}}}{\tau _{\varepsilon }}\le -\underset{\varepsilon \to 0}{\lim }{\varepsilon ^{2}}\log {P_{{x_{0}^{\ast }}}}({\tau _{\varepsilon }}\le M)\\ {} & \hspace{1em}=\underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{{\max _{1\le n\le M}}|{y_{n}}|\ge 1}{{y_{0}}={x_{0}^{\ast }}}}{}}{\inf }I({y_{0}},{y_{1}},{y_{2}},\dots )=\underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{{\max _{1\le n\le M}}|{y_{n}}|\ge 1}{{y_{0}}={x_{0}^{\ast }}}}{}}{\inf }\frac{1}{2}{\sum \limits_{n=1}^{\infty }}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}\\ {} & \hspace{1em}=\underset{1\le N\le M}{\inf }\left(\underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{|{y_{N}}|\ge 1}{{y_{0}}={x_{0}^{\ast }}}}{|{y_{n}}|\lt 1,n=1,\dots ,N}}{\inf }\frac{1}{2}{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}\right),\end{aligned}\]
where the last equality holds because one can choose ${y_{n}}=f({y_{n-1}})$ for all $n\ge N+1$ and get the same infimum.
For the autoregressive case with a linear function $f(x)=ax$, $|a|\lt 1$, and ${\tau _{\varepsilon }}$ as above, Klebaner and Liptser showed that
(11)
\[ \underset{\varepsilon \to 0}{\limsup }{\varepsilon ^{2}}\log {E_{{x_{0}}}}({\tau _{\varepsilon }})\le \frac{1-{a^{2}}}{2}\]
by minimizing the sum. This was done by considering the telescoping sum
\[ {\sum \limits_{n=1}^{N}}{a^{N-n}}({y_{n}}-a{y_{n-1}})={y_{N}}\]
when ${y_{0}}={x_{0}}=0$ and applying the Cauchy–Schwarz inequality to get
(12)
\[ {\sum \limits_{n=1}^{N}}{({y_{n}}-a{y_{n-1}})^{2}}\ge \frac{{y_{N}^{2}}}{{\textstyle\textstyle\sum _{n=1}^{N}}{a^{N-n}}}.\]
Here, $|{y_{N}}|=1$, and the result in (11) follows since M can be arbitrarily large. Note that if we instead were to study exits from a scaled interval $(-c,c)$, $c\gt 0$, so that ${\tau _{\varepsilon }^{c}}=\min \{n\ge 1:|{X_{n}^{\varepsilon }}|\ge c\}$, then $|{y_{N}}|=c$ and the upper bound would be
(13)
\[ \underset{\varepsilon \to 0}{\limsup }{\varepsilon ^{2}}\log {E_{{x_{0}}}}\big({\tau _{\varepsilon }^{c}}\big)\le \frac{(1-{a^{2}}){c^{2}}}{2}.\]

3 Minimizing the sum

In the previous section we saw that the asymptotics for the exit times of processes of the type defined in (7) with $N(0,1)$-normal white noise is determined by the function f through the infimum of the sum of squares in (10). In this section we study some properties of these sums for particular classes of autoregression functions f. Since the infimum of the sum is attained for $|{y_{N}}|=1$, it does not matter how the function f is defined outside of the interval $[-1,1]$, and we only focus on how it is defined on $[-1,1]$.
Lemma 1.
If f is increasing on $[-1,1]$ and $f(0)=0$, the sum can be minimized separately over positive and negative values:
(14)
\[\begin{aligned}{}& \underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{|{y_{N}}|\ge 1}{{y_{0}}=0}}{}}{\inf }\frac{1}{2}{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}\end{aligned}\]
(15)
\[\begin{aligned}{}& \hspace{1em}=\min \left(\underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{{y_{N}}=1,{y_{0}}=0}{{y_{n}}\ge 0,n=1,\dots ,N-1}}{}}{\inf }\frac{1}{2}{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}},\right.\\ {} & \hspace{1em}\hspace{2em}\hspace{0.2778em}\hspace{0.2778em}\hspace{0.2778em}\hspace{0.2778em}\hspace{0.2778em}\hspace{0.2778em}\left.\underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{{y_{N}}=-1,{y_{0}}=0}{{y_{n}}\le 0,n=1,\dots ,N-1}}{}}{\inf }\frac{1}{2}{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}\right).\end{aligned}\]
Proof.
Assume that the infimum of the sum is attained for ${\{{y_{n}^{\ast }}\}_{n=0}^{N}}$, where ${y_{0}^{\ast }}=0$ and ${y_{N}^{\ast }}=1$, and let
\[ {S^{\ast }}={\sum \limits_{n=1}^{N}}{\big({y_{n}^{\ast }}-f\big({y_{n-1}^{\ast }}\big)\big)^{2}}.\]
Then ${S^{\ast }}\le 1$. We will show by induction that ${y_{n}^{\ast }}\ge 0$ for $n=1,\dots ,N-1$. We show first that ${y_{N-1}^{\ast }}\ge 0$. If ${y_{N-1}^{\ast }}\lt 0$, then $f({y_{N-1}^{\ast }})\le 0$ and ${(1-f({y_{N-1}}))^{2}}\ge 1$. Also,
\[ {S^{\ast }}\ge {\big(1-f\big({y_{N-1}^{\ast }}\big)\big)^{2}}+{\big({y_{L}^{\ast }}\big)^{2}},\]
where $L=\min \{i|{y_{i}}\ne 0\}$. It follows that ${S^{\ast }}\gt 1$, which is a contradiction. Thus, ${y_{N-1}^{\ast }}\ge 0$.
Now, assume that ${y_{N}^{\ast }},{y_{N-1}^{\ast }},\dots ,{y_{N-K+1}^{\ast }}\ge 0$ for some $K\lt N$. Make the contrary assumption that ${y_{N-K}^{\ast }}\lt 0$. Then $f({y_{N-K}^{\ast }})\le 0$ and
\[\begin{aligned}{}{S^{\ast }}& ={\big(1-f\big({y_{N-1}^{\ast }}\big)\big)^{2}}+\cdots +{\big({y_{N-K+2}^{\ast }}-f\big({y_{N-K+1}^{\ast }}\big)\big)^{2}}\\ {} & \hspace{1em}+{\big({y_{N-K+1}^{\ast }}-f\big({y_{N-K}^{\ast }}\big)\big)^{2}}+{\big({y_{N-K}^{\ast }}-f\big({y_{N-K-1}^{\ast }}\big)\big)^{2}}+\cdots +{y_{1}^{2}}\\ {} & \ge {\big(1-f\big({y_{N-1}^{\ast }}\big)\big)^{2}}+\cdots +{\big({y_{N-K+2}^{\ast }}-f\big({y_{N-K+1}^{\ast }}\big)\big)^{2}}\\ {} & \hspace{1em}+{\big({y_{N-K+1}^{\ast }}-0\big))^{2}}+0+\cdots +0.\end{aligned}\]
In fact, the inequality above is strict, since
\[\begin{aligned}{}{S^{\ast }}& \ge {\big(1-f\big({y_{N-1}^{\ast }}\big)\big)^{2}}+\cdots +{\big({y_{N-K+2}^{\ast }}-f\big({y_{N-K+1}^{\ast }}\big)\big)^{2}}\\ {} & \hspace{1em}+{\big({y_{N-K+1}^{\ast }}-0\big))^{2}}+{\big({y_{L}^{\ast }}\big)^{2}}\end{aligned}\]
where $L=\min \{i\le N-K|{y_{i}^{\ast }}\ne 0\}$. Thus, ${S^{\ast }}$ is not the minimal sum, which is a contradiction. It follows that ${y_{N-K}^{\ast }}\ge 0$.
If we assume instead that the infimum on the left-hand side in (14) is attained for a sequence ${\{{y^{\prime }_{n}}\}_{n=0}^{N}}$, where ${y^{\prime }_{0}}=0$ and ${y^{\prime }_{N}}=-1$, one can show that ${y^{\prime }_{n}}\le 0$ for $n=1,\dots ,N$ in a similar way.  □
Lemma 2.
If f is increasing on $[-1,1]$, $f(0)=0$ and f is odd, so that $f(-x)=-f(x)$ on $[-1,1]$, we can minimize over only positive values:
(16)
\[ \underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{|{y_{N}}|\ge 1}{{y_{0}}=0}}{}}{\inf }\frac{1}{2}{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}=\underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{{y_{N}}=1,{y_{0}}=0}{{y_{n}}\ge 0,n=1,\dots ,N-1}}{}}{\inf }\frac{1}{2}{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}.\]
Proof.
This follows immediately from Lemma 1, since the two infima on the right-hand side in (14) have the same value.  □
Lemma 3.
If f is increasing, $f(0)=0$ and $|f(x)|\lt |x|$ on $(-1,1)\setminus \{0\}$, the sum should be minimized separately over positive values and increasing sequences or negative values and decreasing sequences:
(17)
\[ \underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{{y_{N}}=1,{y_{0}}=0}{{y_{n}}\ge 0,n=1,\dots ,N-1}}{}}{\inf }\frac{1}{2}{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}=\underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{{y_{N}}=1,{y_{0}}=0}{{y_{0}}\le {y_{1}}\le {y_{2}}\le \cdots \le {y_{N}}}}{}}{\inf }\frac{1}{2}{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}\]
and
(18)
\[ \underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{{y_{N}}=-1,{y_{0}}=0}{{y_{n}}\le 0,n=1,\dots ,N-1}}{}}{\inf }\frac{1}{2}{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}=\underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{{y_{N}}=-1,{y_{0}}=0}{{y_{0}}\ge {y_{1}}\ge {y_{2}}\ge \cdots \ge {y_{N}}}}{}}{\inf }\frac{1}{2}{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}.\]
Proof.
We prove equality (17). Assume that the sum ${\textstyle\sum _{n=1}^{N}}{({y_{n}}-f({y_{n-1}}))^{2}}$ is minimized by the sequence ${\{{y_{n}^{\ast }}\}_{n=0}^{N}}$, where ${y_{0}^{\ast }}=0$, ${y_{N}^{\ast }}=1$ and ${y_{n}^{\ast }}\in [0,1]$ for $n=1,\dots ,N-1$. We show by induction that ${y_{N}^{\ast }}\ge {y_{N-1}^{\ast }}\ge \cdots \ge {y_{1}^{\ast }}\ge {y_{0}^{\ast }}$. It is given that ${y_{N}^{\ast }}\ge {y_{N-1}^{\ast }}$. Assume that ${y_{N}^{\ast }}\ge {y_{N-1}^{\ast }}\ge \cdots \ge {y_{N-k+1}^{\ast }}$ for some k. We show that ${y_{N-k+1}^{\ast }}\ge {y_{N-k}^{\ast }}$.
If ${y_{N-k+1}^{\ast }}=0$, it is clear that the minimum of the sum is attained for ${y_{N-k}^{\ast }}={y_{N-k-1}^{\ast }}=\cdots ={y_{1}^{\ast }}={y_{0}^{\ast }}=0$. Then ${y_{N-k+1}^{\ast }}\ge {y_{N-k}^{\ast }}$.
If ${y_{N-k+1}^{\ast }}\gt 0$, make the contrary assumption that ${y_{N-k+1}^{\ast }}\lt {y_{N-k}^{\ast }}$. Then ${y_{N-k}^{\ast }}\in ({y_{N-m}^{\ast }},{y_{N-m+1}^{\ast }}]$ for some $m\in \{1,\dots ,k-1\}$. It follows that
(19)
\[\begin{aligned}{}{\sum \limits_{n=1}^{N}}{\big({y_{n}^{\ast }}-f\big({y_{n-1}^{\ast }}\big)\big)^{2}}& ={\big(1-f\big({y_{N-1}^{\ast }}\big)\big)^{2}}+\cdots +{\big({y_{N-m+2}^{\ast }}-f\big({y_{N-m+1}^{\ast }}\big)\big)^{2}}\\ {} & \hspace{1em}+{\big({y_{N-m+1}^{\ast }}-f\big({y_{N-m}^{\ast }}\big)\big)^{2}}+{\big({y_{N-m}^{\ast }}-f\big({y_{N-m-1}^{\ast }}\big)\big)^{2}}\\ {} & \hspace{1em}+\cdots +{\big({y_{N-k+1}^{\ast }}-f\big({y_{N-k}^{\ast }}\big)\big)^{2}}+{\big({y_{N-k}^{\ast }}-f\big({y_{N-k-1}^{\ast }}\big)\big)^{2}}\\ {} & \hspace{1em}+\cdots +{\big({y_{2}^{\ast }}-f\big({y_{1}^{\ast }}\big)\big)^{2}}+{\big({y_{1}^{\ast }}\big)^{2}}\\ {} & \ge {\big(1-f\big({y_{N-1}^{\ast }}\big)\big)^{2}}+\cdots +{\big({y_{N-m+2}^{\ast }}-f\big({y_{N-m+1}^{\ast }}\big)\big)^{2}}\\ {} & \hspace{1em}+{\big({y_{N-m+1}^{\ast }}-f\big({y_{N-k}^{\ast }}\big)\big)^{2}}+0+\cdots +0\\ {} & \hspace{1em}+{\big({y_{N-k}^{\ast }}-f\big({y_{N-k-1}^{\ast }}\big)\big)^{2}}+\cdots +{\big({y_{1}^{\ast }}\big)^{2}},\end{aligned}\]
because ${y_{N-m+1}^{\ast }}-f({y_{N-m}^{\ast }})\ge {y_{N-m+1}^{\ast }}-f({y_{N-k}^{\ast }})$. Equality is attained if
(20)
\[ f\big({y_{N-m}^{\ast }}\big)=f\big({y_{N-k}^{\ast }}\big)\hspace{2.5pt}\text{and}\hspace{2.5pt}{y_{N-m}^{\ast }}=f\big({y_{N-m-1}^{\ast }}\big),\dots ,{y_{N-k+1}^{\ast }}=f\big({y_{N-k}^{\ast }}\big).\]
If this is true, f is constant on the interval $[{y_{N-m}^{\ast }},{y_{N-k}^{\ast }}]$, and $f({y_{N-m}^{\ast }})={y_{N-k+1}^{\ast }}$. Also,
\[ f\big({y_{N-m}^{\ast }}\big)=f\big(f\big({y_{N-m-1}^{\ast }}\big)\big)=\cdots =f\big(f\big(\dots \big(f\big({y_{N-k+1}^{\ast }}\big)\big)\big)\big)\lt {y_{N-k+1}^{\ast }},\]
because ${y_{N-k+1}^{\ast }}\gt 0$. Thus, (20) does not hold, which implies that equality is not attained in (19). Then, the sum ${\textstyle\sum _{n=1}^{N}}{({y_{n}}-f({y_{n-1}}))^{2}}$ is not minimized by the sequence ${\{{y_{n}^{\ast }}\}_{n=0}^{N}}$, which is a contradiction. Thus, ${y_{N-k+1}^{\ast }}\ge {y_{N-k}^{\ast }}$.  □
Remark 1.
If f and g are as in Lemma 3 and $|f(x)|\le |g(x)|$ on $[-1,1]$, then
(21)
\[ \underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{|{y_{N}}|\ge 1}{{y_{0}}=0}}{}}{\inf }\frac{1}{2}{\sum \limits_{n=1}^{N}}{\big({y_{n}}-g({y_{n-1}})\big)^{2}}\le \underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{|{y_{N}}|\ge 1}{{y_{0}}=0}}{}}{\inf }\frac{1}{2}{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}.\]
Proof.
By Lemma 3, the minimum on the right-hand side is attained for an increasing sequence ${\{{y_{n}}\}_{n=0}^{N}}$:
\[ \underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{{y_{N}}=1,{y_{0}}=0}{{y_{n}}\ge 0,n=1,\dots ,N-1}}{}}{\inf }\frac{1}{2}{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}=\underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{{y_{N}}=1,{y_{0}}=0}{{y_{0}}\le {y_{1}}\le \cdots \le {y_{N}}}}{}}{\inf }\frac{1}{2}{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}.\]
Now, for each $n=1,\dots ,N$,
\[ {y_{n}}-f({y_{n-1}})\ge {y_{n}}-{y_{n-1}}\ge 0,\]
and the same is true when f is replaced by g. Since $f(x)\le g(x)$ on $[0,1]$,
\[ {y_{n}}-f({y_{n-1}})\ge {y_{n}}-g({y_{n-1}}),\]
and it follows that
\[ {\big({y_{n}}-f({y_{n-1}})\big)^{2}}\ge {\big({y_{n}}-g({y_{n-1}})\big)^{2}}.\]
The statement of the lemma then follows.  □

4 Piecewise linear functions

Consider the process
(22)
\[ {X_{n+1}^{\varepsilon }}=f\big({X_{n}^{\varepsilon }}\big)+\varepsilon {\xi _{n+1}},\]
where ${X_{0}}=0$, ${\{{\xi _{n}}\}_{n\ge 1}}$ is a sequence of i.i.d. standard normal random variables, ε is a small positive parameter and f is a continuous piecewise linear function. We consider exit times from the interval $(-1,1)$, so $\Omega =(-1,1)$ and
(23)
\[ {\tau _{\varepsilon }}=\min \big\{n\ge 1:|{X_{n}^{\varepsilon }}|\ge 1\big\}.\]
As in Section 3, the definition of f outside of $[-1,1]$ does not have an impact on the results in this section.
Proposition 1.
Let f be a function on $\mathbb{R}$ that satisfies
\[ f(x)=\left\{\begin{array}{l@{\hskip10.0pt}l}a(x+b)\hspace{1em}& \mathit{if}\hspace{2.5pt}-1\le x\le -b,\\ {} 0\hspace{1em}& \mathit{if}\hspace{2.5pt}-b\lt x\lt b,\\ {} a(x-b)\hspace{1em}& \mathit{if}\hspace{2.5pt}b\le x\le 1,\end{array}\right.\]
where $0\le a\lt 1$ and $0\le b\lt 1$. Then, for any $M\ge 1$,
(24)
\[ \underset{\varepsilon \to 0}{\limsup }{\varepsilon ^{2}}\log {E_{{x_{0}}}}{\tau _{\varepsilon }}\le \frac{1}{2}\min \bigg(1,\underset{2\le N\le M}{\inf }\bigg(\big(1-{a^{2}}\big)\frac{{(1+\frac{a-{a^{N}}}{1-a}b)^{2}}}{1-{a^{2N}}}\bigg)\bigg).\]
If $a=1$ and $0\le b\lt 1$,
(25)
\[ \underset{\varepsilon \to 0}{\limsup }{\varepsilon ^{2}}\log {E_{{x_{0}}}}{\tau _{\varepsilon }}\le \frac{1}{2}\underset{1\le N\le M}{\inf }\bigg(\frac{{(1+(N-1)b)^{2}}}{N}\bigg)\]
for any $M\ge 1$.
vmsta277_g001.jpg
Fig. 1.
The upper bound on the right-hand side in (24) illustrated for different values of a and b and $N=1$ (when the value is 1/2) and $N=2,\dots 8$. The dotted line shows the value $(1-{a^{2}})/2$ in each case
Proof.
We have
\[\begin{aligned}{}\underset{\genfrac{}{}{0.0pt}{}{|{y_{N}}|\ge 1}{{y_{0}}=0}}{\inf }{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}& =\underset{\genfrac{}{}{0.0pt}{}{|{y_{N}}|=1}{|{y_{0}}|=b}}{\inf }{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}\\ {} & =\underset{\genfrac{}{}{0.0pt}{}{{y_{N}}=1}{{y_{0}}=b}}{\inf }{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}},\end{aligned}\]
because $f(0)=f(b)$ and it is enough to take the infimum over positive values because of Lemma 2. If ${y_{N}}=1$, ${y_{0}}=b$ and ${c_{n}}={a^{N-n}}$ for $n=1,\dots ,N$, we have the following telescoping sum:
\[ {\sum \limits_{n=1}^{N}}{c_{n}}\big({y_{n}}-f({y_{n-1}})\big)=\left\{\begin{array}{l@{\hskip10.0pt}l}1\hspace{1em}& \text{if}\hspace{2.5pt}N=1,\\ {} 1+ab\frac{1-{a^{N-1}}}{1-a}\hspace{1em}& \text{if}\hspace{2.5pt}N\ge 2.\end{array}\right.\]
By the Cauchy–Schwarz inequality,
\[ {\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}\ge \frac{{\textstyle\textstyle\sum _{n=1}^{N}}{c_{n}}({y_{n}}-f({y_{n-1}}))}{({\textstyle\textstyle\sum _{n=1}^{N}}{c_{n}^{2}})},\]
where equality can be attained. Since ${\textstyle\sum _{n=1}^{N}}{c_{n}^{2}}=(1-{a^{2N}})/(1-{a^{2}})$, it follows that
\[ \underset{\varepsilon \to 0}{\limsup }{\varepsilon ^{2}}\log {E_{{x_{0}}}}{\tau _{\varepsilon }}\le \frac{1}{2}\min \bigg(1,\underset{2\le N\le M}{\inf }\bigg(\big(1-{a^{2}}\big)\frac{{(1+\frac{a-{a^{N}}}{1-a}b)^{2}}}{1-{a^{2N}}}\bigg)\bigg),\]
for any $M\ge 1$. The value of the infimum, as well as for which N it is attained, depends on the choices of a and b; in some cases the minimum is 1 and in some cases it it less than one.
If we let $a=1$ and $0\lt b\lt 1$, we can use the same method as above with the telescoping sum. Then ${c_{n}}={a^{N-n}}=1$ for $n=1,\dots ,N$,
\[ {\sum \limits_{n=1}^{N}}{c_{n}}\big({y_{n}}-f({y_{n-1}})\big)=1+(N-1)b\]
and the result is
(26)
\[ \underset{\varepsilon \to 0}{\limsup }{\varepsilon ^{2}}\log {E_{{x_{0}}}}{\tau _{\varepsilon }}\le \frac{1}{2}\underset{1\le N\le M}{\inf }\frac{{(1+(N-1)b)^{2}}}{N}.\]
Here, the infimum is 1 (which is attained for $N=1$) if $b\ge 1/3$. For $b\lt 1/3$, the optimal N is either $\big\lfloor \frac{1}{b}\big\rfloor -1$ or $\big\lfloor \frac{1}{b}\big\rfloor $.  □
Note that if $b=0$ in expression (24), the infimum is attained for $N=M$. The inequality holds for any $M\ge 1$ and also as $M\to \infty $. Then the right-hand side in (24) becomes $(1-{a^{2}})/2$, the result for the autoregressive process (11).
Proposition 2.
Let $a\in (0,1)$ and $c\in (0,1]$ and let f be a function that satisfies
\[ f(x)=\left\{\begin{array}{l@{\hskip10.0pt}l}-ac\hspace{1em}& \mathit{if}\hspace{2.5pt}-1\le x\lt -c,\\ {} ax\hspace{1em}& \mathit{if}\hspace{2.5pt}-c\le x\le c,\\ {} ac\hspace{1em}& \mathit{if}\hspace{2.5pt}c\lt x\le 1.\end{array}\right.\]
Then
(27)
\[ \underset{\varepsilon \to 0}{\limsup }{\varepsilon ^{2}}\log E{\tau _{\varepsilon }}\le \left\{\begin{array}{l@{\hskip10.0pt}l}\frac{1}{2}({(1-ac)^{2}}+(1-{a^{2}}){c^{2}}),\hspace{1em}& \textit{if}\hspace{2.5pt}c\le a,\\ {} \frac{1}{2}(1-{a^{2}}),\hspace{1em}& \textit{if}\hspace{2.5pt}c\ge a.\end{array}\right.\]
vmsta277_g002.jpg
Fig. 2.
The upper bound on the right-hand side in (27) drawn as a function of c for some chosen values of a. The dotted lines show the value $(1-{a^{2}})/2$ for each a
Proof.
By Lemma 2 and Lemma 3 we only need to minimize over positive and increasing sequences. We determine the infimum of
\[ {\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}\]
for ${y_{0}}=0$, ${y_{0}}\le {y_{1}}\le \cdots \le {y_{N-1}}\le {y_{N}}$ and ${y_{N}}=1$. The set of sequences which we minimize over can be split into two parts: Either ${y_{N-1}}\lt c$ or the d last elements in the sequence are larger than c: ${y_{N-d}}\ge c$. If ${y_{N-d}}\ge c$,
\[ {\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}\ge {(1-ac)^{2}}+(d-1){(c-ac)^{2}}+{\sum \limits_{n=1}^{N-d}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}},\]
and this lower bound is attained for ${y_{N-d+1}}=\cdots ={y_{N-1}}=c$. Minimizing the sum on the right-hand side above when $0\le {y_{1}}\le \cdots \le {y_{N-d-1}}\le c$ and ${y_{N-d}}\ge c$ is the same minimizing problem as in (13), since $f({y_{n-1}})=a{y_{n-1}}$ when ${y_{n-1}}\le c$. We get
\[ \underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{{y_{N}}=1,{y_{0}}=0,{y_{N-d}}\ge c}{{y_{0}}\le {y_{1}}\le \cdots \le {y_{N}}}}{}}{\inf }{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}={(1-ac)^{2}}+(d-1){(c-ac)^{2}}+\frac{(1-{a^{2}}){c^{2}}}{1-{a^{2(N-d)}}}.\]
This value is smallest if $d=1$, and it is then
(28)
\[ {(1-ac)^{2}}+\frac{(1-{a^{2}}){c^{2}}}{1-{a^{2(N-1)}}}.\]
On the other hand, if ${y_{N-1}}\lt c$,
\[\begin{aligned}{}& \underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{{y_{N}}=1,{y_{0}}=0,{y_{N-1}}\lt c}{{y_{0}}\le {y_{1}}\le \cdots \le {y_{N}}}}{}}{\inf }{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}\\ {} & \hspace{1em}=\underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{{y_{N}}=1,{y_{0}}=0,{y_{N-1}}\lt c}{{y_{0}}\le {y_{1}}\le \cdots \le {y_{N}}}}{}}{\inf }{\sum \limits_{n=1}^{N}}{({y_{n}}-a{y_{n-1}})^{2}}=\frac{1-{a^{2}}}{1-{a^{2N}}}\end{aligned}\]
if $a\lt c$, because the minimizing problem then coincides with the same problem for the autoregressive process. If $a\ge c$,
\[\begin{aligned}{}& \underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{{y_{N}}=1,{y_{0}}=0,{y_{N-1}}\lt c}{{y_{0}}\le {y_{1}}\le \cdots \le {y_{N}}}}{}}{\inf }{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}\\ {} & \hspace{1em}=\underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{{y_{N}}=1,{y_{0}}=0,{y_{N-1}}\lt c}{{y_{0}}\le {y_{1}}\le \cdots \le {y_{N}}}}{}}{\inf }\Bigg({(1-a{y_{n-1}})^{2}}+{\sum \limits_{n=1}^{N-1}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}\Bigg)\end{aligned}\]
where it is optimal to have ${y_{N-1}}$ close to c, and we get the same infimum as in (28). To summarize,
\[\begin{aligned}{}& \underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{{y_{N}}=1,{y_{0}}=0}{{y_{0}}\le {y_{1}}\le \cdots \le {y_{N}}}}{}}{\inf }{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}\\ {} & \hspace{1em}=\min \bigg\{{(1-ac)^{2}}+\frac{(1-{a^{2}}){c^{2}}}{1-{a^{2(N-1)}}},\frac{1-{a^{2}}}{1-{a^{2N}}}\bigg\}.\end{aligned}\]
Since
\[ \underset{\varepsilon \to 0}{\limsup }{\varepsilon ^{2}}\log {E_{{x_{0}}}}{\tau _{\varepsilon }}\le \underset{1\le N\le M}{\inf }\underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{{y_{N}}=1,{y_{0}}=0}{{y_{0}}\le {y_{1}}\le \cdots \le {y_{N}}}}{}}{\inf }{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}\]
for any positive integer M, we can let M be arbitrarily large. We get
\[ \underset{\varepsilon \to 0}{\limsup }{\varepsilon ^{2}}\log {E_{{x_{0}}}}{\tau _{\varepsilon }}\le \left\{\begin{array}{l@{\hskip10.0pt}l}\frac{1}{2}({(1-ac)^{2}}+(1-{a^{2}}){c^{2}}),\hspace{1em}& \text{if}\hspace{2.5pt}c\le a,\\ {} \frac{1}{2}(1-{a^{2}}),\hspace{1em}& \text{if}\hspace{2.5pt}c\ge a.\end{array}\right.\]
 □
Proposition 3.
Let $0\le a\lt 1$ and let f be a function that satisfies
\[ f(x)=\left\{\begin{array}{l@{\hskip10.0pt}l}0\hspace{1em}& \mathit{if}\hspace{2.5pt}-1\le x\lt 0,\\ {} -ax\hspace{1em}& \mathit{if}\hspace{2.5pt}0\le x\le 1.\end{array}\right.\]
Then
(29)
\[ \underset{\varepsilon \to 0}{\limsup }{\varepsilon ^{2}}\log {E_{{x_{0}}}}{\tau _{\varepsilon }}\le \frac{1}{2}\frac{1}{1+{a^{2}}}.\]
Proof.
We study the sum
\[ {S_{N}}:={\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}.\]
If ${y_{N}}=1$, then ${S_{N}}\ge 1$, since $f(x)\le 0$ for all $x\in [-1,1]$. Also, if ${y_{N}}=-1$ and ${y_{N-1}}\lt 0$, we have ${S_{N}}\ge 1$. In the case when ${y_{N}}=-1$ and ${y_{N-1}}\ge 0$, it is optimal to have ${y_{N-2}}=\cdots ={y_{1}}={y_{0}}=0$. The smallest value that the sum can take is the minimum of ${(-1+ax)^{2}}+{x^{2}}$ for $x\in [0,1]$, which is $1/(1+{a^{2}})$. Thus,
\[ \underset{1\le N\le M}{\inf }\underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{|{y_{N}}|=1,{y_{0}}=0}{{y_{1}},\dots ,{y_{N-1}}\in (-1,1)}}{}}{\inf }{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}=\frac{1}{1+{a^{2}}},\]
and it follows that
\[ \underset{\varepsilon \to 0}{\limsup }{\varepsilon ^{2}}\log {E_{{x_{0}}}}{\tau _{\varepsilon }}\le \frac{1}{2}\frac{1}{1+{a^{2}}}.\]
 □
Note that if $a=0$ in Proposition 3, we have $f(x)=0$ on $[-1,1]$, which is a special case of the autoregressive process with $a=0$. The upper bound is then just $1/2$ which agrees with the result in the autoregressive case.
Proposition 4.
Let f be a function that satisfies
\[ f(x)=\left\{\begin{array}{l@{\hskip10.0pt}l}-bx\hspace{1em}& \mathit{if}\hspace{2.5pt}-1\le x\lt 0,\\ {} -ax\hspace{1em}& \mathit{if}\hspace{2.5pt}0\le x\le 1,\end{array}\right.\]
where $0\lt a\lt 1$, $0\lt b\lt 1$. Then
(30)
\[ \underset{\varepsilon \to 0}{\limsup }{\varepsilon ^{2}}\log {E_{{x_{0}}}}{\tau _{\varepsilon }}\le \frac{1}{2}\min \bigg(\frac{1-{(ab)^{2}}}{1+{a^{2}}},\frac{1-{(ab)^{2}}}{1+{b^{2}}}\bigg).\]
Proof.
Let
\[ S={\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}.\]
Consider the cases ${y_{N}}=1$ and ${y_{N}}=-1$ separately. First, let ${y_{N}}=-1$. Clearly, the sum S is smallest if the sequence ${\{{y_{n}}\}_{n=1}^{N}}$ has alternating signs: ${y_{N-1}}\gt 0,{y_{N-2}}\lt 0,\dots $ . Putting the derivative of S with respect to ${y_{n}}$ equal to zero gives
\[ {y_{n}}-f({y_{n-1}})={f^{\prime }}({y_{n}})\big({y_{n+1}}-f({y_{n}})\big),\]
where ${f^{\prime }}({y_{n}})=-b$ if ${y_{n}}\lt 0$ and ${f^{\prime }}({y_{n}})=-a$ if ${y_{n}}\gt 0$. Let ${s_{n}}:={y_{n}}-f({y_{n-1}})$. Then, if $N=2M$ so that N is even, it is optimal to have ${y_{1}}\gt 0$. Then
\[\begin{aligned}{}{s_{2k}}& =-{a^{-k}}{b^{-k+1}}{s_{1}},\\ {} {s_{2k+1}}& ={a^{-k}}{b^{-k}}{s_{1}},\end{aligned}\]
and it follows that
\[\begin{aligned}{}{y_{N}}& =-{s_{1}}\big[{a^{M}}{b^{M-1}}+{a^{M-2}}{b^{M-1}}+{a^{M-2}}{b^{M-3}}+{a^{M-4}}{b^{M-3}}+\cdots \\ {} & \hspace{1em}+{a^{-M+2}}{b^{-M+1}}+{a^{-M}}{b^{-M+1}}\big].\end{aligned}\]
This is a sum of two geometric sums. Since ${y_{N}}=-1$,
\[ {s_{1}}=\frac{1}{(1+{a^{2}}){(ab)^{N}}{a^{-2}}{b^{-1}}}\cdot \frac{{(ab)^{-2}}-1}{{(ab)^{-2N}}-1}.\]
The value of the sum is
\[ S={\sum \limits_{n=1}^{N}}{s_{n}^{2}}=\frac{1-{(ab)^{2}}}{1+{a^{2}}}\cdot \frac{1}{1-{(ab)^{2N}}}.\]
If N is odd, it is optimal to have ${y_{1}}\lt 0$. Then the same calculations as above follow, if a is replaced by b and vice versa. It follows that the value of the sum is
\[ S={\sum \limits_{n=1}^{N}}{s_{n}^{2}}=\frac{1-{(ab)^{2}}}{1+{b^{2}}}\cdot \frac{1}{1-{(ab)^{2N}}}.\]
The cases when ${y_{N}}=1$ and N is odd or even, give the same values of the sum. The smallest values are attained when N is large, and the statement of the proposition follows.  □
We note that when $a=b$ in Proposition 4, we have the autoregressive process with $f(x)=-ax$ on $[-1,1]$. The upper bound in Proposition 4 reduces to $1/(2(1-{a^{2}}))$ which is the bound for the autoregressive process.
Remark 2.
The TAR(1)-model, where $f(x)=a|x|$ for $|a|\lt 1$, is another piecewise linear example. For this model,
(31)
\[ \underset{\varepsilon \to 0}{\limsup }{\varepsilon ^{2}}\log E{\tau _{\varepsilon }}\le \frac{1-{a^{2}}}{2}.\]
Proof.
If $0\le a\lt 1$,
(32)
\[ \underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{|{y_{N}}|\ge 1}{{y_{0}}=0}}{}}{\inf }\frac{1}{2}{\sum \limits_{n=1}^{N}}{({y_{n}}-a|{y_{n-1}}|)^{2}}=\underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{{y_{N}}=1,{y_{0}}=0}{{y_{n}}\ge 0,n=1,\dots ,N-1}}{}}{\inf }\frac{1}{2}{\sum \limits_{n=1}^{N}}{({y_{n}}-a{y_{n-1}})^{2}},\]
and we have the same infimum as in the autoregressive case (which was treated in [12] and [13]). The case $-1\lt a\le 0$ is treated in a similar way.  □
The upper bound in (31) is sharp; this was shown in [18], where Novikov’s martingale method was used for this process to get the corresponding lower bound (the proof was almost the same as for the autoregressive process).

5 Other nonlinear functions

For a more general function f it is not always possible to minimize the sum and get explicit upper bounds. Numerical calculations may be needed. It is also possible to construct (nonstrict) upper bounds by simply evaluating the sum for some sequence of values instead of actually minimizing it. This is illustrated in the following case. Consider the case when f is quadratic, so we have the process
(33)
\[ {X_{n+1}^{\varepsilon }}=f\big({X_{n}^{\varepsilon }}\big)+\varepsilon {\xi _{n+1}},\]
where ${X_{0}}=0$, ${\{{\xi _{n}}\}_{n\ge 1}}$ is a sequence of i.i.d. standard normal random variables, ε is a small positive parameter and $f(x)=a{x^{2}}$.
Proposition 5.
When $f(x)=a{x^{2}}$ and $0\le a\le 0.5$,
\[ \underset{\varepsilon \to 0}{\limsup }{\varepsilon ^{2}}\log E{\tau _{\varepsilon }}\le \frac{1}{2}.\]
If $a\ge 0.5$,
(34)
\[ \underset{\varepsilon \to 0}{\limsup }{\varepsilon ^{2}}\log E{\tau _{\varepsilon }}\le \frac{1}{2}\bigg(\frac{1}{a}-\frac{1}{4{a^{2}}}\bigg).\]
Proof.
Since $f(x)\ge 0$ and f is even, it is optimal to have ${y_{N}}=1$ and ${y_{n}}\ge 0$ $\forall n$. We have
\[\begin{aligned}{}{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n}})\big)^{2}}& ={\sum \limits_{n=1}^{N}}{\big({y_{n}}-a{y_{n-1}^{2}}\big)^{2}}\\ {} & =1+{y_{n-1}^{2}}(1-2a)+{y_{n-2}^{2}}(1-2a{y_{n-1}})+\cdots +{y_{1}^{2}}(1-2a{y_{2}})\\ {} & \hspace{1em}+{a^{2}}\big({y_{n-1}^{4}}+{y_{n-2}^{4}}+\cdots +{y_{2}^{2}}\big)\ge 1\end{aligned}\]
when $0\le a\le 0.5$, and equality is achieved by putting ${y_{1}}={y_{2}}=\cdots ={y_{N-1}}=0$.
For $a\ge 0.5$,
\[ \underset{1\le N\le M}{\inf }\left(\underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{|{y_{N}}|\ge 1,{y_{0}}=0}{|{y_{n}}|\lt 1,n=1,2,\dots ,n-1}}{}}{\inf }\frac{1}{2}{\sum \limits_{n=1}^{N}}{\big({y_{n}}-a{y_{n-1}^{2}}\big)^{2}}\right)\le \underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{{y_{2}}=1,{y_{0}}=0}{|{y_{1}}|\lt 1,}}{}}{\inf }\frac{1}{2}{\sum \limits_{n=1}^{2}}{\big({y_{n}}-a{y_{n-1}^{2}}\big)^{2}},\]
where the infimum on right-hand side is $\frac{1}{2}(\frac{1}{a}-\frac{1}{4{a^{2}}})$ (it is attained for ${y_{1}}=\sqrt{\frac{1}{a}-\frac{1}{2{a^{2}}}}$). This gives the upper bound in (34).  □
This is not necessarily the best upper bound for all $a\ge 0.5$, since we have only calculated the infimum for $N=2$, but at least we get an upper bound. Numerical calculations suggest that the bound in (34) is the best one for a slightly larger than 0.5, and that the optimal choice of N increases as a increases.
For another example of a nonlinear case, consider the classical deterministic Ricker model defined by
\[ {x_{n+1}}={x_{n}}{e^{r-\gamma {x_{n}}}},\hspace{1em}n=0,1,2,\dots ,\]
where ${x_{t}}$ represents the size or density of a population at time t, $r\gt 0$ models the growth rate and $\gamma \gt 0$ is an environmental factor [9]. By suitably renorming the population, we may take $\gamma =1$. The model then has a fixed point at $x=r$ (and one at $x=0$). If we introduce stochasticity in the model by adding normally distributed white noise, and move the fixed point $x=r$ to the origin, we have the process
\[ {X_{n+1}^{\varepsilon }}=f\big({X_{n}^{\varepsilon }}\big)+\varepsilon {\xi _{n+1}},\]
where $f(x)=(x+r){e^{-x}}-r$ and ${\{{\xi _{n}}\}_{n\ge 1}}$ is a sequence of i.i.d. $N(0,1)$ variables. We can examine the time until the process exits from a suitable neighbourhood of the origin.
vmsta277_g003.jpg
Fig. 3.
The function $f(x)=x{e^{r-x}}$ for $r=1.5$. We have a fixed point at $x=r$. On the right, we see a part of the plot with the fixed point moved to the origin
If $r=1.5$, consider for example the time until exit from the interval $[-0.5,0.5]$. Numerical calculations of the infimum
\[ \underset{1\le N\le M}{\inf }\left(\underset{\genfrac{}{}{0.0pt}{}{\genfrac{}{}{0.0pt}{}{|{y_{N}}|\ge 0.5}{{y_{0}}=0}}{|{y_{n}}|\lt 0.5,n=1,\dots ,N}}{\inf }\frac{1}{2}{\sum \limits_{n=1}^{N}}{\big({y_{n}}-f({y_{n-1}})\big)^{2}}\right)\]
give the approximate value 0.09, so that
\[ \underset{\varepsilon \to 0}{\limsup }{\varepsilon ^{2}}\log E{\tau _{\varepsilon }}\lessapprox 0.09.\]
The derivative of f at the origin is $1-r$, so a suitable linear approximation of the function f is $l(x)=-0.5x$. By replacing f by l, we have an autoregressive process for which the upper bound of the exit time is
\[ \underset{\varepsilon \to 0}{\limsup }{\varepsilon ^{2}}\log E{\tau _{\varepsilon }}\le 0.{5^{2}}\frac{1-0.{5^{2}}}{2}=0.09375.\]
We note that a linear approximation might give good enough approximations of the upper bounds, in cases when the neighbourhood around the origin is chosen to be rather small.

6 Connection to stationary distribution

Mark Kac proved in 1947 that the mean return time of a discrete Markov chain to a point x is the reciprocal of the invariant probability $\pi (x)$. In [7], we explored this idea by comparing the exit time for the process defined by the stochastic difference equation
(35)
\[ {X_{n+1}^{\varepsilon }}=f\big({X_{n}^{\varepsilon }}\big)+\varepsilon {\xi _{n+1}}\]
and the return time to a certain set for the same process. We get an upper bound for the asymptotics of the exit time, and that bound is the reciprocal of the stationary distribution of the process, evaluated at the point where the level curve of the stationary distribution touches the boundary of the set (or interval in the univariate case) from which the process exits.
In the univariate case with $f(x)=ax$, $x\in \mathbb{R}$, (that is, for the autoregressive process), this method gives the same upper bound as the LDP method:
\[ \limsup {\varepsilon ^{2}}\log E{\tau _{\varepsilon }}\le \frac{1-{a^{2}}}{2}.\]
These methods only give an upper bound, but we know that the bound is sharp in the autoregressive case; the corresponding lower bound can be shown by other methods – see [18], where Novikov’s martingale method ([16] and [17]) was used, and [11].
In Section 4, we saw several examples of processes of the type defined in (35) where f was a piecewise linear function. For these examples, it is not straightforward to derive expressions for the stationary distributions of the processes. We also note that in Section 4, where the LDP method was used, the definition of f outside of the interval $[-1,1]$ could be ignored. However, when calculating stationary distributions, the definition of f outside of $[-1,1]$ matters a great deal.
We observe that there is another piecewise linear case for which the stationary distribution is known, and this is a TAR(1) process with a threshold of 0, studied by Anděl et al. [1], where
\[ {X_{n+1}^{\varepsilon }}=f\big({X_{n}^{\varepsilon }}\big)+\varepsilon {\xi _{n+1}},\]
with $f(x)=-|ax|$ for $|a|\lt 1$. In [1], they gave the following formula for the stationary distribution of this process:
\[ \frac{1}{\varepsilon }{\bigg(\frac{2(1-{a^{2}})}{\pi }\bigg)^{1/2}}\exp \bigg(-\frac{1}{2}\big(1-{a^{2}}\big)\frac{{x^{2}}}{{\varepsilon ^{2}}}\bigg)\Phi \bigg(\frac{-ax}{\varepsilon }\bigg),\]
where Φ is the cumulative distribution function of a $N(0,1)$ random variable. We note that
\[ -{\varepsilon ^{2}}\log \bigg(\frac{1}{\varepsilon }{\bigg(\frac{2(1-{a^{2}})}{\pi }\bigg)^{1/2}}\exp \bigg(-\frac{1}{2}\big(1-{a^{2}}\big)\frac{{x^{2}}}{{\varepsilon ^{2}}}\bigg)\Phi \bigg(\frac{-ax}{\varepsilon }\bigg)\bigg)\to \frac{1-{a^{2}}}{2}\]
at the point $x=-1$. This means that in this particular case, the value of the limit of the stationary distribution, evaluated at the point where the process exits the interval, coincides with the upper bound achieved by use of the LDP method in Remark 2.
A preprint of a previous version of this paper has been posted on ArXiv [10].

Acknowledgement

The authors would like to thank the two anonymous referees for their constructive comments which have led to improvement of the paper.

References

[1] 
Anděl, J., Netuka, I., Zvára, K.: On threshold autoregressive processes. Kybernetika 20(2), 89–106 (1984). MR0747062
[2] 
Basak, G.K., Ho, K.-W.R.: Level-crossing probabilities and first-passage times for linear processes. Adv. Appl. Probab. 36, 643–666 (2004). MR2058153. https://doi.org/10.1239/aap/1086957589
[3] 
Baumgarten, C.: Survival probabilities of autoregressive processes. ESAIM Probab. Stat. 18, 145–170 (2014). MR3143737. https://doi.org/10.1051/ps/2013031
[4] 
Di Nardo, E.: On the first passage time for autoregressive processes. Sci. Math. Jpn. 21, 61–76 (2008). MR2399804
[5] 
Hamza, K., Jagers, P., Klebaner, F.C.: On the establishment, persistence and inevitable extinction of populations. J. Math. Biol. 72, 797–820 (2016). MR3459167. https://doi.org/10.1007/s00285-015-0903-2
[6] 
Högnäs, G.: On the lifetime of a size-dependent branching process. Stoch. Models 35, 119–131 (2019). MR3969510. https://doi.org/10.1080/15326349.2019.1578241
[7] 
Högnäs, G., Jung, B.: Analysis of a stochastic difference equation: Exit times and invariant distributions. Fasc. Math. 44, 69–74 (2010). MR2722632
[8] 
Högnäs, G., Jung, B.: Exit times for some autoregressive processes with non-Gaussian noise distributions. Contemp. Math. 668, 111–117 (2016). MR3536695. https://doi.org/10.1090/conm/668/13399
[9] 
Högnäs, G.: On the quasi-stationary distribution of a stochastic Ricker model. Stoch. Process. Appl. 70, 243–263 (1997). MR1475665. https://doi.org/10.1016/S0304-4149(97)00064-1
[10] 
Högnäs, G., Jung, B.: Exit times for some nonlinear autoregressive processes (2019). arXiv:1912.08514.
[11] 
Jung, B.: Exit times for multivariate autoregressive processes. Stoch. Process. Appl. 123, 3052–3063 (2013). MR3062436. https://doi.org/10.1016/j.spa.2013.03.003
[12] 
Klebaner, F., Liptser, R.: Large deviations for past-dependent recursions. Probl. Inf. Transm. 32, 23–34 (1996). MR1441520
[13] 
Klebaner, F., Liptser, R.: Large deviations for past-dependent recursions (2006). arXiv:math/0603407.
[14] 
Klebaner, F.C., Liptser, R.: Moderate deviations for randomly perturbed dynamical systems. Stoch. Process. Appl. 80(2), 157–176 (1999). MR1682255. https://doi.org/10.1016/S0304-4149(98)00075-1
[15] 
Koski, T., Jung, B., Högnäs, G.: Exit times for ARMA processes. Adv. Appl. Probab. 50(A), 191–195 (2018). MR3905100. https://doi.org/10.1017/apr.2018.79
[16] 
Novikov, A.A.: On the first passage time of an autoregressive process over a level and an application to a “disorder” problem. Theory Probab. Appl. 35(2), 269–279 (1991). https://doi.org/10.1137/1135035
[17] 
Novikov, A., Kordzakhia, N.: Martingales and first passage times of AR(1) sequences. Stochastics 80(2-3), 197–210 (2008). MR2402164. https://doi.org/10.1080/17442500701840885
[18] 
Ruths, B.: Exit times for past-dependent systems. Surv. Appl. Ind. Math. (Obozr. Prikl. Prom. Mat.) 15, 25–30 (2008)
[19] 
Varadhan, S.R.S.: Large Deviations and Applications. Society for Industrial and Applied Mathematics (1984). https://epubs.siam.org/doi/abs/10.1137/1.9781611970241. MR0758258. https://doi.org/10.1137/1.9781611970241.bm
Reading mode PDF XML

Table of contents
  • 1 Introduction
  • 2 Large deviation tools and bounds for exit times
  • 3 Minimizing the sum
  • 4 Piecewise linear functions
  • 5 Other nonlinear functions
  • 6 Connection to stationary distribution
  • Acknowledgement
  • References

Copyright
© 2025 The Author(s). Published by VTeX
by logo by logo
Open access article under the CC BY license.

Keywords
Exit time first passage time autoregressive process large deviation principle

MSC2020
60G17 60F10

Metrics
since March 2018
144

Article info
views

31

Full article
views

34

PDF
downloads

16

XML
downloads

Export citation

Copy and paste formatted citation
Placeholder

Download citation in file


Share


RSS

  • Figures
    3
vmsta277_g001.jpg
Fig. 1.
The upper bound on the right-hand side in (24) illustrated for different values of a and b and $N=1$ (when the value is 1/2) and $N=2,\dots 8$. The dotted line shows the value $(1-{a^{2}})/2$ in each case
vmsta277_g002.jpg
Fig. 2.
The upper bound on the right-hand side in (27) drawn as a function of c for some chosen values of a. The dotted lines show the value $(1-{a^{2}})/2$ for each a
vmsta277_g003.jpg
Fig. 3.
The function $f(x)=x{e^{r-x}}$ for $r=1.5$. We have a fixed point at $x=r$. On the right, we see a part of the plot with the fixed point moved to the origin
vmsta277_g001.jpg
Fig. 1.
The upper bound on the right-hand side in (24) illustrated for different values of a and b and $N=1$ (when the value is 1/2) and $N=2,\dots 8$. The dotted line shows the value $(1-{a^{2}})/2$ in each case
vmsta277_g002.jpg
Fig. 2.
The upper bound on the right-hand side in (27) drawn as a function of c for some chosen values of a. The dotted lines show the value $(1-{a^{2}})/2$ for each a
vmsta277_g003.jpg
Fig. 3.
The function $f(x)=x{e^{r-x}}$ for $r=1.5$. We have a fixed point at $x=r$. On the right, we see a part of the plot with the fixed point moved to the origin

MSTA

MSTA

  • Online ISSN: 2351-6054
  • Print ISSN: 2351-6046
  • Copyright © 2018 VTeX

About

  • About journal
  • Indexed in
  • Editors-in-Chief

For contributors

  • Submit
  • OA Policy
  • Become a Peer-reviewer

Contact us

  • ejournals-vmsta@vtex.lt
  • Mokslininkų 2A
  • LT-08412 Vilnius
  • Lithuania
Powered by PubliMill  •  Privacy policy