\documentclass[JEP,XML,SOM,Unicode,published]{cedram}
\datereceived{2022-06-16}
\dateaccepted{2023-09-06}
\dateepreuves{2023-10-05}

\usepackage[scr=rsfs,cal=euler]{mathalfa}
\multlinegap0pt
\let\cal\mathcal
\let\sep:
\let\epsilon\varepsilon

\makeatletter
\def\@settitle{%
\vspace*{-8mm}
\raggedleft\includegraphics[scale=.5]{titre-jep}
\vtop to 50 mm{%
 \parindent=0pt
 {\abstractfont\article@logo\par}
  \hrule
 \vfil
 \begin{center}
 \def\baselinestretch{1.2}\large\vfil
   {\didottitraille\MakeUppercase\@title\par}
 \vfil\vfil
 \begin{minipage}{.8\textwidth}\centering
   \ifx\@empty\smfbyname\else
   {\smf@byfont\smfbyname\ifsmf@byauthor\enspace\else\ \fi}%
   \fi {\smf@authorfont \edef\smfandname{{\noexpand\smf@andfont
         \smfandname}} \andify\authors\authors\par}
 \end{minipage}
 \vfil \vrule height .4pt width .3\textwidth \vfil
 \end{center}}%
 \par\enlargethispage{.5\baselineskip}%
}
\def\@setthanks{\def\thanks##1{\par##1\@addpunct{{\upshape.}}}\vspace*{-5pt}\thankses}
\makeatother

\newenvironment{enumeratei}
{\bgroup\def\theenumi{\roman{enumi}}\def\theenumii{\arabic{enumii}}\begin{enumerate}}
{\end{enumerate}\egroup}

\newcommand{\Psfrac}[2]{(\sfrac{#1}{#2})}
\newcommand{\psfrac}[2]{\sfrac{(#1)}{#2}}
\newcommand{\spfrac}[2]{\sfrac{#1}{(#2)}}
\newcommand{\pspfrac}[2]{\sfrac{(#1)}{(#2)}}

\let\f\frac

\theoremstyle{plain}
\newtheorem{theorem}{Theorem}[section]
\newtheorem{lemma}[theorem]{Lemma}
\newtheorem{corollary}[theorem]{Corollary}
\newtheorem{prop}[theorem]{Proposition}

\theoremstyle{remark}
\newtheorem{ex}[theorem]{Example}
\newtheorem{nota}[theorem]{Notation}
\newtheorem{defi}[theorem]{Definition}
\newtheorem{rem}[theorem]{Remark}
\newtheorem{hyp}[theorem]{Assumption}
\newtheorem*{remark}{Remark}
\newtheorem*{notations}{Notations}

\newcommand{\Co}{\mathcal{C}}

\newcommand\T{\mathbb{T}}
\newcommand\Z{\mathbb{Z}}
\newcommand\N{\mathbb{N}}
\newcommand\R{\mathbb{R}}
\newcommand\E{\mathbb{E}}
\renewcommand\P{\mathbb{P}}


\datepublished{2023-10-17}
\begin{document}
\frontmatter
\title[Individual-based evolutionary models and Hamilton-Jacobi equations]{Filling the gap between\\ individual-based evolutionary models and\\ Hamilton-Jacobi equations}

\author[\initial{N.} \lastname{Champagnat}]{\firstname{Nicolas} \lastname{Champagnat}}
\address{Université de Lorraine, CNRS, Inria, IECL\\
F-54000 Nancy, France}
\email{nicolas.champagnat@inria.fr}
\urladdr{http://nchampagnat.perso.math.cnrs.fr/}

\author[\initial{S.} \lastname{Méléard}]{\firstname{Sylvie} \lastname{Méléard}}
\address{École Polytechnique \& Institut Universitaire de France, CNRS, Institut polytechnique de Paris\\
route de Saclay, 91128 Palaiseau Cedex, France}
\email{sylvie.meleard@polytechnique.edu}
\urladdr{https://sites.google.com/view/sylvie-meleard/accueil}

\author[\initial{S.} \lastname{Mirrahimi}]{\firstname{Sepideh}\nobreakauthor \lastname{Mirrahimi}}
\address{Institut Montpelliérain Alexander Grothendieck, Univ. Montpellier, CNRS\\
Place Eugène Bataillon, 34090 Montpellier, France}
\email{sepideh.mirrahimi@umontpellier.fr}
\urladdr{https://imag.umontpellier.fr/~mirrahimi/}

\author[\initial{V. T.} \lastname{Tran}]{\firstname{Viet Chi} \lastname{Tran}}
\address{LAMA, Univ Gustave Eiffel, Univ Paris Est Creteil, CNRS\\
F-77454 Marne-la-Vallée, France}
\email{chi.tran@univ-eiffel.fr}
\urladdr{https://perso.math.u-pem.fr/tran.viet-chi/}

\thanks{This work is funded by the European Union (ERC, SINGER, 101054787). Views and opinions expressed are however those of the author(s) only and do not necessarily reflect those of the European Union or the European Research Council. Neither the European Union nor the granting authority can be held responsible for them. This work has also been supported by the Chair ``Modélisation Mathématique et Biodiversité'' of Veolia Environnement-École Polytechnique-Museum National d’Histoire Naturelle-Fondation X. S.Mirrahimi was partially supported by the ANR project DEEV ANR-20-CE40-0011-01. V.C.T. also acknowledges support from Labex Bézout (ANR-10-LABX-58)}

\begin{abstract}
We consider a stochastic model for the evolution of a discrete population structured by a trait with values on a finite grid of the torus, and with mutation and selection. We focus on a parameter scaling where population is large, individual mutations are small but not rare, and the grid mesh is much smaller than the size of mutation steps.  When considering the evolution of the population in long time scales, the contribution of small sub-populations may strongly influence the dynamics. Our main result quantifies the asymptotic dynamics of subpopulation sizes on a logarithmic scale. We establish that under our rescaling, the stochastic discrete process converges to the viscosity solution of a Hamilton-Jacobi equation. The proof makes use of almost sure maximum principles and careful control of the martingale parts.
\end{abstract}

\keywords{Stochastic birth death models, large population approximation, selection, mutation, viscosity solution, maximum principle, Hamilton-Jacobi equation}

\subjclass{92D25, 92D15, 60J80, 60F99, 35F21}

\altkeywords{Processus de naissances et morts, processus aléatoire, approximation grande population, sélection, mutation, solution de viscosité, principe du maximum, équation de Hamilton-Jacobi}

\alttitle{Des modèles stochastiques d'évolution aux équations de Hamilton-Jacobi}

\begin{altabstract}
Nous considérons un modèle stochastique pour l'évolution d'une population discrète structurée en trait à valeurs dans une grille finie du tore, avec mutation et sélection. On se place dans une limite d'échelle de grande population, de petites mutations (mais pas rares), et où le maillage tend vers zéro. En temps long, la contribution de petites sous-populations peut fortement influencer la dynamique. Nous montrons que dans ce cadre, le processus stochastique discret converge sur une échelle logarithmique vers la solution de viscosité d'une équation de Hamilton-Jacobi. La preuve fait appel à un principe du maximum presque-sûr et à des estimées fines des parties martingales.
\end{altabstract}

\maketitle
\vspace*{-\baselineskip}
\tableofcontents
\mainmatter

\section{Introduction and presentation of the model}
\label{sec:introduction}

Long-term evolutionary dynamics of biological populations may be strongly influenced by small populations and local extinction in some areas of the phenotypical trait space. Survival of small populations in very large populations is crucial when evolution proceeds by selective sweeps \cite{keelingpalmer2008} or for the evolution of antibiotic resistance for bacteria \cite{ochmanlawrencegroisman,stewartlevin}. For example, in bacterial populations involving horizontal transfer, it was shown in~\cite{billiardcolletferrieremeleardtran,calvez20} that the individual-based long-term dynamics is very sensitive to random survival of very small populations, which may either drive the population to evolutionary suicide or to cyclic dynamics. In such context, the bacterial population is very large making the tracking of small populations very challenging.

From a point of view of mathematical modeling, one wishes to consider large population scalings allowing for survival of much smaller populations. Two approaches emerged: a purely deterministic one, based on partial differential equations (PDE), and a stochastic one, based on birth and death processes (so-called individual-based models in biology). Both approaches describe exponentially small populations sizes and characterize the dynamics of the exponents.

An analytical approach allowing to deal with negligible but non-extinct populations was proposed in~\cite{diekmannjabinmischlerperthame} {and then widely developed (see for instance~\cite{BP.GB:08,Barles2009,LMP}) for the asymptotic study of parabolic integro-differential selection-mutation models}. Let us present it in a setting close to~\cite{Barles2009}. We consider a population whose individuals are differentiated by a trait $x\in \mathbb{T}$, the torus of dimension 1, identified below with the interval $[0,1)$. The trait can vary from an individual to the other.
The evolution of the population is driven by two effects: mutation of the traits, and selection as the reproductive and survival abilities of an individual depend on its trait $x$.
For an individual of trait $x\in \mathbb{T}$, let us denote by $b(x)$ (resp. $d(x)$ and $p(x)$) the clonal birth rate (resp. the death rate and the birth rate with mutation), and by $G(h)$ the mutation kernel. Assuming that the population density solves the PDE
\begin{equation}
\label{eq:PDE}
\begin{cases} \varepsilon\partial_t u_\varepsilon(t,x)=u_\varepsilon(t,x)\left(b(x)-d(x)\right) & \\
\hspace{2.5cm}+\dpl\int_{\mathbb{T}}\frac{1}{\varepsilon}G\left(\psfrac{x-y}{\varepsilon}\right)
p(y)u_\varepsilon(t,y)\mathrm dy, & (t,x)\in\R_+\times\T,\\[8pt]
u_\varepsilon(t,0)=\exp(\sfrac{\beta_0(x)}{\varepsilon}), & x\in\T
\end{cases}
\end{equation}
in the limit $\varepsilon\to 0$ of small mutations and large time, and applying the Hopf-Cole transformation
\begin{equation}
\label{eq:Hopf-Cole}
\beta_\varepsilon(t,x)=\varepsilon\log u_\varepsilon(t,x),\quad\text{or}\quad u_\varepsilon(t,x)=\exp(\sfrac{\beta_\varepsilon(t,x)}{\varepsilon}),
\end{equation}
it is proved in~\cite{Barles2009} (in a slightly different setting{, considering $x\in \R$ and taking into account a competition term}) that $\beta_\varepsilon$ converges to the unique viscosity solution $\beta$ of the Hamilton-Jacobi equation
\begin{equation}
\label{eq:HJ_intro}
\begin{cases}
\dpl\frac{\partial}{\partial t} \beta(t,x)=b(x)-d(x)+ p(x) \int_{\R} G(h)e^{h\partial_x \beta(t,x)} dh, & (t,x)\in\R_+\times\T,\\
\beta(0,x)=\beta_0(x), & x\in\T.
\end{cases}
\end{equation}

Scaling limits of individual-based models on a discrete trait space with rare mutations and large population via a scaling parameter $K$ that will tend to infinity, and allowing to deal with negligible populations and local extinction, were proposed in~\cite{durrettmayberry,Bovier2019,champagnatmeleardtran2021,coquille2021,blathpaultobias}. These references focus on population sizes of the order of $K^\beta$, and characterize the asymptotic dynamics of the exponent $\beta$. In particular, local extinction is possible when the exponent $\beta$ hits 0. The fact that the trait space is discrete allows to describe separately the dynamics of each small sub-population.
However, this makes the detailed description of the
asymptotic dynamics very complicated (see~\cite{champagnatmeleardtran2021,coquille2021}).
Note that mutations are assumed individually rare in these references, but they are more frequent than in the scaling limits of adaptive dynamics (see e.g.~\cite{metzgeritzmeszenajacobsheerwaarden,dieckmannlaw,champagnat06,champagnatmeleard2011}), where negligible populations either fixate or go extinct fast, due to the fact that the populational mutation rate vanishes here.
Scaling limits with non-vanishing populational mutation rates partly solve the criticisms raised by biologists~\cite{WaxmanGavrilets2005} concerning the too slow evolutionary speed in adaptive dynamics, particularly in microorganism populations. Biological criticisms were also raised for the analytical approach~\cite{perthame2010}, because
of the so-called {\it tail problem}: exponentially small populations, which may actu\-ally be extinct, can have a strong influence on the future evolutionary dynamics of the population. In particular, evolutionary branching is too fast.
Modifications of the Hamilton-Jacobi equation were proposed in~{\cite{perthame2010,Mirrahimi2012,jabin2012}} to solve this problem, but we believe that an individual-based approach is crucial to provide a more realistic and biologically relevant solution to the tail problem.

The purpose of our work is to provide a stochastic individual-based justification of Hamilton-Jacobi equations. To our knowledge, this is the first proof of this kind in the literature. Note however that there are many examples of spatial branching processes with space-, time- or type-dependent branching rate, for which the exponential growth can be expressed using variational formulas over paths (see e.g. \cite{biggins,berestyckibrunetharrisharrisroberts,callegaroroberts,mallein,maillardraoultourniaire}). These can be seen as Hopf-Lax variational formulas of certain Hamilton-Jacobi equations.
Note also that the Hopf-Cole transformation~\eqref{eq:Hopf-Cole} is reminiscent of large deviations scalings. Our scaling is more of a law of large numbers type. As far as we know, the large deviations interpretation of the Hamilton-Jacobi equation can be done through a Feynman-Kac interpretation of the PDE~\eqref{eq:PDE}~\cite{Champagnat-Henry-2019}. However, the stochastic process involved in the Feynman-Kac formula does not seem to be directly related to the biological population process, even though some works suggest that it may be related to the ancestral trait process of living individuals~\cite{foriengarnierpatout}. None of the references above use a direct approach from individual-based models to Hamilton-Jacobi equations.

We follow an individual-based approach, assuming a continuous trait space with a vanishing discretization step $\delta_K$, where $K$ is a scaling parameter such that the population is of the order of $K^{\widetilde{\beta}^K(t,x)}$, assuming frequent and small mutations. In the individual-based model, individuals with trait $x\in\mathbb{T}$ give birth to a clone at rate $b(x)$, die at rate $d(x)$ or give birth to a mutant at rate $p(x)$. Mutant traits are drawn according to a discretization of the distribution $\log (K) G(\log (K) \cdot)$. Mutation steps are of the order of $1/\log (K)$ and the discretization step $\delta_K$ is assumed much smaller than $1/\log (K)$. In this first work, we focus on the understanding of the relevant scales allowing to capture the limiting Hamilton-Jacobi dynamics. Thus, we consider a simplified model where the birth rate $b$ is assumed larger than the death rate $d$, making the stochastic process super-critical, and the trait space has no boundary. Generalization is a work in progress.

The proof of our main result makes use of uniform Lipschitz bounds on the finite variation part of $\widetilde{\beta}^K$, obtained using an almost sure maximum principle and careful bounds for the martingale part. The identification of the limit is done by checking that it is almost surely a viscosity solution of~\eqref{eq:HJ}. We describe the model and state our main result in Section~\ref{sec:model-main-result}. The proof is divided into two main steps---proof of tightness and identification of the limit---which are detailed in Sections~\ref{sec:main-steps} and~\ref{sec:identif}, respectively.

\subsubsection*{Acknowledgements}
The authors thank the anonymous reviewers and also Anouar Jeddi for their comments.

\section{Model and main result}
\label{sec:model-main-result}
\subsection{The model}

We consider a super-critical stochastic birth-death-mutation model describing an asexual population of individuals characterized by a quantitative phenotypic or genetic trait $x\in \T$. Starting from a finite population whose initial size is parameterized by $K\in \N$, our goal is to recover, in the limit $K\to +\infty$, an evolutionary dynamics described by the Hamilton-Jacobi partial differential equation~\eqref{eq:HJ_intro}. For this, we consider a discretization of the trait space $\T$ with step $\delta_K\to 0$. For the sake of simplicity, we will consider in what follows that $1/\delta_K\in \N$. Then, the population is composed of individuals with traits belonging to the discrete space
\[
{\cal X}_K:=\Bigl\{i\delta_K:i\in \{0, 1,\dots, \frac{1}{\delta_K}-1\}\Bigr\},
\]
embedded with the torus distance: $\forall x, y \in [0,1)$,
\begin{align*}
\rho(x,y) & =\min\left\{|x'-y'|,\ x'=x\bmod1,\ y'=y\bmod1\right\} \\ & =\min\big(|x-y|,1-|x-y|\big).\end{align*}
It is enough to define $\rho(x,y)$ for $x,y \in \T$ by considering their representative in $[0,1)$.

The number of individuals with trait $i \delta_K$ is described by the stochastic process $(N_i^K(t), t\ge 0)$. The total population size at time $t$ is then given by
\[
N^K(t)=\sum_{i=0}^{1/\delta_K-1}N_i^K(t).
\]
An individual with trait $x\in {\cal X}_K$
\begin{itemize}
\item gives birth to a new individual with the same trait $x$ at rate $b(x)$;

\item dies at rate $d(x)$;

\item gives birth to a mutant individual with trait $y\in{\cal X}_K$ at rate
\begin{equation}\label{mutation-rate}
p(x)\delta_K \log(K)\, G\bigl((\overline{y-x})\,\log(K)\bigr),
\end{equation}
where
$\overline{z}$ is the unique real number of $[-1/2,1/2)$ that is equal to $z$ modulo 1, \ie $\overline{z}=z-\lfloor z+1/2\rfloor$, where $\lfloor\cdot\rfloor$ is the integer part function.
\end{itemize}

{Jumps of $N^K$ are thus of $+1$ or $-1$. In the rest of the paper, the following assumptions are made:}

\skpt
\begin{hyp}\label{hypmodele}
\begin{enumerate}
\item\label{hypmodele1}
We assume that $b$, $d$ and $p$ are nonnegative Lipschitz continuous functions defined on $\T$, such that for all $x\in\T$,
\begin{equation*}%\label{hyp:supercritical}
b(x)>d(x) \hbox{ and } p(x)>0.
\end{equation*}
This means that the birth-death process for each trait is super-critical. In the sequel, we denote by $\bar{b}$, $\bar{p}$ and $\bar{d}$ the upper bounds of these functions on $\T$, by $\,\underline p>0$ the lower bound of $p$, and by $\|b\|_{\text{Lip}}$, $\|d\|_{\text{Lip}}$ and $\|p\|_{\text{Lip}}$ their Lipschitz norm.

\item\label{hypmodele2}
The function $G$ defined on $\mathbb{R}$ is a nonnegative continuous density function (satisfying $\int_\R G(y)\,dy=1$) and has finite exponential moments of any order. Moreover, we assume that there exists $R>0$ such that $G$ is nonincreasing on $[R,+\infty)$ and nondecreasing on $(-\infty,-R]$.\\
An example of function $G$ satisfying Assumption 2 is given by the Gaussian kernel $G(h)=\frac{1}{\sqrt{2\pi}\sigma}e^{-h^2/2\sigma^2}$.

\item\label{hypmodele3}
There exists a constant $a_1>0$ such that, for all $K\in \N$ and all $i\in\{0, 1,\dots, \sfrac{1}{\delta_K}-1\}$,
\begin{equation}\label{hyp:condinit}
N^K_i(0)\geq K^{a_1}.
\end{equation}

\item\label{hypmodele4}
There exists $a_2<a_1$ such that
\begin{equation}\label{hyp:deltaK}
K^{-a_2/4}\ll \delta_K\ll \frac{1}{\log(K)}\quad \mbox{ as }\quad K\to +\infty.
\end{equation}
\end{enumerate}
\end{hyp}

Point \eqref{hypmodele4} above implies that
\begin{equation}\label{def:hk}
h_K:=\delta_K \log(K)\ll 1.
\end{equation}
Thus, the interpretation of \eqref{mutation-rate} is that the rate at which $x$ gives birth to a mutant individual is close to $ p(x)$. Indeed, for an individual with trait $x_K=i_K \delta_K$ with $i_K=\lfloor x/\delta_K\rfloor$ and $x\in\T$ fixed,
\begin{multline*}%\label{eq:conv_integrale}
\lim_{K\to +\infty}p(x_K)\sum_{j=0}^{\sfrac{1}{\delta_K}-1} h_K G\left(\overline{(i_K-j)\delta_K}\, \log(K)\right) \\
\begin{aligned} &= \lim_{K\to +\infty}p(x_K)\sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor} h_K G\left(h_K\ell\right) =p(x)\int_{\mathbb{R}}G(y)\,dy=p(x),
\end{aligned}
\end{multline*}
where we used Assumption~\ref{hypmodele}.2 to control the tails of the Riemann sum.

Therefore, the individual mutation rate is order 1
and mutation steps are small: conditionally on being a mutant, the trait $y$ of the offspring has the distribution $G$ scaled by a factor $1/\log(K)$ representing the order of magnitude of the mutation steps.
Note also that \eqref{def:hk} means that the mesh size is much smaller than the mutation step.

Our goal is to study the asymptotic behavior (when $K$ tends to infinity) of the population sizes $N_i^K$ when $N_i^K(0)$ is of the order of $K^{\alpha_i}$ for some $\alpha_i>0$. Note that in the case where $p(x)=0\,$ for all $x\in\T$, the process $N^K_i(t)$ is a super-critical one-dimensional branching process, hence $\E(N^K_i(t))=\E(N^K_i(0)) e^{(b(i \delta_K)-d(i\delta_K))t}$. Therefore, if the initial condition is of order $K^{\alpha_i}$, then
\[
\E[ N^K_i(t\log(K))]\sim K^{\alpha_i + (b(i \delta_K)-d(i\delta_K))t}.
\]
This suggests to study
\begin{equation*}%
\label{beta}
\beta_i^K(t)=\f{\log(N_i^K(t \log(K)))}{\log(K)},
\end{equation*}
with the convention that $\beta^K_i(t)=0\,$ if $\,N^K_i(t\log(K))=0$. So the sub-population of trait $i\delta_K$ at time $t\log(K)$ has size $N^K_i(t \log(K))=K^{\beta^K_i(t)}$.

We make the following Lipschitz assumption on the initial condition $\beta^K_i(0)$.

\begin{hyp}\label{hypconvergence}
\noindent Assume that there exists a constant $A>0$ such that
\begin{equation*}%\label{hyp:condlip}
\lim_{K\to+\infty}\mathbb{P}\Big(\sup_{i\neq j}\frac{|\beta^K_i(0)-\beta^K_j(0)|}{\rho(i\, \delta_K,j\, \delta_K)}> A\Big)=0.
\end{equation*}
\end{hyp}

\skpt
\begin{notations}
\begin{enumeratei}
\item
We shall use the following Riemann approximation repeatedly in the proofs: for all $L>0$,
\begin{multline*}%\label{eq:momentexpo:G}
\lim_{K\to +\infty}\sum_{j=0}^{\sfrac{1}{\delta_K}-1} h_K G\left(\overline{(i_K-j)\delta_K}\,\log(K)\right) e^{ L\,\log(K)\, \rho(j \delta_K,i_K \delta_K)}\\
=\lim_{K\to +\infty} \sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor} h_K G\left(h_K\ell\right)e^{ L\,{h_K} \, | {\ell} |}
=\int_{\mathbb{R}}e^{L |y|}G(y)\,dy,
\end{multline*}
and thus, there exists a constant $\overline{G}(L)$ depending only on $\alpha>0$ such that 
\begin{equation}\label{eq:conv_moment-expo}
\sup_{K\geq 1}\,\sum_{j=0}^{\sfrac{1}{\delta_K}-1} h_K G\left(\overline{(i_K-j)\delta_K}\,\log(K)\right)e^{ L\,\log(K)\, \rho(j \delta_K,i_K \delta_K)} =: \overline G(L)<+\infty.
\end{equation}

\item
In what follows and for any function $f$ on $ \{0, 1,\dots, \sfrac{1}{\delta_K}-1\}$, we will use the notation
\[\Delta_Kf_i= \frac{f_{i+1}-f_i}{\delta_K},\]
with the convention that $f_{1/\delta_K}=f_0$.
\end{enumeratei}
\end{notations}

\subsection{The main result - Sketch of the proof}

\par Since we are interested in the convergence of the quantities $\beta_i^K$ to a continuous function defined on the trait space $\T$, when $K\to+\infty$ and $\delta_K \to 0$, we introduce the following affine interpolation of the $\beta_i^K$'s:
for all $x\in\T$ and $K\geq 1$, let $i\in \{0, 1,\dots, \sfrac{1}{\delta_K}-1\}$ be such that $x\in [i\delta_K,(i+1)\delta_K)$, and define
\begin{equation}
\label{lafonction}
\widetilde \beta^K_{t}(x) =\beta_i^K(t)\Big(1-\f {x}{\delta_K}+i\Big)+\beta_{i+1}^K(t)\Big(\f {x}{\delta_K}-i\Big),
\end{equation}
with the convention that $\beta^K_{1/\delta_K}(t)=\beta^K_0(t)$.
For $T>0$, the sequence of processes $(\widetilde \beta^K_{t}, t\in[0,T])_{K}$ belongs to $\mathbb{D}([0,T], \Co(\T,\mathbb{R}))$, where $\Co(\T,\mathbb{R})$ is endowed with the topology of uniform convergence and $\mathbb{D}([0,T], \Co(\T,\mathbb{R}))$ is the Skorokhod space of càdlàg paths with the associated Skorokhod topology.

Let us state our main theorem.

\begin{theorem}\label{thm:main}
Let $T>0$. Under the Assumptions \ref{hypmodele} and \ref{hypconvergence}, and assuming that $\widetilde{\beta}^K_0(\cdot)$ converges in probability for the topology of uniform convergence on $\Co(\T,\mathbb{R})$ to a deterministic function $\beta_0(\cdot)\in \Co(\T,\mathbb{R})$,
the sequence $(\widetilde{\beta}^K)_K$ converges in probability in $\mathbb{D}([0,T], \Co(\T,\mathbb{R}))$ to the unique Lipschitz viscosity solution of the Hamilton-Jacobi equation
\begin{equation}
\label{eq:HJ}
\begin{cases}
\dpl\frac{\partial}{\partial t} \beta(t,x)=b(x)-d(x)+ p(x) \int_{\R} G(h)e^{h\partial_x \beta(t,x)} dh, & (t,x)\in (0,T]\times \T, \\
\beta(0,x)=\beta_0(x), & x\in \T.
\end{cases}
\end{equation}
\end{theorem}

The proof of Theorem \ref{thm:main} will be classically obtained in two steps: tightness and identification of the limiting values. Therefore we will prove the two next results, respectively in Sections~\ref{sec:main-steps} and~\ref{sec:identif}.

\begin{theorem}
\label{thm:tight}
The distributions of the processes $(\widetilde \beta^K_{t}, t\in[0,T])_{K}$ form a C-tight sequence in ${\cal P}(\mathbb{D}([0,T], \Co(\T,\mathbb{R})))$, the set of probability measures on $\mathbb{D}([0,T], \Co(\T,\mathbb{R}))$. In addition, for all $T>0$, there exists a constant $L$ such that, for any $\beta$ distributed as a limiting value of the laws of $(\widetilde \beta^K_{t}, t\in[0,T])_{K}$, we have almost surely
\begin{equation}
\label{eq:Lipschitz-estimate}
\sup_{t\in[0,T]}\sup_{x,y\in\T\ s.t.\ x\neq y}\frac{|\beta(t,x)-\beta(t,y)|}{\rho(x,y)}\leq L.
\end{equation}
\end{theorem}

\begin{theorem}
\label{thm:identification}
The limiting values $\beta$ of $(\tilde \beta^K_{t}, t\in[0,T])_{K}$ are characterized as {the} unique Lipschitz viscosity solution of the Hamilton-Jacobi equation \eqref{eq:HJ}.
\end{theorem}

The proofs of these two results require to control the increments of the functions $\widetilde \beta^K_{t}(.)$.
These functions
can also be written as \[\widetilde
\beta^K_{t}(x)
=(x-i\delta_K)\Delta_K\beta^K_i(t)+\beta^K_i(t).\]
From this expression, we observe that two technical steps are required: to estimate uniformly $\beta_{i}^K(t)$ and to control uniformly $\ \Delta_K\beta^K_i(t)=\psfrac{\beta_{i+1}^K(t)-\beta_{i}^K(t)}{\delta_K}\ $ (the second estimate being the harder part, it is a major difficulty and constitutes the technical interest of the paper). Such estimates are also obtained in the deterministic derivation of Hamilton-Jacobi equations of type \eqref{eq:HJ} from parabolic integro-differential equations using the maximum principle and the Bernstein method which consists in applying again the maximum principle to the equation satisfied by the increments (see \cite{Barles2009}). Here, since we have stochastic processes we cannot apply the Bernstein method directly. Using the Doob-Meyer decomposition, the stochastic processes can be separated into a finite variation part and a martingale part. We show indeed that the martingale part remains small with our rescaling and we apply the maximum principle almost surely on the finite variation part.

Let us detail now the semi-martingale Doob-Meyer decomposition of the processes~$\beta_{i}^K$, $i\in \{0,\dots, 1/\delta_{K}-1\}$. Using standard arguments~\cite{bansayemeleard},
\begin{multline}
N^K_i(t)-N^K_i(0)-\int_0^t(b(i\delta_K)-d(i\delta_K))N^K_i(s) ds \\ -\sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor}\int_0^t h_Kp((i+\ell)\delta_K)G(h_{K}\ell)N^K_{i+\ell}(s)ds \label{eq:DM-N^K_i}
\end{multline}
is a square integrable martingale with quadratic variation
\[
\int_0^t(b(i\delta_K)+d(i\delta_K))N^K_i(s) ds
+\sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor}\hspace*{-2mm}\int_0^t\hspace*{-1mm} h_Kp((i+\ell)\delta_K)G(h_{K}\ell)N^K_{i+\ell}(s)ds.
\]
It then follows from Itô's formula for jump processes that
\begin{equation}\label{decomposition}
\beta_{i}^K(t)=M^K_{i}(t) + A^K_{i}(t)
\end{equation}
with
\begin{align} A_{i}^K(t)= \beta_{i}^K(0) + & \frac{1}{\log(K)}\int_{0}^{t\log(K)}\Bigl(b(i \delta_{K})N^K_{i}(s) \log\bigl(1+\sfrac{1}{N^K_{i}(s)}\bigr) \label{def:A}\\
& \hspace{2cm}+ d(i \delta_{K})N^K_{i}(s) \log\bigl(1-\sfrac{1}{N^K_{i}(s)}\bigr)\Bigr) ds \\
+ & \frac{1}{\log(K)} \sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor}
h_{K}\, p((i+\ell) \delta_{K}) G(h_{K}\ell) \\
& \hspace{2cm} \times \int_{0}^{t\log(K)}N^K_{i+\ell}(s)\log\left(1+\sfrac{1}{N^K_{i}(s)}\right) ds ,
\end{align}
with the conventions that, when the index $j\notin \{0,\dots, 1/\delta_K-1\}$,
\begin{equation}\label{eq:convention}
\begin{cases}
\begin{aligned}
N^K_{j}&=N^K_{j-\lfloor j\delta_K\rfloor/
\delta_K},\\
p(j\delta_K)&=p((j-\lfloor j\delta_K\rfloor/
\delta_K)\delta_K),
\end{aligned}
&\text{when }j\geq 1/\delta_K, \\[10pt]
\begin{aligned}
N^K_{j}&=N^K_{j+\lceil | j|\delta_K\rceil/
\delta_K},\\
p(j\delta_K)&=p((j+\lceil |j|\delta_K\rceil/
\delta_K)\delta_K),
\end{aligned}
&\text{when }j<0. 
\end{cases}
\end{equation}
The process $M^K_{i}$ is a local martingale with predictable quadratic variation
\begin{align}
\langle M^K_{i}\rangle_{t}& = \frac{1}{\log^2(K)}\int_{0}^{t\log(K)}\Big(b(i \delta_{K})N^K_{i}(s)\log^2\bigl(1+\sfrac{1}{N^K_{i}(s)}\bigr) \label{def:M}\\
& \qquad\qquad\qquad + d(i \delta_{K})N^K_{i}(s)\log^2\bigl(1-\sfrac{1}{N^K_{i}(s)}\bigr)\Big) ds \\
& +\frac{1}{\log^2(K)}\sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor}
h_{K} p((i+\ell) \delta_{K}) G(h_{K}\ell)\\
& \qquad \qquad \qquad \times \int_{0}^{t\log(K)}N^K_{i+\ell}(s)\log^2\left(1+\sfrac{1}{N^K_{i}(s)}\right) ds.
\end{align}

In Section \ref{sec:main-steps}, we will prove technical uniform estimates on the martingale part $\Delta_K M^K_{i}(t)$ and the finite variation part $\Delta_K A^K_{i}(t)$ of the processes $\Delta_K\beta^K_{i}(t)$. In that aim, let us introduce sequences of stopping times playing an important role in the proofs.

Let $a\in(a_2,a_1)$ be fixed during the rest of the proof, $a_1$ being defined in \eqref{hyp:condinit} and $a_2$ in \eqref{hyp:deltaK}. For any $K$, we define
\begin{equation}
\label{theta}
\tau'_{K} = \inf\Bigl\{t\ge 0\sep \exists i \in \{0, 1,\dots, \sfrac{1}{\delta_K}-1\}, N^K_i(t\log(K)) < K^a\Bigr\}.
\end{equation}
Note that $\beta^K_{i}(t)>0$ for all $t\le\tau'_K$.

For all $L>0$, we also define
\begin{equation} \label{def:tau}
\tau_K(L)=\inf\Bigl\{t\geq 0\sep\exists i\in \{0, 1,\dots, \sfrac{1}{\delta_K}-1\},\,|\beta^K_{i+1}(t)-\beta^K_i(t)|>L\delta_K \Bigr\},
\end{equation}with the usual convention that $\beta^K_{1/\delta_K}=\beta^K_0$.
It is easy to check that
\[
\tau_K(L)=\inf\Bigl\{t\geq 0:\exists i,j\in \{0, 1,\dots, \sfrac{1}{\delta_K}-1\},
|\beta^K_{i}(t)-\beta^K_j(t)|>L\rho(i\delta_K,j\delta_K) \Bigr\}
\]
and also
\begin{equation} \label{def:tau2}
\tau_K(L)=\inf\Bigl\{t\geq 0:\exists x,y\in \T,\,|\widetilde{\beta}^K_{t}(x)-\widetilde{\beta}^K_t(y)|>L\ \rho(x,y) \Bigr\}.
\end{equation}

We will study the processes until the stopping time
\begin{equation}
\theta_K(L)=\tau_K(L)\wedge \tau'_{K}.\label{def:thetaK(L)}
\end{equation}
Before the stopping time $\theta_K(L)$, the functions $\widetilde{\beta}^K_t$ are Lipschitz and the population size of each trait is bounded from below by $K^a$, by definition. For each $L$ fixed, we~will provide uniform estimates on the martingale parts $M^K_{i}(t)$ and $\Delta_K M^K_{i}(t)$ and the finite variation parts $A^K_{i}(t)$ and $\Delta_K A^K_{i}(t)$ of the processes $\beta^K_{i}(t)$ and $\Delta_K\beta^K_{i}(t)$, before the stopping time $\theta_K(L)$. This will allow us to prove that for all $T$, $\theta_K(L)$ is larger than $T$ for $L$ large enough (see Proposition \ref{prop:Lipschitz-estimate}).

\section{Proof of the Theorem~\ref{thm:tight}}
\label{sec:main-steps}

We will use the criterion of Theorem 3.1 in Jakubowski \cite{jakubowski}. To prove the tightness of the sequence $(\widetilde{\beta}^K)_{K\geq 1}$ in $\mathbb{D}([0,T],\Co(\T,\mathbb{R}))$, it is sufficient to prove:

\begin{enumeratei}
\item\label{enum:i}
For each $\varepsilon>0$, there exists a compact set $C_{\varepsilon}\subset \Co(\T,\mathbb{R})$ such that
\[\forall K,\ \mathbb{P}\left(\widetilde \beta^K\in \mathbb{D}([0,T], C_{\varepsilon})\right) >1-\varepsilon.\]
\item\label{enum:ii}
For each $f\in \Co(\T,\mathbb{R})$, the sequence of laws of real-valued processes
\begin{equation}\label{def-Xf}X^K_{f}(\cdot) = \int_\T \widetilde \beta^K(.,x) f(x) dx\end{equation}
is tight.
\end{enumeratei}

Point \eqref{enum:i} is the hard part of the proof. Using Ascoli's characterization of compact subsets of $\Co(\T,\mathbb{R})$, we need to obtain estimates related to equi-boundedness and to equi-continuity for the processes $\widetilde \beta^K_{t}(.)$.
The proof relies on Lipschitz estimates (in $x$) of the functions $\widetilde{\beta}^K_t$. In Section \ref{sec:control-martingale}, we show that the martingale part of $\beta^K_i(t\wedge\theta_K(L))$ remains small with our rescaling and in Section \ref{sec:control-fv}, we apply the maximum principle almost surely on the finite variation part of $\beta^K_i(t\wedge\theta_K(L))$. This allows us to prove that $\theta_K(L)$ is large enough in Section \ref{sec:control-incr}. The proof of the tightness is ended in Section \ref{sec:tight}.

\subsection{Control of the martingale part}\label{sec:control-martingale}

Our first estimate will be useful to prove the tightness of the laws of $\widetilde{\beta}^K$. In the sequel, $C$ denotes a constant not depending on any parameter and that may change from line to line.

The following estimate will be used repeatedly: by~\eqref{def:tau2}, for all $t\leq\tau_K(L)$ and all $i,j\leq 1/\delta_K-1$,
\begin{equation}
\label{eq:borne-N_i/N_j}
\frac{N^K_j(t\log(K))}{N^K_{i}(t\log(K))} = \exp(\log(K) (\beta_{j}(t)-\beta_{i}(t)) \le e^{ L\, \rho(j \delta_K,i \delta_K)\, \log(K) }.
\end{equation}

\begin{lemma}
\label{lem:M}
For all $T>0$, $L>0$, $K>0$ and $a\in (a_2,a_1)$, there exists a constant~$C$ independent of $K$, $T$, $L$ and $i$ such that, almost surely, for all $t\leq T$ and all $i\in\{0,\ldots,1/\delta_K-1\}$,
\begin{equation}
\label{eq:lem-M-1}
\langle M^K_i\rangle_{t\wedge\theta_K(L)}\leq \Bigl(\frac{C\overline{G}(L)}{K^a\log(K)}\Bigr)t.
\end{equation}
Further, for all $A>0$,
\begin{equation}
\label{maison}
\mathbb{P}\Big(\sup_{t\le T\wedge \theta_{K}(L)}\sup_{i}|M^K_i(t)|\geq A\Big)\leq\frac{C\overline{G}(L)T}{A^2\delta_K K^a\log(K)}.
\end{equation}
\end{lemma}

\begin{proof}
It follows from~\eqref{def:M} that
\begin{multline*}
\langle M^K_i\rangle_{t\wedge\theta_K(L)} \leq \frac{C(\bar b+\bar d)}{\log^2 (K)}\int_0^{(t\wedge \theta_K(L))\log(K)}\frac{ds}{N^K_i(s)}\\
+\frac{\bar p}{\log^2 (K)}\int_0^{(t\wedge\theta_K(L))\log(K)}\quad
\sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor}
h_{K} G(h_{K}\ell)\frac{N^K_{i+\ell}(s)}{(N^K_{i}(s))^2}\,ds.
\end{multline*}
Therefore, using~\eqref{theta} and~\eqref{eq:borne-N_i/N_j},~\eqref{eq:lem-M-1} follows. Then
\begin{align*}
\mathbb{P}\Big(\sup_{t\le T\wedge \theta_{K}(L)}\sup_{i}|M^K_i(t)|\geq A\Big) & \leq\sum_{i=0}^{1/\delta_K-1}\mathbb{P}\Big(\sup_{t\le T\wedge \theta_{K}(L)}|M^K_i(t)|\geq A\Big) \\ & \leq \frac{1}{A^2}\sum_{i=0}^{1/\delta_K-1}\mathbb{E}\Big(\sup_{t\le T\wedge \theta_{K}(L)}|M^K_i(t)|^2\Big),
\end{align*}
so we obtain~\eqref{maison} using Doob's inequality and~\eqref{eq:lem-M-1}.
\end{proof}

The second lemma gives controls on the martingale increments.

\begin{lemma}
\label{corol:accrM}
For all $T>0$, $K>0$ and $a\in (a_2,a_1)$, let us set $\varepsilon_{K}= \delta_K^{-1}(K^{a}\log(K))^{-1/4}$. For $L>0$, we define the event
\begin{equation}
\Omega_{K}(L)=\biggl\{\sup_{\substack{0\leq i\leq 1/\delta_K-1\\ t\le T\wedge \theta_K(L)}} |\Delta_KM^K_{i}(t)| \le \varepsilon_{K}\biggr\}.\label{event:OmegaK}
\end{equation}
Then there exists a constant $C>0$ such that
\begin{equation}
\label{eq:estim-Omega}
\mathbb{P}(\Omega_K^c(L))
\leq \frac{C\sqrt{\overline{G}(L)T}}{\delta_K (K^a
\log(K))^{1/4}} .
\end{equation}
In addition,
\begin{equation}
\label{eq:borne-martingale-simple}
\mathbb{P} \biggl(\sup_{\substack{0\leq i,j\leq 1/\delta_K-1\\ t\le T\wedge \theta_K(L)}}\left|M^K_i(t)-M^K_j(t)\right|>\varepsilon_K \biggr)\leq C\sqrt{\overline{G}(L)T}
\varepsilon_K.
\end{equation}
\end{lemma}

\begin{proof}
We first prove that, for all $\varepsilon>0$ and any $i\in\{0,\ldots,1/\delta_K-1\}$, for some constant $C$ independent of $t$, $K$, $\varepsilon$ and $L$,
\begin{equation}
\label{eq:mg-max-ineq}
\mathbb{P}\Big(\sup_{s\le t\wedge \theta_K(L) } |\Delta_KM^K_{i}(s)|>\varepsilon\Big)\leq \frac{C}{\varepsilon} \sqrt{ \frac{ \overline{G}(L) t}{\delta_K^2 K^{a}\log(K)} },
\end{equation}
where the constant $\overline{G}(L)$ is defined in~\eqref{eq:conv_moment-expo}.

By the submartingale maximal lemma, we have that \[ \mathbb{P}\Big(\sup_{s\le t} |\Delta_KM^K_{i}(s)|>\varepsilon\Big)\leq \frac1\varepsilon\,\mathbb{E}(|\Delta_KM^K_{i}(t)|) \leq \frac1\varepsilon\, \mathbb{E}\big(|\Delta_KM^K_{i}(t)|^2\big)^{1/2}.\]
Now, Lemma~\ref{lem:M} yields
\[
\langle\Delta_KM^K_{i} \rangle_{t\wedge \theta_K}
\leq \frac{2}{\delta_K^2}\left(\langle M^K_{i+1} \rangle_{t\wedge \theta_K} + \langle M^K_{i} \rangle_{t\wedge \theta_K}\right)
\leq \frac{ 4 C\overline{G}(L) t}{\delta_K^2 K^{a}\log(K)}.
\]
Hence~\eqref{eq:mg-max-ineq} is proved.

Now, we choose $\varepsilon=\varepsilon_K$ and obtain
\begin{align*}
\mathbb{P}(\Omega_K^c(L))\leq & \sum_{0\leq i\leq 1/\delta_K-1}\mathbb{P}\Bigl(\sup_{t\le T\wedge \theta_K(L)} |\Delta_KM^K_{i}(t)| > \varepsilon_{K}
\Bigr)\\
\leq & \frac{C\sqrt{\overline{G}(L)T}}{\delta_K\varepsilon_K \delta_K \sqrt{K^{a}\log(K)}}.
\end{align*}
This ends the proof of~\eqref{eq:estim-Omega}.

To complete the proof of
Lemma~\ref{corol:accrM}, it is sufficient to notice that, since $|j-i|\delta_K \le 1$ for all $0\leq i,j\leq 1/\delta_K-1$, we have that
\[
\Omega_K(L)\subset\biggl\{\sup_{\substack{0\leq i,j\leq 1/\delta_K-1\\ t\le T\wedge \theta_K(L)}}\left|M^K_i(t)-M^K_j(t)\right|\leq\varepsilon_K\biggr\}. \qedhere
\]
\end{proof}

Note that, by \eqref{hyp:deltaK}, $\delta_K^4 K^{a}\log(K)$ tends to infinity as $K$ goes to infinity, so~$ \mathbb{P}(\Omega_{K}(L))$ tends to $1$ and the probability of $\Omega_{K}(L)$ converges to 1.
From now on, we will work on the probability subspace $\Omega_{K}(L)$.

\subsection{Control of the finite variation part}\label{sec:control-fv}
Let us now focus on the finite variation part $A^K_i$. We will prove that

\begin{prop}
\label{est-varpart}
Let $T>0$. Then, there exists a constant $C_1$ such that for $K$ large enough, for all $t\leq T$ and all $i\in\{0,\ldots,1/\delta_K-1\}$ the following inequality holds almost surely on $\Omega_K(L)$:\vspace*{-3pt}\enlargethispage{\baselineskip}
\begin{equation*}
%\label{max-varpart}
|A^K_i(t\wedge \theta_K(L)) |\leq \max_{0\leq j\leq 1/\delta_K-1}\beta^K_j(0) + C_1 t.
\end{equation*}
\end{prop}

\begin{proof}
We first provide the proof of the upper bound on $A_i^K(t\wedge \theta_K(L))$. For simplicity we will omit the dependency in $L$ of $\theta_K(L)$ when there is no ambiguity.

Let $t$ and $s$ be less than $T$ such that $s<t$. Using that $ \log (1+x)\leq x$ and \eqref{eq:borne-N_i/N_j}, and neglecting the non positive death term, we have\vspace*{-5pt}
\begin{multline*}
A_{i}^K(t\wedge \theta_K) - A_{i}^K(s\wedge \theta_K)\\
\begin{aligned}
&= \frac{1}{\log(K)}\int_{(s\wedge \theta_K) \log(K)}^{(t\wedge \theta_K)\log(K)}\Big(b(i \delta_{K})N^K_{i}(u) \log\left(1+\sfrac{1}{N^K_{i}(u)}\right)\\[-3pt]
&
\hspace{4cm} + d(i \delta_{K})N^K_{i}(u) \log\left(1-\sfrac{1}{N^K_{i}(u)}\right)\Big) du \\[-3pt]
& +\frac{1}{\log(K)} \sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor} h_{K} p((i+\ell) \delta_{K}) G(h_{K}\ell)\\[-5pt]
& \hspace{4cm} \times \int_{(s\wedge \theta_K) \log(K)}^{(t\wedge \theta_K)\log(K)}N^K_{i+\ell}(u)\log\left(1+\sfrac{1}{N^K_{i}(u)}\right) du \\[-3pt]
&\leq C\bar b (t- s) \\
& \, + \frac{1}{\log(K)} \sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor} h_{K} p((i+\ell) \delta_{K}) G(h_{K}\ell)\int_{(s\wedge \theta_K) \log(K)}^{(t\wedge \theta_K)\log(K)}\frac{N^K_{i+\ell}(u)}{N^K_{i}(u)} du\\[-3pt]
&\leq C\bar b (t - s)
\\[-3pt]
& \, +\hspace*{-1mm} \sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor}\hspace*{-1mm} h_{K} p((i+\ell) \delta_{K}) G(h_{K}\ell)\int_{s\wedge \theta_K }^{t\wedge \theta_K}\hspace*{-1mm} \exp\bigl(\log(K) (\beta^K_{i+\ell}(u)-\beta^K_{i}(u))\bigr) du.
\end{aligned}
\end{multline*}
Recall that on $\Omega_K(L)$,\vspace*{-3pt}
\[\beta^K_{j}(u)-\beta^K_{i}(u) =A_{j}^K(u) - A_{i}^K(u)+ M_{j}^K(u) -M_{i}^K(u) \leq A_{j}^K(u) - A_{i}^K(u)+\varepsilon_K
.\]
Thus we obtain that, on the event $\Omega_K(L)$,\vspace*{-5pt}
\begin{multline*} A_{i}^K(t\wedge\theta_K) - A_{i}^K(s\wedge\theta_K)
\leq C\bar b (t - s)
+ \sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor} h_{K} p((i+\ell) \delta_{K}) G(h_{K}\ell) \\[-2pt] \int_{s\wedge \theta_K }^{t\wedge \theta_K} e^{\varepsilon_K \log(K) } \exp\bigl(\log(K) (A^K_{i+\ell}(u)-A^K_{i}(u))\bigr) du.
\end{multline*}
We deduce that, almost surely on $\Omega_K(L)$ and for all $t\leq\theta_K(L)$,\vspace*{-5pt}
\begin{multline*}
\frac{dA^K_{i}(t)}{dt} \leq C\bar b
+ \sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor} h_{K} p((i+\ell) \delta_{K}) G(h_{K}\ell) \\[-8pt] \times e^{\varepsilon_K\log(K) } \exp\bigl(\log(K) (A^K_{i+\ell}(t)-A^K_{i}(t))\bigr).
\end{multline*}
Defining $\mathbf{A}^K_i(t) = A^K_{i}(t) - C\bar bt - 2 \bar p t$, we deduce that for any $t\leq \theta_K(L)$,\vspace*{-5pt}
\begin{multline}
\label{eq:prop-est-varpart}
{d \mathbf{A}^K_{i}(t)\over dt}\leq \bar p e^{\varepsilon_K\log(K) } \sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor} h_{K} G(h_{K}\ell) \\[-5pt]
\times \exp(\log(K) (\mathbf{A}^K_{i+\ell}(t)- \mathbf{A}^K_{i}(t))) - 2 \bar p.
\end{multline}
Let us introduce
\[(i_{K}, t_{K}) = (i_K(\omega),t_K(\omega))=\text{argmax}_{{i\in \{0,\dots, \sfrac{1}{\delta_{K}}-1\}, t\in [0, \theta_{K}(\omega)\wedge T]}} \mathbf{A}^K_{i}(t).\]
We can prove that $t_{K}=0.$
Indeed, if conversely we assume that $t_{K}>0$, then the right term of \eqref{eq:prop-est-varpart}, for $K$ large enough, is negative for $i=i_{K}$ {and $t=t_K$} and then the
left term is negative, contradicting the fact that $\mathbf{A}^K_{i_K}(t)$ is maximal for $t=t_K$. Hence, we have proved that {for $K$ large enough}, almost surely on
the event $\Omega_K$, for all $t\leq \theta_K(L)\wedge T$ and for all $i \in \{0, . . . , 1/\delta_K - 1 \}$,
\begin{align*}
A^K_i(t)=\mathbf{A}^K_i(t)+C\bar bt + 2 \bar p t&\leq\max_{0\leq j\leq 1/\delta_K}\mathbf{A}^K _j(0)+C\bar b t + 2 \bar p t\\
&= \max_{0\leq j\leq 1/\delta_K} \beta^K _j(0)+C\bar b t + 2 \bar p t.
\end{align*}

For the lower bound, we observe that, for $t\leq \theta_K(L)\wedge T$, $N^K_i(t)\geq K^a$ and hence, for $K$ large enough,
\[
\log\left(1+\sfrac{1}{N^K_i(t)}\right)\geq 0 \quad\text{and}\quad \log\left(1-\sfrac{1}{N^K_i(t)}\right)\geq -\frac{2}{N^K_i(t)}.
\]
The proof then follows from the same argument as above.
\end{proof}


The last result has a consequence that will be useful to prove the tightness of $\widetilde{\beta}^K$ in Section~\ref{sec:tight}.

\begin{corollary}
\label{lem:upper-bound}
For all $T>0$, there exists $C(T)$ such that,
\[
\lim_{K\to+\infty}\mathbb{P}\Bigl(\sup_{0\leq i\leq 1/
\delta_K-1}\sup_{t\in[0,T\wedge\theta_K(L)]}\beta^K_i(t)\geq C(T)\Bigr)=0.
\]
\end{corollary}

\begin{proof}
We use the semi-martingale decomposition~\eqref{decomposition} of $\beta^K_i$, the result of Proposition~\ref{est-varpart},~\eqref{maison} with $A=1$ and Lemma~\ref{corol:accrM}, to deduce that, for all $t\leq T$ {and $K$ large enough} and $i \in \{0, 1,\dots, \sfrac{1}{\delta_K}-1\}$,
\[
|\beta^K_i(t\wedge\theta_K)|\leq \sup_{0\leq j\leq 1/\delta_K-1}|\beta^K_j(0)|+C_1T+1
\]
with probability at least
\[
1-\frac{C\overline{G}(L)T}{\delta_K K^a\log(K)} -\frac{C\sqrt{\overline{G}(L)T}}{\delta_K(K^a \log(K))^{1/4}}.
\]
Since $\widetilde{\beta}^K_i(0)$ converges in probability to $\beta_0$, $\mathbb{P}(\sup_i \beta^K_i(0)\geq \|\beta_0\|_\infty+1)$ converges to~$0$ when $K$ goes to $+\infty$. Hence the result follows with $C(T)=\|\beta_0\|_\infty+C_1T+2$.
\end{proof}

\Subsection{Estimates on $\theta_K(L)$}\label{sec:control-incr}

\begin{lemma}\label{lem:lower-bound} Under Assumption~\ref{hypmodele}\eqref{hypmodele1} and~\ref{hypmodele}\eqref{hypmodele3}, for all $T>0$,
\[
\lim_{K\to+\infty}\mathbb{P}(\tau'_K\geq T) =1.
\]
\end{lemma}

\begin{proof}[Proof of Lemma \ref{lem:lower-bound}]
By~\eqref{eq:DM-N^K_i}, neglecting incoming mutations, it is easy to prove using standard coupling arguments that for each $i$, the process $(N_i^K(t))_{t}$ is pathwisely bounded below by a branching process $(Z_i^K(t))_{t}$ with birth rate $b(i\delta_K)$, death rate $d(i\delta_K)$ and initial condition $K^{a+\varepsilon}$ for $\varepsilon=a_1-a>0$.
In addition, the processes $(Z_i^K(t))_{t}$ for $0\leq i\leq \sfrac{1}{\delta_K}-1$ are independent. Let us define
\[
\theta''_{K} = \inf\Big\{t\ge 0\sep \exists i \in \{0, 1,\dots, \sfrac{1}{\delta_K}-1\},\, Z^K_i(t\log(K)) < K^a\Big\} .
\]
In order to prove that $\ \lim_{K\to \infty} \mathbb{P}(\tau'_{K}>T) =1$, it is enough to prove that
\[\ \lim_{K\to \infty} \mathbb{P}(\theta''_{K}=+\infty) =1.\]
We have
\begin{align*}
\mathbb{P}(\theta''_{K}=+\infty) &= \mathbb{P}\Big(\forall i\in \{0,1,\dots, \sfrac{1}{\delta_K}-1\} ,\ \forall t\geq 0,\ Z^K_i(t \log(K)) >K^a\Big)\\
&= \prod_{i=0}^{\sfrac{1}{\delta_K}-1}\mathbb{P}\Big(\inf_{t\ge 0} Z^K_i(t\log(K)) >K^a\Big).
\end{align*}
Fix $i\in\{0,\ldots,\sfrac{1}{\delta_K}-1\}$. It is usual to prove (by time change) that the probability $\mathbb{P}\big(\inf_{t\ge 0} Z^K_i(t\log(K)) >K^a\big)$ is equal to the probability that a random walk
\[
M\Bigl({b(i\delta_K)\over b(i\delta_K)+d(i\delta_K)}, {d(i\delta_K)\over b(i\delta_K)+d(i\delta_K)}\Bigl)
\]
on $\Z_{+}$ (adding $+1$ with probability $\spfrac{b(i\delta_K)}{b(i\delta_K)+d(i\delta_K)}$ and $-1$ with probability $\spfrac{d(i\delta_K)}{b(i\delta_K)+d(i\delta_K)}$) with initial value $K^{a+\varepsilon}$ never attains $K^a$.
This quantity is well known and equal to
\[1 -\Big({d(i\delta_K)\over b(i\delta_K)}\Big)^{K^{a+\varepsilon}-K^a}.\]
Since $\alpha=\max_{x\in\T}d(x)/b(x)<1$, it follows from \eqref{hyp:deltaK} that
\begin{align*}
\mathbb{P}(\theta''_{K}=+\infty) &\geq \exp\Bigl({1\over \delta_K} \log\big(1 -\alpha^{K^{a+\varepsilon}-K^a}\big)\Bigr)\\
&\sim \exp\Bigl(- {1\over \delta_K}\alpha^{K^{a+\varepsilon}-K^a}\Bigr)\gg 1-K^{{a_2}/4} \alpha^{K^{a+\varepsilon}-K^a},
\end{align*}
which tends to $1$ when $K$ tends to infinity.
\end{proof}

\begin{prop}
\label{prop:Lipschitz-estimate}
Under the Assumptions \ref{hypmodele} and \ref{hypconvergence}, for all $T>0$, there exists $L_T$ in the definition \eqref{def:thetaK(L)} of $\theta_K(L)$ such that
\[
\lim_{K\to+\infty}\mathbb{P}(\theta_K(L_T)>T)=1.
\]
\end{prop}

\begin{proof}[Proof of Proposition \ref{prop:Lipschitz-estimate}]
In view of Lemma \ref{lem:lower-bound} it is enough to prove that there exists $L_T$ such that
$\lim_{K\to +\infty} P(\tau_K(L_T)>T)=1$. For $i\in\{0,\ldots,1/\delta_K-1\}$, let us consider the increments\vspace*{-3pt}\enlargethispage{\baselineskip}
\begin{align*}
\Delta_K\beta^K_{i}(t\wedge \theta_K)&= \frac{\beta_{i+1}^K(t\wedge \theta_K)-\beta_{i}^K(t\wedge \theta_K)}{\delta_K}\\
&= \Delta_KM^K_{i}(t\wedge \theta_K) + \Delta_KA^K_{i}(t\wedge \theta_K).
\end{align*}
We also introduce\vspace*{-3pt}
\begin{equation}
\label{def:g}
g^K_{i}(t)=\Delta_K A^K_{i}(t\wedge \theta_K)+\frac{\|p\|_{\text{Lip}}}{\underline p} A^K_{i+1}(t\wedge\theta_K).
\end{equation}
To prove the result we will show that we can control $\Delta_K\beta^K_{i}(t\wedge \theta_K)$. To this end, we will first control $g_i^K(t)$ using an almost sure maximum principle. We will then use the fact that $\Delta_K M^K_{i}(t)$ is small in $\Omega_K(L)$ to obtain a control on $\Delta_K\beta^K_{i}(t\wedge \theta_K)$. We~provide the proof in several steps.

\subsubsection*{Step 1} As a first step we prove that for all $t\leq\theta_K(L)\wedge T$,\vspace*{-3pt}
\[
{d g^K_{i}(t)\over dt} \le C(K,L)
+ \bar p \log(K)\sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor}
h_{K} G(h_{K}\ell) \left[g^K_{\ell+i}(t) - g^K_{i}(t)\right]^+ \, e^{h_{K}L|\ell|}.
\]

Using \eqref{def:A}, we obtain for $K$ large enough and $0\leq s<t\leq T$,
\begin{multline}\label{eq:horrible}
 g^K_{i}(t)-g^K_{i}(s) \\
 =\frac{1}{h_K}\int_{(s\wedge\theta_K)\log(K)}^{(t\wedge\theta_K)\log(K)}\Big[b((i+1)\delta_K)N^K_{i+1}(u) \log\left(1+\sfrac{1}{N^K_{i+1}(u)}\right)\\
 \hspace{7cm} -b(i\delta_K)N^K_{i}(u)
\log\left(1+\sfrac{1}{N^K_{i}(u)}\right)\Big] du \\
 +\frac{1}{h_K}\int_{(s\wedge\theta_K)\log(K)}^{(t\wedge\theta_K)\log(K)} \Big[d((i+1)\delta_K)N^K_{i+1}(u) \log\left(1-\sfrac{1}{N^K_{i+1}(u)}\right)\\
 \hspace{7cm} -d(i\delta_K)N^K_{i}(u)
\log\left(1-\sfrac{1}{N^K_{i}(u)}\right)\Big]du  \\
 \hspace*{-2.4cm}+\frac{1}{h_K}\int_{(s\wedge\theta_K)\log(K)}^{(t\wedge\theta_K)\log(K)} \sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor} h_{K} G(h_{K}\ell)
\\
\hskip 1cm \Big[ p((i+1+\ell) \delta_{K})N^K_{i+1+\ell}(u)\log\left(1+\sfrac{1}{N^K_{i+1}(u)}\right)\\
 \hspace{6cm} -p((i+\ell) \delta_{K})N^K_{i+\ell}(u)\log\left(1+\sfrac{1}{N^K_{i}(u)}\right)\Big]du  \\
\hspace*{.8cm}+\frac{\|p\|_{\text{Lip}}}{\underline p\log(K)} \int_{(s\wedge\theta_K)\log(K)}^{(t\wedge\theta_K)\log(K)}\left[b((i+1)\delta_K)N^K_{i+1}(u) \log\left(1+\sfrac{1}{N^K_{i+1}(u)}\right)\right.\\
 \hspace{5.5cm}+\left. d((i+1)\delta_K)N^K_{i+1}(u) \log\left(1-\sfrac{1}{N^K_{i+1}(u)}\right)\right]du  \\
\hspace*{.9cm} +\frac{\|p\|_{\text{Lip}}}{\underline p\log(K)} \int_{(s\wedge\theta_K)\log(K)}^{(t\wedge\theta_K)\log(K)} \sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor} h_{K} G(h_{K}\ell)\
p((i+1+\ell)\delta_K) \\
 \hspace{7cm} \times N^K_{i+1+\ell}(u)\log\left(1+\sfrac{1}{N^K_{i+1}(u)}\right)du
\end{multline}\vspace*{-3\baselineskip}
\begin{multline*}
\hspace*{1cm} \leq \frac{C(\bar b+\bar d)}{ h_K}\int_{(s\wedge\theta_K)\log(K)}^{(t\wedge\theta_K)\log(K)}\Bigl[\frac{1}{N^K_{i+1}(u)}+\frac{1}{N^K_{i}(u)}\Bigr]du \\
 \hspace{5cm} +(C_1(\bar b+\bar d)+\|b\|_{\text{Lip}}+\|d\|_{\text{Lip}})(t\wedge \theta_K-s\wedge \theta_K)  \\
\hspace*{-.7cm} +\frac{1}{h_K}\int_{(s\wedge\theta_K)\log(K)}^{(t\wedge\theta_K)\log(K)} \sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor} h_{K} p((\ell+i)\delta_K) G(h_{K}\ell)\\
 \hspace{2cm}\times \left[N_{\ell+i+1}^K(u)\log \left(1+\sfrac{1}{N_{i+1}^K(u)}\right)-N_{\ell+i}^K(u)\log \left(1+\sfrac{1}{N_{i}^K(u)}\right)
\right]du  \\
\hspace*{.1cm} +\frac{3\|p\|_{\text{Lip}}}{\underline{p}\log(K)}\int_{(s\wedge\theta_K)\log(K)}^{(t\wedge\theta_K)\log(K)} \sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor}h_Kp((\ell+i)\delta_K)G(h_K\ell)\\
 \hspace{6cm}N_{\ell+i+1}^K(u)\log \left(1+\sfrac{1}{N_{i+1}^K(u)}\right)du,
\end{multline*}
where we used to prove the last inequality that, for all $x$ such that $|x|\leq 1/2$, we have\vspace*{-3pt}
\begin{align}
\Bigl|\frac{1}{x}\log(1+x)\Bigr|&\leq C,
\label{approxlog3}\\
\Bigl |\frac{1}{x}\log(1+x)-\frac{1}{y}\log(1+y) \Bigr|&\leq C(|x|+|y|),
\label{approxlog4}
\end{align}
and the fact that (recalling the convention~\eqref{eq:convention} and that $p$ is periodic)\vspace*{-3pt}
\begin{align*}%\label{eq:borne-Lipschitz-ppe-max}
\left|p((\ell+i+1)\delta_K)-p((\ell+i)\delta_K)\right|&= p((\ell+i)\delta_K)\frac{\left|p((\ell+i+1)\delta_K)-p((\ell+i)\delta_K)\right|}{p((\ell+i)\delta_K)} \\
&\leq\frac{\|p\|_{\text{Lip}}\delta_K}{\underline{p}}p((\ell+i)\delta_K).
\end{align*}
Note also that, to obtain the last inequality \eqref{eq:horrible}, we have taken $K$ large enough such that $\sfrac{\|p\|_{\text{Lip}}\delta_K}{\underline{p}}\leq 1$, so that\vspace*{-3pt}
\[
p((\ell+i+1)\delta_K)\leq 2p((\ell+i)\delta_K).
\]
Next, notice that for all {$x',y',x,y$ such that $|x|,|y|\leq 1/2$,\vspace*{-3pt}
\begin{equation*}%\label{approx-bis}
{1 \over y'}\log(1+{ y}) -{ 1\over x'}\log(1+{ x}) \le {y\over y'} - {x\over x'} + C\Big({y^2\over y'}+{x^2\over x'}\Big).
\end{equation*}
Using this inequality and \eqref{eq:borne-N_i/N_j}, we have\vspace*{-3pt}
\begin{align*}
\frac{1}{h_K} & \int_{(s\wedge\theta_K)\log(K)}^{(t\wedge\theta_K)\log(K)} \sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor} h_{K} p((\ell+i)\delta_K) G(h_{K}\ell)\\
& \hspace{1cm} \times \left[N_{\ell+i+1}^K(u)\log \left(1+\sfrac{1}{N_{i+1}^K(u)}\right)-N_{\ell+i}^K(u)\log \left(1+\sfrac{1}{N_{i}^K(u)}\right)
\right]du \\
& \leq \frac{1}{h_K}\int_{(s\wedge\theta_K)\log(K)}^{(t\wedge\theta_K)\log(K)} \sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor} h_Kp((\ell+i)\delta_K)G(h_K\ell)\\[-5pt]
& \hspace{7cm} \times \biggl[\frac{N_{\ell+i+1}^K(u)}{N_{i+1}^K(u)}-\frac{N_{\ell+i}^K(u)}{N_{i}^K(u)}
\biggr]du \\
& +\frac{C\bar p}{h_K}\int_{(s\wedge\theta_K)\log(K)}^{(t\wedge\theta_K)\log(K)} \sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor} h_KG(h_K\ell)e^{Lh_K|\ell|}\\[-5pt]
& \hspace{7cm} \times \left[\frac{1}{N_{i+1}^K(u)}+\frac{1}{N_{i}^K(u)}\right]\ du.
\end{align*}
Therefore, using \eqref{eq:conv_moment-expo} and the definition of $\tau'_K$, we have proved that\vspace*{-.5\baselineskip}\enlargethispage{\baselineskip}
\begin{multline*}
g^K_{i}(t) - g^K_{i}(s)
\begin{aligned}[t]
& \leq \biggl(\frac{C\bar G(L)\log(K)}{ K^a \ h_K}+C\biggr)(t-s) \\
& +\frac{1}{h_K}\int_{(s\wedge\theta_K)\log(K)}^{(t\wedge\theta_K)\log(K)} \sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor} h_Kp((\ell+i)\delta_K)G(h_K\ell)\\
& \hspace{6cm}\times \biggl[\frac{N_{\ell+i+1}^K(u)}{N_{i+1}^K(u)}-\frac{N_{\ell+i}^K(u)}{N_{i}^K(u)}
\biggr]\ du \\
& +\frac{{3}\|p\|_{\text{Lip}}}{\underline{p}\log(K)}\int_{(s\wedge\theta_K)\log(K)}^{(t\wedge\theta_K)\log(K)} \sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor} h_Kp((\ell+i)\delta_K)G(h_K\ell)\\
& \hspace{6cm} \times \frac{N_{\ell+i+1}^K(u)}{N_{i+1}^K(u)}du.
\end{aligned}
\end{multline*}
Using that for any real numbers $\lambda,\alpha$, $e^\lambda\le e^\alpha+ e^\lambda(\lambda-\alpha)$, we have
\begin{align*}
\frac{1}{h_K}&\int_{(s\wedge\theta_K)\log(K)}^{(t\wedge\theta_K)\log(K)} \sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor} h_Kp((\ell+i)\delta_K)G(h_K\ell)\biggl[\frac{N_{\ell+i+1}^K(u)}{N_{i+1}^K(u)}-\frac{N_{\ell+i}^K(u)}{N_{i}^K(u)}
\biggr]\ du
\\
& \leq \frac{\log(K)}{ h_K}\int_{(s\wedge\theta_K)\log(K)}^{(t\wedge\theta_K)\log(K)} \sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor} h_Kp((\ell+i)\delta_K)G(h_K\ell) \\[-3pt]
& \hspace{2cm}\times \biggl(\beta^K_{\ell+i+1}\Bigl(\frac{u}{\log(K)}\Bigr)-\beta^K_{\ell+i}\Bigl(\frac{u}{\log(K)}\Bigr)\\
& \hspace{4cm} -\Bigl(\beta^K_{i+1}\Bigl(\frac{u}{\log(K)}\Bigr)-\beta^K_i\Bigl(\frac{u}{\log(K)}\Bigr)\Bigl)\biggr) \frac{N^K_{\ell+i+1}(u)}{N^K_{i+1}(u)}du \\[-3pt]
& \leq \int_{(s\wedge\theta_K)\log(K)}^{(t\wedge\theta_K)\log(K)} \sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor} h_Kp((\ell+i)\delta_K)G(h_K\ell)\\[-3pt]
& \hspace*{2cm}\times\left(\Delta_K\beta^K_{\ell+i}\Bigl(\frac{u}{\log(K)}\Bigr)-\Delta_K\beta^K_{i}\Bigl(\frac{u}{\log(K)}\Bigr)\right)
\frac{N^K_{\ell+i+1}(u)}{N^K_{i+1}(u)}du \\
&\leq \int_{(s\wedge\theta_K)\log(K)}^{(t\wedge\theta_K)\log(K)} \sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor} h_Kp((\ell+i)\delta_K)G(h_K\ell)\\[-3pt]
& \hspace{2cm} \times\biggl(\Delta_KM^K_{\ell+i}\Bigl(\frac{u}{\log(K)}\Bigr)-\Delta_KM^K_{i}\Bigl(\frac{u}{\log(K)}\Bigr) \\[-3pt]
& \hspace{3.5cm}+ \Delta_KA^K_{\ell+i}\Bigl(\frac{u}{\log(K)}\Bigr)-\Delta_KA^K_{i}\Bigl(\frac{u}{\log(K)}\Bigr)\biggr) \frac{N^K_{\ell+i+1}(u)}{N^K_{i+1}(u)}du ,
\end{align*}
using~\eqref{decomposition}. Thus, using~\eqref{hyp:deltaK} and~\eqref{def:g}, we deduce that\vspace*{-.5\baselineskip}
\begin{multline*}
g^K_{i}(t)-g^K_{i}(s)
- \int_{(s\wedge\theta_K)\log(K)}^{(t\wedge\theta_K)\log(K)} \sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor}
h_{K} p((\ell+i)\delta_K)G(h_{K}\ell) \\[-5pt]
\times \Bigl(g^K_{\ell+i}\Bigl(\frac{u}{\log(K)}\Bigr)-g^K_i\Bigl(\frac{u}{\log(K)}\Bigr)\Bigr)
\frac{N^K_{\ell+i+1}(u)}{N^K_{i+1}(u)}du
\end{multline*}
\begin{align*}
&\leq C_0(K,L)(t-s)\\
& +\frac{{3}\|p\|_{\text{Lip}}}{\underline{p}\log(K)}\int_{(s\wedge\theta_K)\log(K)}^{(t\wedge\theta_K)\log(K)} \sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor}h_Kp((\ell+i)\delta_K)G(h_K\ell)
\frac{N_{\ell+i+1}^K(u)}{N_{i+1}^K(u)}du \\
& -\frac{\|p\|_{\text{Lip}}}{\underline{p}}\int_{(s\wedge\theta_K)\log(K)}^{(t\wedge\theta_K)\log(K)} \sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor}
h_{K} p((\ell+i)\delta_K)G(h_{K}\ell)\\
& \hspace{2cm}\times\left(\beta_{\ell+i+1}\Bigl(\frac{u}{\log(K)}\Bigr)-\beta_{i+1}\Bigl(\frac{u}{\log(K)}\Bigr)\right)
\frac{N^K_{\ell+i+1}(u)}{N^K_{i+1}(u)}du \\
& + \frac{\|p\|_{\text{Lip}}}{\underline{p}}\int_{(s\wedge\theta_K)\log(K)}^{(t\wedge\theta_K)\log(K)} \sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor}
h_{K} p((\ell+i)\delta_K)G(h_{K}\ell)\\
& \hspace{2cm}\times\left(M_{\ell+i+1}\Bigl(\frac{u}{\log(K)}\Bigr)-M_{i+1}\Bigl(\frac{u}{\log(K)}\Bigr)\right)
\frac{N^K_{\ell+i+1}(u)}{N^K_{i+1}(u)}du\\
& + \int_{(s\wedge\theta_K)\log(K)}^{(t\wedge\theta_K)\log(K)} \sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor}
h_{K} p((\ell+i)\delta_K)G(h_{K}\ell)\\
& \hspace{2cm}\times\left(\Delta_KM^K_{\ell+i}\Bigl(\frac{u}{\log(K)}\Bigr)-\Delta_KM^K_{i}\Bigl(\frac{u}{\log(K)}\Bigr)\right)
\frac{N^K_{\ell+i+1}(u)}{N^K_{i+1}(u)}du,
\end{align*}
where \[C_0(K,L) = C+\frac{C\bar G(L)}{\delta_K K^a}.\] Now, using \eqref{eq:borne-N_i/N_j}, we have on the event $\Omega_K(L)$ (recall \eqref{event:OmegaK}) whose probability tends to 1 (by Lemma \ref{corol:accrM}) that
\begin{multline*}
\int_{(s\wedge\theta_K)\log(K)}^{(t\wedge\theta_K)\log(K)} \sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor}
h_{K} p((\ell+i)\delta_K)G(h_{K}\ell)\\[-3pt]
\shoveright{\times\left|\Delta_KM^K_{\ell+i}\Bigl(\frac{u}{\log(K)}\Bigr)-\Delta_KM^K_{i}\Bigl(\frac{u}{\log(K)}\Bigr)\right|
\frac{N^K_{\ell+i+1}(u)}{N^K_{i+1}(u)}du} \\
\leq 2\bar p \log(K) (t-s)\varepsilon_K\sum_{\ell\in\Z} h_K G(h_k\ell)e^{L h_K|\ell|}\leq 2\bar{p}\log(K) \varepsilon_K \bar G(L)(t-s),
\end{multline*}
using \eqref{eq:conv_moment-expo} again.
Similarly, using~\eqref{eq:borne-martingale-simple}, we have that on the event $\Omega_K(L)$, with a probability tending to 1,
\begin{multline*}
\int_{(s\wedge\theta_K)\log(K)}^{(t\wedge\theta_K)\log(K)} \sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor}
h_{K} p((\ell+i)\delta_K)G(h_{K}\ell)\\
\times \left|M_{\ell+i+1}\Bigl(\frac{u}{\log(K)}\Bigr)-M_{i+1}\Bigl(\frac{u}{\log(K)}\Bigr)\right|
\frac{N^K_{\ell+i+1}(u)}{N^K_{i+1}(u)}du\\
\leq \bar{p}\log(K) \varepsilon_K \bar G(L)(t-s).
\end{multline*}
To conclude, we use the inequality {$e^x(3-x)\leq e^2$} for all $x\in\mathbb{R}$ to deduce that
\[
\frac{N^K_{\ell+i+1}(u)}{N^K_{i+1}(u)}\biggl[{3}-\log(K)\biggl(\beta_{\ell+i+1}\Bigl(\frac{u}{\log(K)}\Bigr)-\beta_{i+1}\Bigl(\frac{u}{\log(K)}\Bigr)\biggr)\biggr]\leq {e^2}.
\]
Combining the four previous inequalities, we deduce that on the event $\Omega_K(L)$,
\begin{multline*}g^K_{i}(t)-g^K_{i}(s) \leq C(K,L)(t-s)\\
+
\bar p \log(K) \int_{(s\wedge\theta_K)}^{(t\wedge\theta_K)} \sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor}
h_{K} G(h_{K}\ell) \left[g^K_{\ell+i}(v) - g^K_{i}(v)\right]^+ \, e^{h_{K}L|\ell|}dv,\end{multline*}
where $[x]^+=x\vee 0$ is the positive part of $x$ and
\[
C(K,L)=C+\frac{C\bar G(L)}{\delta_K K^
a} +2\bar{p}\log(K) \varepsilon_K \bar G(L)+ \frac{\|p\|_{\text{Lip}}}{\underline{p}}\bar{p}e^2.
\]
Thus, for all $t\leq\theta_K(L)\wedge T$, and on the event $\Omega_K(L)$,
\[
{d g^K_{i}(t)\over dt} \le C(K,L)
+ \bar p \log(K)\sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor}
h_{K} G(h_{K}\ell) \left[g^K_{\ell+i}(t) - g^K_{i}(t)\right]^+ \, e^{h_{K}L|\ell|}.
\]

\subsubsection*{Step 2} We next provide an upper bound on $g_i^K(t)$. To this end, we will
use the maximum principle for $\omega\in\Omega_K(L)$ fixed.
Defining $\widetilde g^K_{i}(t) = g^K_{i}(t) -2C(K,L) t$,
we deduce that for any $t\le \theta_{K}(L)\wedge T$,
\begin{equation}\label{derive}
{d \widetilde g^K_{i}(t)\over dt} < \bar p \log(K)
\sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor}
h_{K} G(h_{K}\ell) \left[\widetilde g^K_{\ell+i}(t) - \widetilde g^K_{i}(t)\right]^+ \, e^{h_{K}L|\ell|}.
\end{equation}
Let us introduce
\[(i_{K}, t_{K}) = (i_K(\omega),t_K(\omega))=\text{argmax}_{{i\in \{0,\dots, \sfrac{1}{\delta_{K}}-1\}, t\in [0, \theta_{K}(\omega)\wedge T]}}\ \widetilde g^K_{i}(t)\]
and let us prove that \[t_{K}=0.\]
By contradiction, if we assume that $t_{K}>0$, then the right term of \eqref{derive} is non-positive for $i=i_{K}$ and then the
left term is negative, contradicting the fact that $\widetilde g^K_{i_K}(t)$ is maximal for $t=t_K$. Hence, we have proved that, almost surely on
the event $\Omega_K(L)$, for all $t\leq \theta_K(L)\wedge T$ and $0\leq i\leq 1/\delta_K-1$,
\begin{align*}
g^K_i(t)= \widetilde g^K_i(t)+2C(K,L) t
&\leq \max_{0\leq j\leq 1/\delta_K-1}\widetilde g^K _j(0)+2 C(K,L) t\\
&= \max_{0\leq j\leq 1/\delta_K-1} g^K _j(0)+2C(K,L) t.
\end{align*}

\subsubsection*{Step 3} We next provide an upper bound on $|\Delta_K\beta_i^K|$ which will allow us to conclude the proof.
By Proposition~\ref{est-varpart}, for $t\leq T$,
\begin{multline*}
\Delta_K\beta_i^K(t\wedge\theta_K) =g_i^K(t) - \frac{\|p\|_{\text{Lip}}}{\underline{p}} A_{i+1}^K(t\wedge\theta_K)+
\Delta_KM_i^K(t\wedge\theta_K) \\
\leq \max_{0\leq j\leq 1/\delta_K} g^K _j(0)+2 C(K,L) t+ \frac{\|p\|_{\text{Lip}}}{\underline{p}}\Bigl(\max_{0\leq j\leq 1/\delta_K-1}\beta^K_i(0)+C_1 t\Bigl)+\varepsilon_K.
\end{multline*}
A similar argument applied to $(\beta^K_i(t\wedge\theta_K)-\beta^K_{i-1}(t\wedge\theta_K))/\delta_K$ gives the converse inequality, so, recalling that $A^K_i(0)=\beta^K_i(0)$, we finally obtain that there exists a constant $C$ independent of $K$, $i$, $t$ and $L$ such that, almost surely on the event $\Omega_K(L)$, for all $i\in\{0,\ldots,1/\delta_K-1\}$ and $t\leq T$,
\begin{multline*}
|\Delta_K\beta_i^K(t\wedge\theta_K)|\leq C\Big[\max_{0\leq j\leq 1/\delta_K-1} \left(|\Delta_K \beta^K_i(0)|+\beta^K_i(0)\right)+1+T\\
+\bar{G}(L) T\Bigl(\frac{1}{\delta_K K^
a}+\varepsilon_K\log(K)\Bigr)\Big].
\end{multline*}
Finally, defining $\widetilde \Omega_K$ as the event of probability converging to $1$ where
\[\max_{0\leq j\leq 1/\delta_K-1} |\Delta_K \beta^K_i(0)|+\beta^K_i(0)\le A+\|\beta_0\|_\infty+1,\] where the constant $A$ comes from Assumption~\ref{hypconvergence}, on the event $\Omega_K(L) \cap \widetilde \Omega_K$, we have for $t\leq T$,
\begin{equation*}
|\Delta_K\beta_i^K(t\wedge\theta_K)|\leq C\biggl[A+\|\beta_0\|_\infty+2 +T+ \bar G(L) T\Bigl(\frac{1}{\delta_K K^
a}+\varepsilon_K\log(K)\Bigr)\biggr].
\end{equation*}

To conclude the proof of Proposition~\ref{prop:Lipschitz-estimate}, we first fix $T>0$, set
\begin{equation*}%\label{eq:L0}
L_T=C(A+\|\beta_0\|_\infty+3+T)
\end{equation*}
and choose $K$ large enough such that $C\,T\,\bar G(L)\bigl(\frac{1}{\delta_K K^
a}+\varepsilon_K\log(K)\bigr)< 1$.
Then
\[\P(\tau_K(L_T)>T)\ge \P(\Omega_K(L_T)\cap \widetilde \Omega_K)\xrightarrow[K\to\infty]{} 1.\]
Combining the last estimate with Lemma~\ref{lem:lower-bound} ends the proof of Proposition~\ref{prop:Lipschitz-estimate}.}
\end{proof}

\subsection{Proof of Theorem~\ref{thm:tight}}
\label{sec:tight}

In a first step, we prove that the sequence of laws of $(\tilde \beta^K_{t}, t\in[0,T])_{K}$ is tight in ${\cal P}(\mathbb{D}([0,T], \Co(\T,\mathbb{R})))$. We will then check that it is actually $C$-tight.

Let us recall that the random functions $\ \tilde \beta^K\in\mathbb{D}([0,T], \Co(\T,\mathbb{R}))$ are defined in \eqref{lafonction} as follows. For all $x\in\T$, let $i\in\{0,\ldots,1/\delta_K-1\}$ be such that $x\in [i\delta_K,(i+1)\delta_K)$, and set
\[
\widetilde \beta^K_{t}(x):=\widetilde \beta^K(t,x) =\beta_i^K(t)\Big(1-\f {x}{\delta_K}+i\Big)+\beta_{i+1}^K(t)\Big(\f {x}{\delta_K}-i\Big),
\]
where, by convention, $\beta^K_{1/\delta_K}(t)=\beta^K_0(t)$.
Let us recall that the proof of Theorem~\ref{thm:tight} is based on the criterion of Jakubowski \cite{jakubowski} recalled in Section~\ref{sec:main-steps}. Our goal is to prove Conditions~\eqref{enum:i} and~\eqref{enum:ii} therein.

Let us first prove \eqref{enum:i}. By Ascoli's theorem, we know that a compact set $K_{\varepsilon}$ is a set of equi-continuous and equi-bounded functions. By Corollary \ref{lem:upper-bound} and Proposition \ref{prop:Lipschitz-estimate}, we have, on the event $\{\theta_K(L_T)>T\}$ of probability converging to $1$ when $K$ tends to infinity,
that, for all $x\in\T$ and all $t\in[0,T]$,
\[\widetilde \beta^K_{t}(x) =(x-i\delta_K)\Delta_K\beta^K_i(t) + \beta_{i}^K(t) \leq L_T\delta_K +C(T),\]
so the sequence $(\widetilde \beta^K_{t}, t\in[0,T])_{K}$ is equi-bounded.
Furthermore, recall that, by~\eqref{def:tau2}, for $x,y\in\T$,
\[\left|\widetilde \beta^K_{t}(x)-\widetilde \beta^K_{t}(y)\right|= \rho(x,y)\sup_{0\leq j\leq 1/\delta_K-1}\left|\Delta_K\beta^K_j(t)\right|
\leq L_T \rho(x,y).\]
We deduce that the sequence is equi-continuous and~(i) is proved.

Let us now prove \eqref{enum:ii}, \ie that for all
$f\in \Co(\T,\mathbb{R})$, the sequence of laws of the real-valued processes
\begin{multline*}
X^K_f(t)= \int_\T \widetilde \beta^K(t,x) f(x) dx =\sum_{i=0}^{1/\delta_K-1} \bigg[\beta^K_{i}(t) \int_{i\delta_{K}}^{(i+1)\delta_{K}} \Big(1+i -{x\over \delta_{K}} \Big)f(x)dx \\
+\beta^K_{i+1}(t) \int_{i\delta_{K}}^{(i+1)\delta_{K}} \Big({x\over \delta_{K}}-i\Big) f(x)dx \bigg]
\end{multline*}
is tight. Recalling that $\beta_{i}^K(t)= A^K_{i}(t) + M^K_{i}(t)$, the process $X^K_f$ is a local semi-martingale with Doob-Meyer decomposition
$X^K_{f} = A^K_{f} + M^K_{f} $, where
\[
A^K_{f} (t) = \sum_{i} \bigg[A^K_{i}(t) \int_{i\delta_{K}}^{(i+1)\delta_{K}} \Bigl(1+i -{x\over \delta_{K}}\Bigr)f(x)dx +A^K_{i+1}(t) \int_{i\delta_{K}}^{(i+1)\delta_{K}} \Bigl({x\over \delta_{K}}-i\Bigr) f(x)dx \bigg]
\]
and $M^K_f$ is defined similarly using $M^K_i(t)$ instead of $A^K_i(t)$.

We use Aldous and Rebolledo criteria (see for example Joffe-Métivier \cite{joffemetivier}) to prove the tightness of the sequence $(X^K_f)$.
Let $S$ be a stopping time for the filtration of the underlying Poisson point measures, a.s. in $[0,T]$.
We need to estimate for $\alpha>0$, the quantity
$\mathbb{P}(|A^K_{f}((S+\alpha)\wedge T) -A^K_{f} (S)|>\eta)$ for $\eta>0$.
From~\eqref{def:A}, we deduce
\begin{multline*}
A^K_{f}((S+\alpha)\wedge T) -A^K_{f} (S) =
 \sum_{i} \biggl\{\bigg(\int_{i\delta_{K}}^{(i+1)\delta_{K}} \Bigl(1+i -{x\over \delta_{K}}\Bigr)f(x)dx \biggr) \\
\bigg[
\frac{1}{\log(K)}\int_{S \log(K)}^{((S+\alpha)\wedge T)\log(K)}\Bigl(b(i \delta_{K})N^K_{i}(s)\log\left(1+\sfrac{1}{N^K_{i}(s)}\right)\\
\hspace*{-1cm}+ d(i \delta_{K})N^K_{i}(s)\log\left(1-\sfrac{1}{N^K_{i}(s)}\right)\Bigr) ds
\\
\hspace*{2cm}+{\frac{1}{\log(K)}}\, \sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor}h_Kp((\ell+i)\delta_K)G(h_K\ell)\\
\times\int_{S \log(K)}^{((S+\alpha)\wedge T)\log(K)} N^K_{\ell+i}(s)\log\left(1+\sfrac{1}{N^K_{i}(s)}\right) ds\bigg]
\end{multline*}%
\vspace*{-1.5\baselineskip}
\begin{multline*}
+ \biggl(\int_{i\delta_{K}}^{(i+1)\delta_{K}} ({x\over \delta_{K}}-i) f(x)dx\biggr)\\
\times \bigg[
\frac{1}{\log(K)}\int_{S \log(K)}^{((S+\alpha)\wedge T)\log(K)}\bigg(b((i+1) \delta_{K})N^K_{i+1}(s)\log\left(1+\sfrac{1}{N^K_{i+1}(s)}\right)\\
+ d((i +1) \delta_{K})N^K_{i+1}(s)\log\left(1-\sfrac{1}{N^K_{i+1}(s)}\right)\bigg) ds \\
\hspace*{2cm} +{\frac{1}{\log(K)}} \sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor}h_Kp((\ell+i+1)\delta_K)G(h_K\ell)
\\
\times \int_{S \log(K)}^{((S+\alpha)\wedge T)\log(K)} N^K_{\ell+i+1}(s)\log\left(1+\sfrac{1}{N^K_{i+1}(s)}\right) ds\bigg]\bigg\}.
\end{multline*}
Using \eqref{approxlog3} and the definition of $\theta_K(L_T)$, proceeding as in the proof of Proposition~\ref{prop:Lipschitz-estimate} we have
\begin{multline*}
\mathbb{E}(|A^K_{f}((S+\alpha)\wedge\theta_K(L_T)\wedge T) -A^K_{f} (S\wedge\theta_K(L_T))|) \\
 \leq {C\over \log(K)}\sum_{i=0}^{1/\delta_K-1} \int_{i\delta_{K}}^{(i+1)\delta_{K}} |f(x)|dx\bigg\{ 2(\bar b+\bar d)\,\alpha\log(K)\\ 
 \qquad+ \sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor}\hspace*{-3pt}h_Kp((\ell+i)\delta_K)G(h_K\ell)\times \mathbb{E}\Big(\int_{(S\wedge\theta_K(L_T)) \log(K)}^{((S+\alpha)\wedge\theta_K(L_T)\wedge T)\log(K)} {N^K_{\ell+i}(s)\over N^K_{i}(s)}ds\Bigr)\\
 + \sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor}h_Kp((\ell+i+1)\delta_K)G(h_K\ell) \\
\shoveright{\times \mathbb{E}\Bigl(\int_{(S\wedge\theta_K(L_T)) \log(K)}^{((S+\alpha)\wedge\theta_K(L_T)\wedge T)\log(K)} {N^K_{\ell+i+1}(s)\over N^K_{i+1}(s)}ds\Bigr)\!\biggr\}}\\
\leq \alpha C\left[2(\bar b+\bar d)+2\bar p\overline{G}(L) \right]\|f\|_\infty.\hspace*{5.5cm} 
\end{multline*}
By Proposition \ref{prop:Lipschitz-estimate}, $\theta_K(L_T)>T$ with probability converging to 1, so we deduce from Markov's inequality that, for all $\varepsilon>0$ and $\eta>0$, there exists $\alpha$ such that,
\[
\limsup_{K\to+\infty}\sup_S\P(|A^K_{f}((S+\alpha)\wedge T) -A^K_{f} (S)|>\eta)\leq\varepsilon,
\]
where the supremum is taken over all stopping times $S\leq T$. This is Aldous criterion for $A^K_{f}(t)$.

It remains to prove a similar property replacing $A^K_{f}$ by $\langle M^K_{f} \rangle$. This can be done similarly using~\eqref{def:M}. Computations are actually simpler by Lemma~\ref{lem:M}.

Hence we have proved that the sequence of laws of $(\tilde \beta^K_{t}, t\in[0,T])_{K}$ is relatively compact in ${\cal P}(\mathbb{D}([0,T],{\cal C}(\mathbb{T},\mathbb{R})))$. We prove below that this sequence is actually $C$-tight and that the Lipschitz estimate~\eqref{eq:Lipschitz-estimate} is satisfied with the value of $L_T$ of Proposition~\ref{prop:Lipschitz-estimate}.

Since $|\beta_{i}^K(t)-\beta_{i}^K(t-)| \leq C/\log(K)$ for any $K$, $i$ and $t$, we have
\[\lim_{K} \mathbb{P}(\sup_{t\le T} \|\widetilde\beta^K_{t}-\widetilde\beta^K_{t-}\|_\infty >\varepsilon) = 0.\]
Then, we deduce from Proposition 3.26 in Jacod-Shiryaev \cite[p.\,315]{jacod} that, for all $f\in \Co(\T,\R)$, the sequence of laws of $X_f^K$ defined in~\eqref{def-Xf} is $C$-tight. We proceed by contradiction to deduce that $(\widetilde{\beta}^K)$ is also $C$-tight: if this is not true, there exists an event $\Omega_1$ of positive probability such that, for all $\omega\in\Omega_1$, there exists $t_0(\omega)$, $\alpha(\omega)$ and a ball $B(\omega)\subset \T$ of positive radius such that, for all $x\in B$, \begin{equation}\label{etape1}
|\widetilde{\beta}^K_{t_0}(x)-\widetilde{\beta}^K_{t_0-}(x)|>\alpha.
\end{equation}Therefore, there exists non-random $\alpha>0$ and $\varepsilon>0$ and $i\in\{0,1,\ldots, \lfloor 1/\varepsilon\rfloor-1\}$ and an event $\Omega_2\subset\Omega_1$ of positive probability such that~\eqref{etape1} holds true for all $x\in[i\varepsilon,(i+1)\varepsilon]$ and for this non-random $\alpha$. Now, we define $f_i\in \Co(\T,\R)$ with support in $[i\varepsilon,(i+1)\varepsilon]$ and positive on $(i\varepsilon,(i+1)\varepsilon)$. Then, for all $\omega\in\Omega_2$,
\[
\liminf_{K\to+\infty}|X_{f_i}^K(t_0)-X_{f_i}^K(t_0-)|\geq \alpha \inf_{x\in [(i+1/3)\varepsilon,(i+2/3)\varepsilon]}|f_i(x)|>0.
\]
This is a contradiction with the $C$-tightness of $(X^K_{f_i})$.

We now prove the Lipschitz estimate for $\beta$. Using the Skorohod representation theorem, we can construct copies $\hat{\beta}^K$ of $\tilde{\beta}^K$ and $\hat{\beta}$ of $\beta$ in the Polish space $\mathbb{D}([0,T], \Co(\T,\mathbb{R}))$, such that $(\hat{\beta}^K)$ converges (up to a subsequence) almost surely for the $\mathbb{L}^\infty$ norm on $[0,T]$ to $\hat{\beta}$. We then define
\[
\hat{\tau}^K=\inf_{i\neq j \in \{0,\ldots,1/\delta_K-1\}}\inf\biggl\{t\geq 0: \frac{|\hat{\beta}^K(t,i\delta_K)-\hat{\beta}^K(t,j\delta_K)|}{\rho(i\delta_K,j\delta_K)}>L_T\biggr\}.
\]
Then $\hat{\tau}_K$ is distributed as $\tau_K$ in~\eqref{def:tau}. It then follows from Proposition~\ref{prop:Lipschitz-estimate} that $\hat{\tau}_K>T$ with probability converging to 1. Hence, for all $x\neq y\in\T$, almost surely
\[
|\hat{\beta}(t,x)-\hat{\beta}(t,y)|=\lim_{K\to+\infty}|\hat{\beta}^K(t,x)-\hat{\beta}^K(t,y)|\leq L_T \,\rho(x,y).
\]
By continuity of $\hat{\beta}$ we deduce that this property holds, almost surely, for all $x,y\in\T$.

\section{Identification of the limit as a viscosity solution of\\ a Hamilton-Jacobi equation}\label{sec:identif}

Theorem~\ref{thm:main} will be deduced from Theorems~\ref{thm:tight} and~\ref{thm:identification}.
In the previous section, we proved Theorem~\ref{thm:tight}, \ie the $C$-tightness of the laws of $(\widetilde \beta^K_{t}, t\in[0,T])_{K}$ for all $T>0$. Hence, the sequence of laws of $(\widetilde \beta^K_{t}, t\in[0,T])_{K}$ admits at least one limiting value. Our aim is now to prove Theorem~\ref{thm:identification}, \ie to identify the limiting path as the unique viscosity solution of the Hamilton-Jacobi equation \eqref{eq:HJ}.

Let $\beta$ be distributed as a limiting value of the laws of $(\tilde \beta^K_{t}, t\in {[0,T]})_{K}$. By Theorem~\ref{thm:tight}, $\beta$ belongs to $C([0,T]\times\T,[0,+\infty))$. In the sequel and with an abuse of notation, we denote again by $(\tilde \beta^K_{t}, t\in[0,T])_{K}$ the subsequence that converges in distribution to $\beta$.

We define $\widetilde{A}^K$ as the piecewise affine interpolation of the $A^K_i$ as we did for $\widetilde{\beta}^K$ in~\eqref{lafonction}. It follows from Lemma~\ref{lem:M} that
$
\widetilde{A}^K-\widetilde{\beta}^K
$ converges in law, and thus in probability, to $0$. Therefore $(\widetilde{A}^K-\widetilde{\beta}^K,\widetilde{\beta}^K)$ converges in law to $(0, \beta)$ and thus $(\widetilde{A}^K,\widetilde{\beta}^K)$ converges in law to $(\beta, \beta)$.
We apply the Skorokhod's representation theorem to the random variables $(\widetilde{\beta}^K,\widetilde{A}^K,\beta)$ on the Polish space $\mathbb{D}([0,T],{\cal C}(\mathbb{T}))\times{\cal C}([0,T]\times\mathbb{T})^2$: there exist a new probability space $(\widehat{\Omega},\widehat{\mathcal{A}},\widehat{\mathbb{P}})$, and random processes $\widehat{\beta}^K,\widehat{A}^K,\widehat{\beta}$ and $\widehat{M}^K=\widehat{\beta}^K-\widehat{A}^K$ on this space such that $(\widehat{A}^K,\widehat{\beta}^K)$ has the same distribution as $(\widetilde{A}^K,\widetilde{\beta}^K)$ and converges almost surely to $(\widehat{\beta}, \widehat{\beta})$, where $\widehat{\beta}$ has the same distribution as $\beta$. Let us denote by $\widehat{\Omega}_0$ the event where the convergence holds.

We also define for the value $L_T$ introduced in Proposition~\ref{prop:Lipschitz-estimate}
\[
\widehat{\Omega}_K=\Bigl\{\omega\in \widehat{\Omega} \sep \sup_{x\in\T}\,\sup_{t\in[0,T]}|\widehat{M}^K(t,x)|\leq \varepsilon'_K,
\|\widehat{\beta}^K\|_\text{Lip}\leq L_T,\! \inf_{(t,x)\in[0,T]\times\mathbb{T}} \widehat{\beta}^K(t,x)\geq a \Bigr\},
\]
where $\varepsilon'_K=\delta_K^{-1/2} K^{-a/2}$ converges to 0 by~\eqref{hyp:deltaK}.
It follows from Lemma~\ref{lem:M}, Proposition~\ref{prop:Lipschitz-estimate} and Lemma \ref{lem:lower-bound} that $\P(\widehat{\Omega}_K)\to 1$ when $K\to +\infty$.

Notice that, for $K$ fixed, both $\widehat{A}^K$ and $\widehat{\beta}^K$ are càdlàg functions almost surely with values in the set of piecewise affine functions of $x$ between the grid points $i\delta_K$ (this set of càdlàg functions is a measurable subset of $\mathbb{D}([0,T],{\cal C}(\mathbb{T}))$). We set
\[
\widehat{A}^K_i(t)=\widehat{A}^K(t,i\delta_K),\quad\widehat{\beta}^K_i(t)=\widehat{\beta}^K(t,i\delta_K)\quad\text{and}\quad\widehat{N}^K_i(t)=K^{\widehat{\beta}^K_i(t/\log(K))}
\]
for all $i\in \{0,\dots ,1/\delta_K-1\}$. In addition, the processes $\widehat{A}^K_i$ and $\widehat{N}^K_i$ satisfy almost surely the relation~\eqref{def:A} for all $t\in [0,T]$, because~\eqref{def:A} involves measurable functions of the processes $\widehat{A}^K_i$ and $\widehat{N}^K_i$. We define $\widehat{\Omega}_1$ as the event where these almost-sure properties hold true and we notice that the set
\[
\Omega_0=\widehat{\Omega}_0\cap \widehat{\Omega}_1\cap \limsup_{K\to+\infty} \widehat{\Omega}_K
\]
has probability 1.

To prove that $\beta$ is a viscosity sub-solution of Equation \eqref{eq:HJ}, we work $\omega$ by $\omega$ in~$\Omega_0$.
Let $\omega\in\Omega_0$ and $T>0$ and consider a smooth function $\varphi:[0,T]\times\T$ (depending on~$\omega$) such that $\widehat{\beta}(\omega)-\varphi$ attains a strict {global} maximum on {$[0,T]\times\T$} at the point $(\bar t(\omega),\bar x(\omega))$ such that $\bar{t}(\omega)>0$. We will prove that
\[
\frac{\partial}{\partial t} \varphi(\bar t,\bar x) \leq {b(\bar x)-d(\bar x)}+ p(x) \int_{\R} G(h)e^{h\partial_x \varphi(\bar t,\bar x)} dh.
\]
Since $\widehat{A}^K(\omega)$ is continuous and converges in $L^\infty([0,T]\times\T)$ to $\beta$, there exists for $K$ large enough a local maximum of $\widehat{A}^K(\omega)-\varphi$ on $[0,T]\times\T$ at a point $(t_K(\omega),x_K(\omega))$ such that $(t_K(\omega),x_K(\omega))\to (\bar t(\omega),\bar x(\omega))$ as $K\to \infty$. Assume $K$ is large enough so that $t_K(\omega)>0$. 

From now on, we will omit the dependencies with respect to $\omega\in\Omega_0$ to avoid heavy notation.

Defining $i_K\in\{0,\ldots,1/\delta_K-1\}$ such that $i_K\delta_K\leq {x_K}<(i_K+1)\delta_K$, we have\vspace*{-3pt}
\begin{align*}
\frac{\partial}{\partial t} &\widehat A^K(t_K,x_K) =
\Bigl(1-\f {x_K}{\delta_K}+i_K\Bigr)\frac{d}{dt} \widehat A_{i_K}^K(t_K)+\Bigl(\f {x_K}{\delta_K}-i_K\Bigr) \frac{d}{dt} \widehat A_{i_K+1}^K(t_K)\\
&=\Bigl(1-\f {x_K}{\delta_K}+i_K\Bigr) \widehat N_{i_K}^K(t_K \log(K))\Big(b(i_K\delta_K) \log \big(1+\sfrac{1}{ \widehat N_{i_K}^K(t_K\log(K))} \big)\\
& \hspace{5.5cm} +d(i_K\delta_K) \log \big(1-\sfrac{1}{ \widehat N_{i_K}^K(t_K\log(K))} \big)\Big)\\
&+\Bigl(\f {x_K}{\delta_K}-i_K\Bigr) \widehat N_{i_K+1}^K(t_K\log(K))\Big(b\big((i_K+1)\delta_K\big) \log \big(1+\sfrac{1}{ \widehat N_{i_K+1}^K(t_K\log(K))} \big) \\
& \hspace{5cm} +d\big((i_K+1)\delta_K\big) \log \big(1-\sfrac{1}{ \widehat N_{i_K+1}^K(t_K\log(K))} \big)\Big)\\
&+ \Bigl(1-\f {x_K}{\delta_K}+i_K\Bigr) \sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor}h_Kp((\ell+i_K)\delta_K)G(h_K\ell) \\
& \hspace{5cm} \times \widehat N_{\ell+i_K}^K(t_K\log(K))\log \big(1+\sfrac{1}{ \widehat N_{i_K}^K(t_K\log(K))} \big)\\
&+ \Bigl(\f {x_K}{\delta_K}-i_K\Bigr) \sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor}h_Kp((\ell+i_K+1)\delta_K)G(h_K\ell)\\
& \hspace{4.2cm} \times \widehat N_{\ell+i_K+1}^K(t_K\log(K))\log \big(1+\sfrac{1}{ \widehat N_{i_K+1}^K(t_K\log(K))} \big).
\end{align*}
Using that for all $x\geq -1/2$,\vspace*{-3pt}\enlargethispage{\baselineskip}
\begin{equation*}
b\log(1+x)+d\log(1-x)\le (b-d) x,\label{approxlog2}
\end{equation*}
and using $\widehat{\beta}^K(t,x)\geq a$ for $K$ large enough (by definition of $\Omega_0$), we deduce that\vspace*{-3pt}
\begin{align*}
\frac{\partial}{\partial t} \widehat A^K&(t_K,x_K) \leq \Bigl(1-\f {x_K}{\delta_K}+i_K\Bigr) \big(b(i_K\delta_K) -d(i_K\delta_K)\big)\Bigl(1+\frac{C}{K^a}\Bigr)\\
&+\Bigl(\f {x_K}{\delta_K}-i_K\Bigr)\big(b((i_K+1)\delta_K) -d((i_K+1)\delta_K)\big)\Bigl(1+\frac{C}{K^a}\Bigr)\\
&+ \Bigl(1-\f {x_K}{\delta_K}+i_K\Bigr) \sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor}h_Kp((\ell+i_K)\delta_K)G(h_K\ell) \\[-5pt]
& \hspace{5cm} \times e^{\log(K)(\widehat \beta_{\ell+i_K}^K(t_K)-\widehat\beta_{i_K}^K(t_K))}
\Bigl(1+\frac{C}{K^a}\Bigr)
\notag\\
&+ \Bigl(\f {x_K}{\delta_K}-i_K\Bigr) \sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor}h_Kp((\ell+i_K+1)\delta_K)G(h_K\ell)\\[-5pt]
& \hspace{5cm} \times e^{\log(K)(\widehat \beta_{\ell+i_K+1}^K(t_K)-\widehat \beta_{i_K+1}^K(t_K))} \Bigl(1+\frac{C}{K^a}\Bigr).
\end{align*}
We next use the fact that $\mu$, $b$ and $d$ are $C^1$ functions in $\T$ to obtain, modifying the constant $C$ if necessary,\vspace*{-5pt}
\begin{multline}\label{eq:calcul-penible}
\frac{\partial}{\partial t} \widehat A^K(t_K,x_K) \leq \big(b(x_K)-d(x_K)+C\delta_K)\Bigl(1+\frac{C}{K^a}\Bigr)\\
+(1-\f {x_K}{\delta_K}+i_K) \sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor}h_K \bigl(p(x_K)+C\delta_K(|\ell|+1)\bigr) G(h_K\ell)  \\[-5pt]
 \times e^{\log(K)(\widehat \beta_{\ell+i_K}^K(t_K)-\widehat \beta_{i_K}^K(t_K))} \Bigl(1+\frac{C}{K^a}\Bigr)
\end{multline}
\vspace*{-1.2\baselineskip}
\begin{multline*}
+ (\f {x_K}{\delta_K}-i_K) \sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor}h_K \bigl(p(x_K)+C\delta_K(|\ell|+1)\bigr) G(h_K\ell) \\[-5pt]
\times e^{\log(K)(\widehat \beta_{\ell+i_K+1}^K(t_K)-\widehat\beta_{i_K+1}^K(t_K))} \Bigl(1+\frac{C}{K^a}\Bigr).
\end{multline*}

Since $(t_K,x_K)$ is a maximum point of $\widehat A^K-\varphi$, we deduce that
\begin{multline*}
\widehat \beta_j^K(t_K)-\widehat\beta_{i_K}^K(t_K)=\widehat{\beta}^K(t_K,j\delta_K)-\widehat{\beta}^K(t_K,x_K)-\left(\widehat\beta_{i_K}^K(t_K)-\widehat{\beta}^K(t_K,x_K)\right) \\
\leq\varphi(t_K,j\delta_K)-\varphi(t_K,x_K)+\widehat{M}^K(t_K,j\delta_K)-\widehat{M}^K(t_K,x_K)\\
-\left(\widehat\beta_{i_K}^K(t_K)-\widehat{\beta}^K(t_K,x_K)\right)
\end{multline*}
and similarly for $\widehat \beta_{j+1}^K(t_K)- \widehat\beta_{i_K+1}^K(t_K)$.
In addition,
\[
\varphi(t_K,j\delta_K)- \varphi(t_K,x_K)
\leq (j-i_K)\delta_K\partial_x \varphi(t_K,x_K)+O(|x_K-i_K\delta_K|)+O(|j-i_K|^2\delta_K^2).
\]
Therefore, since $\omega\in\limsup \widehat{\Omega}_K$, there exists a subsequence in $K$ (still denoted $K$) along which
\begin{align*}
\widehat\beta_j^K(t_K)-\widehat\beta_{i_K}^K(t_K) & \leq (j-i_K)\delta_K\partial_x \varphi(t_K,x_K)+C\left(|x_K-i_K\delta_K|+|j-i_K|^2\delta_K^2+\varepsilon_K\right) \\ & \leq (j-i_K)\delta_K\partial_x \varphi(t_K,x_K)+C (|j-i_K|^2\delta_K^2+\delta_K+\varepsilon_K)
\end{align*}
and
\[
\widehat\beta_{j+1}^K(t_K)-\widehat\beta_{i_K+1}^K(t_K)\leq (j-i_K)\delta_K\partial_x \varphi(t_K,x_K) +C(|j-i_K|^2\delta_K^2+\delta_K+\varepsilon_K).
\]
Combining these inequalities with~\eqref{eq:calcul-penible}, we obtain
\begin{multline}\label{ineq:xK}
\frac{\partial}{\partial t}\varphi(t_K,x_K) \leq\frac{\partial}{\partial t} \widehat A^K(t_K,x_K)
\leq \big(b(x_K)-d(x_K)+C\delta_K) \Bigl(1+\frac{C}{K^a} \Bigr)\\
+\sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor}h_K (p(x_K)+C\delta_K(|\ell|+1)) G(h_K\ell) \\[-5pt]
\times
e^{h_K\ell\partial_x\varphi(t_K,x_K)+C\frac{(h_K\ell)^2}{\log(K)}+o(1)} \Bigl(1+\frac{C}{K^a} \Bigl).
\end{multline}
Note that the first inequality above comes from the fact that $(t_K,x_K)\in (0,T]\times \T $ is a maximum point of $\widehat A^K-\varphi$ in $[0,T]\times \T$, with the maximum being attained possibly on $\{T\}\times \T$.
Since $p(x_K)\to p(\bar{x})$ when $K\to+\infty$, to prove the convergence of the sum in the right-hand side of \eqref{ineq:xK}, it is sufficient to study the convergence of
\[
S= \sum_{\ell=-\lfloor 1/2\delta_K\rfloor}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor} h_K G\big(h_K\ell \big)e^{h_K\ell\partial_x\varphi(t_K,x_K)+C\sfrac{(h_K\ell)^2}{\log(K)}}.
\]
Recall from Assumption \ref{hypmodele} that $G$ is continuous and that there exists $R>0$ such that $G$ is nonincreasing on $[R,+\infty)$ and nondecreasing on $(-\infty,-R]$. We first notice that
\[
S_0=\sum_{\ell=\lfloor -R/h_K \rfloor}^{\lfloor R/h_K\rfloor } h_K G\big(h_K\ell \big)e^{h_K\ell\partial_x\varphi(t_K,x_K)+C\sfrac{(h_K\ell)^2}{\log(K)}}
\]
is a Riemann sum which converges to $\int_{-R}^R G(y) e^{y\partial_x\varphi(\bar t,\bar x)}dy$. Hence we only have to deal with the remainder $S-S_0$. We detail the analysis for\vspace*{-3pt}
\[
S_+=\sum_{\ell= \lfloor R/h_K\rfloor +1}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor}h_K G\big(h_K\ell \big)e^{h_K\ell\partial_x\varphi(t_K,x_K)+C\sfrac{(h_K\ell)^2}{\log(K)}}.
\]
A similar computation applies to the lower tail.

For all $\varepsilon>0$, there exists $K_0$ such that for $K\geq K_0$, $|\partial_x\varphi(t_K,x_K)-\partial_x\varphi(\bar{t},\bar{x})| \leq \varepsilon$. Hence, setting $q=\partial_x\varphi(\bar{t},\bar{x})$ and recalling that $G$ is nonincreasing on $[R,+\infty)$, for $K$ large enough,\vspace*{-3pt}\enlargethispage{\baselineskip}
\begin{align*}
S_+ & \leq \sum_{\ell= \lfloor R/h_K \rfloor +1}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor} h_K G\big(h_K\ell \big)e^{qh_K\ell+C\frac{(h_K\ell)^2}{\log(K)}+\varepsilon h_K|\ell|} \\
& \leq \sum_{\ell=\lfloor R/h_K \rfloor +1}^{1/\delta_K-1-\lfloor 1/2\delta_K\rfloor} \int_{h_K(\ell-1)}^{h_K\ell } G\big(h_K\ell \big)e^{q h_K\ell+C\frac{(h_K\ell)^2}{\log(K)}+\varepsilon h_K|\ell|} \ dy\\
& \leq \int_{R}^{h_K/2\delta_K} G(y) e^{q y+|q|h_K+C\frac{y^2+h_K(2|y|+h_K)}{\log(K)}+\varepsilon (|y|+h_K)}\ dy \\
& \leq \int_{R}^{\log(K)} G(y) e^{q y+C\frac{y^2+2h_K|y|}{\log(K)}+\varepsilon |y|}(1+\varepsilon)\ dy.
\end{align*}
Observing that, for $|y|\leq (\log(K))^{1/3}$,\vspace*{-3pt}
\[\frac{y^2+2h_K|y|}{\log(K)}\leq (\log(K))^{-1/3}+2\delta_K(\log(K))^{1/3}\to 0\]
when $K\to+\infty$ and that, for $|y|\leq \log(K)$, $\psfrac{y^2+2h_K|y|}{\log(K)}\leq y+2h_K$, we can decompose the domain of integration as
\[[R,\log(K)]=[R,(\log(K))^{1/3})\cup[(\log(K))^{1/3},\log(K)]\]to deduce that, for $K$ large enough,\vspace*{-3pt}
\begin{multline*}
S_+ \leq (1+2\varepsilon)\int_{R}^{(\log(K))^{1/3}}G(y)e^{q y+\varepsilon|y|}dy +(1+2\epsilon)\int_{(\log(K))^{1/3}}^{\log(K)}G(y) e^{(q+C)y+\varepsilon|y|}dy \\
\leq(1+2\varepsilon)\int_{R}^{+\infty}G(y)e^{q y+\varepsilon|y|}dy+\frac{1+2\epsilon}{e^{(\log(K))^{1/3}}}\int_\mathbb{R} G(y) e^{(q+C+2)|y|}dy.
\end{multline*}
For the second inequality, we used that, for $\varepsilon<1$, $e^{\varepsilon|y|}\leq e^{2|y|-(\log (K))^{1/3}}$ for all $y\geq (\log (K))^{1/3}$. Now, by dominated convergence,\vspace*{-3pt}
\[
\int_{R}^{+\infty}G(y)e^{q y+\varepsilon|y|}dy\xrightarrow[\varepsilon\to 0]{}\int_{R}^{+\infty}G(y)e^{q y}dy.
\]
To conclude, recalling that $q=\partial_x\varphi(\bar{t},\bar{x})$, we have proved that\vspace*{-3pt}
\[
\limsup_{K\to+\infty} S\leq \int_\mathbb{R}G(y)e^{ y\partial_x\varphi(\bar{t},\bar{x})}dy.
\]
Therefore,\vspace*{-3pt}
\[
\frac{\partial}{\partial t}\varphi(\bar{t},\bar{x})\leq b(\bar x)-d(\bar x)+p(\bar x)\int_{\R} G(h) e^{h\partial_x\varphi(\bar{t},\bar{x})}dh.
\]
We conclude that $\beta$ is a viscosity sub-solution of~\eqref{eq:HJ} in $(0,T]\times \T$.

Following similar arguments, we can prove that $\beta$ is a viscosity super-solution, and hence a viscosity solution of~\eqref{eq:HJ} in $(0,T]\times \T$. The result then follows from uniqueness of a Lipschitz viscosity solution of~\eqref{eq:HJ}~\cite{GB:94}.

\backmatter
\bibliographystyle{jepplain+eid}
\bibliography{champagnat-et-al}
\end{document}

