Systems & Control Letters 172 (2023) 105443
Contents lists available atScienceDirect
Systems & Control Letters
journal homepage:www.elsevier.com/locate/sysconle
Nonzero-sum risk-sensitive stochastic differential games: A multi-parameter eigenvalue problem approach
✩Mrinal K. Ghosh
a, K. Suresh Kumar
b, Chandan Pal
c, Somnath Pradhan
d,∗aDepartment of Mathematics, Indian Institute of Science Bangalore, Bengaluru, India
bDepartment of Mathematics, Indian Institute of Technology Bombay, Powai, Mumbai 400076, India
cDepartment of Mathematics, Indian Institute of Technology Guwahati, Guwahati 781039, India
dDepartment of Mathematics and Statistics, Queen’s University, Kingston, ON, Canada
a r t i c l e i n f o
Article history:
Received 22 June 2022
Received in revised form 22 November 2022 Accepted 20 December 2022
Available online xxxx Keywords:
Risk-sensitive cost criterion
Parametric family of Markov generators Principal eigenvalue
Nash equilibrium
Hamilton–Jacobi–Bellman equations
a b s t r a c t
We study nonzero-sum stochastic differential games with risk-sensitive ergodic cost criterion. Under certain conditions, using multi parameter eigenvalue approach, we establish the existence of a Nash equilibrium in the space of stationary Markov strategies. We achieve our results by studying the relevant systems of coupled Hamilton–Jacobi–Bellman (HJB) equations. Exploiting the stochastic representation of the principal eigenfunctions we completely characterize Nash equilibrium points in the space of stationary Markov strategies. The complete characterization of Nash equilibrium points is established under an additive structural assumption on the running cost and the drift term.
©2022 Elsevier B.V. All rights reserved.
1. Introduction
We study non zero-sum risk-sensitive stochastic differential games in a multi parameter eigenvalue problem framework. In the literature of stochastic differential games, one usually con- siders the expectation of the integral of costs ([1–3] etc.). This is the so called risk-neutral situation where the players (i.e., the decision makers or controllers) ignore the risk. If the players are risk-sensitive (i.e., risk-averse or risk-seeking), then one of the most appropriate cost criteria is the expectation of the exponen- tial of the integral of costs as it leads to certainty equivalence [4].
Since the cost criterion is the expectation of the exponential of the integral costs, it is multiplicative as opposed to the additive nature of the cost criterion in the expectation of the integral costs case. Due to this, the analysis of the risk-sensitive case is significantly different from its risk-neutral counterpart. To our knowledge, the risk-sensitive criterion was first introduced by Bellman [5]; see [6] and the references therein. Though this criterion has been studied extensively for stochastic optimal con- trol problems [7–21], the corresponding literature in the context of stochastic differential games is rather limited. Some excep- tions are [22–25]. Basar [22] proves the existence of a Nash equilibrium for finite horizon nonzero-sum risk sensitive games.
✩ This paper is dedicated to the memory of Ari Arapostathis.
∗ Corresponding author.
E-mail addresses: [email protected](M.K. Ghosh),[email protected] (K.S. Kumar),[email protected](C. Pal),[email protected](S. Pradhan).
El-Karoui and Hamadene [25] study risk-sensitive control, zero- sum and nonzero-sum game problems. They prove the existence of an optimal control, a saddle-point and a Nash equilibrium point for relevant cases. In [25], authors use Pontryagin’s mini- mum principle to characterize the optimality condition and the adjoint problem leads to some special backward stochastic dif- ferential equations. Basu and Ghosh [23] study infinite horizon risk-sensitive zero-sum stochastic differential games and estab- lish the existence of saddle points which are mini–max selectors of the associated Hamilton–Jacobi–Isaacs (HJI) equation. In a re- cent work Biswas and Saha [24] consider risk-sensitive zero-sum stochastic differential games for controlled diffusion process in Rd. Under fairly general conditions on the drift and the diffusion coefficients (e.g., the coefficients are locally Lipschitz continuous and have some global growth condition), they study the ergodic cost criterion. They completely characterize saddle point equi- libria in the space of stationary Markov strategies, under the assumption that running cost function satisfies either small cost condition or dominated by some inf-compact function.
In the framework of reflecting diffusions Ghosh and Prad- han [26] (in bounded domain), [27] (in orthant) have studied similar nonzero-sum game problem for risk-sensitive ergodic cost criterion. They studied the game problems by studying the asso- ciated system of coupled HJB equations. In the reflecting diffusion setup, the associated coupled systems are semi-linear elliptic pdes with some oblique boundary conditions. The authors used the principal eigenvalue approach to completely characterize all possible Nash equilibria in the space of stationary Markov strate- gies. Due to the presence of these nontrivial boundary conditions,
https://doi.org/10.1016/j.sysconle.2022.105443 0167-6911/©2022 Elsevier B.V. All rights reserved.
in order to establish the existence of principal eigenpair to as- sociated coupled HJB equations, the authors in [27] crucially used the fact that the drift term, diffusion matrix are uniformly bounded and the running cost function satisfies certain small cost condition. Together with the ergodic cost criterion, in [27] the authors studied the game problem for discounted cost criterion as well.
For controlled diffusion models, similar game problem under discounted cost criterion is studied in [28]. By studying the as- sociated system of coupled HJB equations, which is in this case is a coupled system of semi-linear parabolic pdes, they have established the existence of Nash equilibrium points in the class of eventually stationary Markov strategies. The uniform bound- edness assumptions on the diffusion coefficients and the running cost functions play important role in the analysis of this game problem.
In this paper, we address the existence of Nash equilibria for stochastic differential games where the state of the system is governed by a controlled diffusion processes in the whole space Rd. We consider the risk-sensitive ergodic cost evaluation crite- rion. We analyze this game problem by analyzing the associated system of coupled HJB equation, which is a system of coupled semi-linear elliptic pdes in Rd. Compared to [26–28], under a relatively weaker set of assumptions on diffusion coefficients (e.g., the drift term and diffusion matrix are locally Lipschitz continuous and have some global growth condition) (seeAssump- tion 1), using principal eigenvalue approach we establish the existence of a Nash equilibrium in the space of stationary Markov strategies. Also, in this present study, we are allowing our running cost function to be unbounded as well (seeAssumption 2(ii)).
In order to establish the existence of principal eigenpair of the associated coupled system of Hamilton–Jacobi–Bellman (HJB) equation, we first study the corresponding Dirichlet eigenvalue problem on smooth bounded domains inRd. Applying a version of non-linear Krein–Rutman theorem we show that principal eigenpair exists for Dirichlet eigenvalue problem. Then increas- ing these domains toRdand employing Fan’s fixed point theo- rem [29], we establish the existence of principal eigenpair to the associated coupled system of HJB equation in the whole spaceRd, which lead to the existence of a Nash equilibrium. Furthermore, exploiting the stochastic representation of the principal eigen- functions we completely characterize all possible Nash equilibria in the space of stationary Markov strategies. Thus, the main results of this article can be roughly described as follows.
•
Existence and uniqueness of solution to the coupled HJB equa- tion:Using Principal eigenvalue approach, we establish the existence and uniqueness of solution to the associated cou- pled HJB equation in an appropriate function space.•
Characterization of Nash equilibrium:Using Fan’s fixed point theorem we first establish the existence of Nash equilibrium in the space of stationary Markov strategies. Then utilizing the stochastic representation of the principal eigenfunctions we completely characterize all possible Nash equilibria in the space of stationary Markov strategies.The rest of this paper is organized as follows. Section2deals with the problem description. In Section 3 we discuss the principal eigenvalue problem for controlled diffusion operators on smooth bounded domains. Section4is devoted to study the eigenvalue problem for controlled diffusion operator in whole spaceRd. The complete characterization of Nash equilibrium in the space of stationary Markov strategies is presented in Section5.
2. Problem description
For the sake of notational simplicity we treat two player case.
Let Ui
,
i=
1,
2 be compact metric spaces and Vi=
P(Ui), the space of probability measures on the compact metric space Ui with the topology of weak convergence. Letb¯ =
(b¯
1, . . . ,
b¯
d):
Rd×
U1×
U2→
Rd,¯
ri:
Rd×
U1×
U2→ [
0, ∞
),
i=
1,
2,σ :
Rd→
Rd×dbe given functions satisfyingAssumption 1(to be described below).Defineb
=
(b1, . . . ,
bd):
Rd×
V1×
V2→
Rd,
ri:
Rd×
V1×
V2→ [
0, ∞
) bybk(x
, v
1, v
2)=
∫
U2
∫
U1
b
¯
k(x,
u1,
u2)v
1(du1)v
2(du2),
ri(x, v
1, v
2)=
∫
U2
∫
U1
r
¯
i(x,
u1,
u2)v
1(du1)v
2(du2),
x∈
Rd, v
1∈
V1, v
2∈
V2,
k=
1, . . . ,
d,
i=
1,
2.
We consider a nonzero-sum stochastic differential game whose state is evolving according to a controlled diffusion process given by the solution of the following stochastic differential equation (s.d.e.)dX(t)
=
b(X(t), v
1(t), v
2(t))dt+ σ
(X(t))dW(t),
(2.1) where W(·
) is anRd-valued standard Wiener process,v
i(·
) is a Vi-valued process which is a non-anticipative functional of the state processX(·
), i.e.,v
i(t)=
fi(t,
X([
0,
t]
)) whereX([
0,
t]
)(s)=
X(s∧
t) for alls∈ [
0, ∞
) andfi: [
0, ∞
)×
C([
0, ∞
);
Rd)→
Vi. Such a strategy is called an admissible strategy. Fori=
1,
2,Aidenotes the space of all admissible strategies of Playeri. In order to ensure the existence of a solution to Eq.(2.1)and the existence of Nash equilibrium (to be describe in(2.6)), we impose following conditions on the drift termb, and the dispersion matrix
¯ σ
. Assumption 1.(i) Local Lipschitz continuity: The function
σ =
[σ
ij]:
Rd→
Rd×d and b¯ :
Rd×
U1×
U2→
Rd are locally Lipschitz continuous inx(uniformly with respect to the rest), i.e., for each R≥
0, there exists a constantCR>
0 depending on R>
0, such that|¯
b(x,
u1,
u2)− ¯
b(y,
u1,
u2)|
2+ ∥ σ
(x)− σ
(y)∥
2≤
CR|
x−
y|
2 for all x,
y∈
BR (:= {
x∈
Rd: |
x| <
R}
), i=
1,
2 and (u1,
u2)∈
U1×
U2, where∥ σ ∥ :=
√tr(
σ σ
T) and b¯ =
(b¯
1
, . . . ,
b¯
d)T. Also, we assume that b
,
ri are jointly continuous in (x,
u1,
u2) fori=
1,
2.(ii)Affine growth condition: b
¯
andσ
satisfy a global growth condition of the formsup
u1∈U1,u2∈U2
⟨¯
b(x,
u1,
u2),
x⟩
++∥ σ
(x)∥
2≤
C0( 1+|
x|
2)∀
x∈
Rd,
for some constantC0>
0.(iii) Nondegeneracy:For eachR
>
0, it holds thatd
∑
i,j=1
aij(x)zizj
≥
CR−1|
z|
2∀
x∈
BR,
and for allz
=
(z1, . . . ,
zd)T∈
Rd, wherea= [
aij] :=
12
σ σ
T. Also, we assume that the running cost functionsr¯
i:
Rd×
U1×
U2→
R+i=
1,
2 are jointly continuous in (x,
u1,
u2) and locally Lipschitz continuous in x (uniformly with respect to the rest), i.e., for allR≥
0 andx,
y∈
BR there exists a constant CR>
0 depending onR>
0, such that|¯ri(x,u1,u2)− ¯ri(y,u1,u2)|2 ≤ CR|x−y|2 for all (u1,u2)∈U1×U2.
2
M.K. Ghosh, K.S. Kumar, C. Pal et al. Systems & Control Letters 172 (2023) 105443
It is well known that, underAssumption 1, for any (
v
1, v
2)∈
A1
×
A2 and initial conditionX(0)=
x, the s.d.e.(2.1)admits a unique weak solution which is a strong Markov process (see [30, Theorem 2.2.11, p.42]). For the stochastic differential game, the controlled diffusion given by (2.1) has the following inter- pretation: The ith player controls the state dynamics, i.e., the controlled diffusion given above, through the choice of her/his strategyv
i. The functionr¯
irepresents the running cost function of Player i. If the strategyv
ihas the formv
i(t)= ¯ v
i(t,
X(t)),
t≥
0 for somev ¯
i: [
0, ∞
)×
Rd→
Vi, thenv
ior by an abuse of notationv ¯
iis called a Markov strategy for Playeri. LetMi
= { v
i: [
0, ∞
)×
Rd→
Vi| v
iis measurable}
be the set of all Markov strategies for Playeri. Under a pair of Markov strategies the s.d.e.(2.1)admits a unique strong solution which is a strong Markov process (see [30, Theorem 2.2.12, p.45]). Ifv
i does not have explicit depen- dence ont, i.e.,v ¯
i(t,
x)= ¯ v
i(x),
x∈
Rd,
t≥
0, it is said to be a stationary Markov strategy for Playeri. The set of all stationary Markov strategies for Player iis denoted by Si,
i=
1,
2. We topologize Si,
i=
1,
2, using a metrizable weak* topology on L∞(Rd;
Ms(Ui)), where Ms(Ui) denotes the space of all signed measures on Ui with weak* topology. Since Si is a subset of the unit ball of L∞(Rd;
Ms(Ui)), it is compact under the above weak* topology. One also has the following characterization of the topology given by the following convergence criterion: For i=
1,
2,v
in→ v
i inSiasn→ ∞
if and only iflim
n→∞
∫
Rd
f(x)
∫
Ui
g(x,ui)vin(x)(dui)dx =
∫
Rd
f(x)
∫
Ui
g(x,ui)vi(x)(dui)dx, (2.2) for allf
∈
L1(Rd)∩
L2(Rd),
g∈
Cb(Rd×
Ui); see [30, p. 57], for details.For
v
i∈
Vi,
i=
1,
2, letLv1,v2:
C2(Rd)→
C(Rd), be given by Lv1,v2f(x)=
aij(x)∂
2f(x)∂
xi∂
xj+
bi(x, v
1, v
2)∂
f(x)∂
xi,
f∈
C2(Rd),
(2.3) where Einstein summation convention is used. Further, let G1v2f=
infv1∈V1
[
Lv11,v2f+
r1(x, v
1, v
2(x))f] , v
2∈
S2,
(2.4)G2v1f
=
infv2∈V2
[
Lv21,v2f+
r2(x, v
1(x), v
2)f] , v
1∈
S1,
f∈
C2(Rd),
where forf∈
C2(Rd),Lv11,v2f(x)
=
Lv1,v2(x)f(x)∀ v
1∈
V1, v
2∈
S2 andLv21,v2f(x)
=
Lv1(x),v2f(x)∀ v
1∈
S1, v
2∈
V2.
For (v
1, v
2)∈
S1×
S2, it is easy to see thatLv11(x),v2f(x)
=
Lv21,v2(x)f(x)=
Lv1(x),v2(x)f(x)=
aij(x)∂
2f(x)∂
xi∂
xj+
bi(x, v
1(x), v
2(x))∂
f(x)∂
xi.
The analysis of our game problem will be based on the analysis of the eigenvalue problems of the above defined operators.
2.1. Ergodic cost criterion
Given the running cost functionsri
:
Rd×
V1×
V2→
R+,
i=
1,
2, for any (v
1, v
2)∈
A1×
A2, the associated risk-sensitive ergodic cost of Playeriis defined byρ
i(x, v
1, v
2)=
lim supT→∞
1
T logEvx1,v2
[ e
∫T
0 ri(X(t),v1(t),v2(t))dt]
,
i=
1,
2.
(2.5)The definition of a Nash equilibrium is standard, i.e., (
v
1∗, v
2∗)∈
A1
×
A2 is a Nash equilibrium among the class of admissible strategies ifρ
1(x, v
∗1, v
2∗)≤ ρ
1(x, v
1, v
∗2),
for allv
1∈
A1,
(2.6)ρ
2(x, v
∗1, v
2∗)≤ ρ
2(x, v
1∗, v
2),
for allv
2∈
A2,
for allx∈
Rd.
We assume that our running cost functions ri,i=
1,
2 satisfy Assumption 1(i). Now for each (v
1, v
2)∈
A1×
A2, defineλ
1(x, v
2)=
infv′1∈A1
ρ
1(x, v
1′, v
2), λ
1(v
2)=
infx∈Rd
λ
1(x, v
2),
(2.7) Λ1(x, v
2)=
infv′1∈S1
ρ
1(x, v
1′, v
2),
Λ1(v
2)=
infx∈Rd
Λ1(x
, v
2), λ
2(x, v
1)=
infv′2∈A2
ρ
2(x, v
1, v
′2), λ
2(v
1)=
infx∈Rd
λ
2(x, v
1),
Λ2(x, v
1)=
infv′2∈S2
ρ
2(x, v
1, v
′2),
Λ2(v
1)=
infx∈Rd
Λ2(x
, v
1).
Now we outline our program for establishing the existence of a Nash equilibrium. We analyze our game problem by analyzing the corresponding system of coupled Hamilton–Jacobi–Bellman (HJB) equations. Suppose that one of the players, say Player 2 announces his strategyv
2∈
S2in advance, then Player 1 tries to minimize associated costρ
1(x, v
1, v
2) (see, Eq.(2.5)) over allv
1∈
A1, which is a (stochastic) optimal control problem for Player 1.
Such an optimal control problem has been studied in [7,13,14]
and it is shown that one can characterize the optimal value and optimal controls by analyzing the corresponding HJB equation given by
λ
1ψ
1(x)=
Gv12ψ
1(x) withψ
1(0)=
1.
(2.8) It is well known that (see [7]) the principal eigenvalue of the HJB equation is the optimal valueλ
1(v
2) and any minimizing selector of (2.8) (which is same as the minimizing selector of (2.4)), i.e., anyv
1∗∈
S1which satisfiesλ
1(v
2)ψ
1(x)=
G1v2ψ
1(x)=
Lv∗ 1,v2
1
ψ
1+
r1(x, v
∗1(x), v
2(x))ψ
1,
is an optimal control for Player 1. In particular,v
1∗∈
S1 is an optimal response for Player 1 corresponding to the announced strategyv
2of Player 2. Note thatv
1∗depends onv
2and the mapv
2(∈
S2)→
the optimal responses of Player 1may be multi-valued. Analogous result holds for Player 2 if Player 1 announces his strategy
v
1∈
S1 in advance. From the above discussion, it is easy to see that for any given pair of strategies (v
1, v
2)∈
S1×
S2, one can construct a set of pairs of optimal responses{
(v
∗1, v
2∗)∈
S1×
S2}
from their corresponding HJB equations. Clearly any fixed point of this multi-valued map is a Nash equilibrium. The above discussion leads to the following program for finding a pair of Nash equilibrium strategies for ergodic cost criterion. Suppose that there exist a pair of stationary strategies (v
1∗, v
2∗)∈
S1×
S2, a pair of scalars (λ
1, λ
2) and a pair of functions (ψ
1, ψ
2) in an appropriate function space satisfying the following coupled HJB equationsλ
1ψ
1=
Gv∗ 2 1
ψ
1=
Lv∗ 1,v∗2
1
ψ
1+
r1(x, v
1∗(x), v
∗2(x))ψ
1λ
2ψ
2=
Gv∗ 1 2
ψ
2=
Lv∗ 1,v∗2
2
ψ
2+
r2(x, v
1∗(x), v
∗2(x))ψ
2,
then (
v
1∗, v
2∗) will be a pair of Nash equilibrium. The above discus- sion leads us to study the principal eigenvalues associated with the above coupled equations in the subsequent sections.3. Dirichlet eigenvalue problem for controlled diffusion oper- ators
In this section, we discuss the principal eigenvalue problem associated with the nonlinear operatorsGivj on smooth bounded
3
domains D
⊂
Rd. The generalized principal eigenvalue of the semi-linear operatorGvij with Dirichlet boundary condition onD is defined byλ
+i (v
j,
D)=
inf{ λ ∈
R|
for someϕ ∈
W2,p(D)∩
C(D)¯ ,
p>
d, ϕ >
0,
Givjϕ ≤ λϕ
inD} ,
(3.1) fori̸=
j,i,
j=
1,
2. Now we prove the existence of the principal eigenvalues of a certain parametric family of semi-linear elliptic pdes.Theorem 3.1. Suppose thatAssumption1holds. Let
v
j∈
Sjand D be a bounded smooth domain inRd. Then there exists (unique up to a scalar multiplication)ψ
D∈
W2,p(D)∩
C(D)¯ ,
p>
d, ψ
D>
0such thatGivj
ψ
D= λ
+i (v
j,
D)ψ
D,
(3.2)ψ
D=
0 on∂
D,
i,
j=
1,
2 with i̸=
j.
Proof. We takei
=
1,
j=
2. Supposer1≤
0 (this will be dropped shortly). Forφ ∈
C01(D)(:=
C0(D)¯ ∩
C1(D)),f∈
Lp(D), letΓ1(φ,f)(x)= − inf v1∈V1
{bi(x, v1, v2(x))∂φ(x)
∂xi
+r1(x, v1, v2(x))φ(x)}+f(x),
and consider aij(x)
∂
2φ ˆ
(x)∂
xi∂
xj=
Γ1(φ,
f)(x),
withφ ˆ =
0 on∂
D.
(3.3) Then by [31, Theorem 9.15, p.241], [31, Theorem 9.14, p.240], there exists a unique solutionφ ˆ ∈
W2,p(D)∩
C(D),¯
p>
d, satisfying∥ ˆ φ ∥
W2,p(D)≤ κ
1(∥ ˆ φ ∥
∞+ ∥
Γ1(φ,
f)∥
Lp(D)),
(3.4) for some positive constantκ
1= κ
1(p,
D) which is independent ofφ ˆ
,φ
,f. From [31, Theorem 9.1, p.220], we deduce that∥ ˆ φ ∥
∞≤ κ
2∥
Γ1(φ,
f)∥
Ld(D),
for some constant
κ
2>
0. Hence, from(3.4), we obtain∥ ˆ φ ∥
W2,p(D)≤ κ
3∥
Γ1(φ,
f)∥
Lp(D) (3.5) for some positive constantκ
3. Now consider an operatorTmap- pingφ ∈
C01(D) to the corresponding solutionφ ˆ
of (3.3), i.e., T(φ
)= ˆ φ
. Since the embeddingW2,p(D)↪ →
C1,α(D) forp>
d andα ∈
(0,
1−
dp) is compact, the operator Tis compact and continuous. Now we want to show that the following space of functions{ φ ∈
C01(D): φ = ν
T(φ
) for someν ∈ [
0,
1]} ,
is bounded inC01(D). Suppose that there exists a sequence (
φ
n, ν
n) with∥ φ
n∥
C10(D)
→ ∞
andν
n→ ν ∈ [
0,
1]
asn→ ∞
. Scalingφ
nappropriately we assume that
∥ φ
n∥
C10(D)
=
1. Hence, in view of the estimate(3.5), extracting a suitable subsequence, there exists a nontrivialφ ˜
satisfyingaij(x)∂2φ˜(x)
∂xi∂xj
= −ν inf
v1∈V1
{bi(x, v1, v2(x))∂φ˜(x)
∂xi
+r1(x, v1, v2(x))φ˜(x)}, with
φ ˜ =
0 on∂
D. This is a contradiction to the strong maximum principle [31, Theorem 9.6, p.225]. This implies that the above space is bounded. Hence, by the Leray–Schauder fixed point theorem [31, Theorem 11.3, p.280], it follows that Tadmits a fixed pointϕ ∈
W2,p(D)∩
C(D) , i.e., we have¯
G1v2
ϕ
(x)=
f(x),
withϕ =
0 on∂
D.
Also, by the strong maximum principle [31, Theorem 9.6] it is clear that
ϕ
satisfying the equation is unique.LetX
=
C0(D) andCthe cone of non-negative functions inX.Now define an operatorT
ˆ
which maps f∈
Xto corresponding solutionϕ ∈
W2,p(D)∩
C(D) satisfying¯
G1v2
ϕ
(x)= −
f(x),
withϕ =
0 on∂
D.
From the above discussion it is easy to see that the opera- tor T
ˆ
is well defined. Thus, combining [31, Theorem 9.1] and [31, Theorem 9.14], we deduce that∥ ϕ ∥
W2,p(D)≤ κ
1supD
| ϕ | ,
(3.6)for some positive constant
κ
1. From (3.6), it is clear that Tˆ
is compact and continuous. Also, from the definition one can see that Tˆ
is 1-homogeneous (i.e.,T(ˆ λ ˜
f)= ˜ λ
T(fˆ
) for allλ ˜ ≥
0).Suppose T(f
ˆ
k)= ϕ
k, k=
1,
2, with f1≤
f2. Thus, we have G1v2ϕ
1(x)≥
Gv12ϕ
2(x). SinceG1v2 is concave, it follows thatG1v2(ϕ
2− ϕ
1)(x)≤
0. Hence, applying [32, Theorem 3.1] we obtainϕ
2≥ ϕ
1and iff1
<
f2(i.e.,f1≤
f2 andf1̸=
f2) then we haveϕ
2> ϕ
1(see [32, Lemma 3.1]). This implies thatT
ˆ
is order preserving. Letφ ˜ ∈
Cbe nontrivial nonnegative function with compact support, hence from the above discussion we deduce thatT(ˆ φ ˜
)>
0. Thus, one can chooseκ
2>
0 such thatκ
2T(ˆ φ ˜
)− ˜ φ >
0 inD. Therefore, by Krein–Rutman theorem (seeTheorem A.1), we conclude that there exists (λ, ψ ˆ
D)∈
R+×
W2,p(D)∩
C(D) with¯ ψ
D>
0 satisfying G1v2ψ
D= ˆ λψ
D in D,
andψ
D=
0 on∂
D.
(3.7) Whereψ
D is unique up to scalar multiplication. Now, r1≥
0 (which is the case by our assumption), since r1 is bounded in D¯
replacingr1by (r1− ∥
r1∥
∞,D), following the above arguments there exists (λ
D, ψ
D)∈
R×
W2,p(D)∩
C(D) with¯ ψ
D>
0 satisfying (3.7).Next, we show that
λ
D= λ
+1(v
2,
D).
Clearly,λ
D≥ λ
+1(v
2,
D).
(3.8)Suppose
λ
+1(v
2,
D)< λ
D. Then for eachε >
0, there existsε
′≤ ε
andϕ
′∈
W2,p(D)∩
C(D)¯ , ϕ
′>
0 such thatG1v2
ϕ
′≤
(λ
+1(v
2,
D)+ ε
′)ϕ
′.
(3.9) Chooseε
′>
0 small enough such thatλ
+1(v
2,
D)+ ε
′< λ
D. Also, we haveG1v2
ψ
D−
(λ
+1(v
2,
D)+ ε
′)ψ
D>
G1v2ψ
D− λ
Dψ
D=
0.
(3.10) Hence byTheorem A.3, it follows thatψ
D=
tϕ
′for somet>
0.This gives a contradiction. Therefore we get
λ
D= λ
+1(v
2,
D). This completes the proof. □4. Eigenvalue problem for controlled diffusion operators inRd In this section we explore the existence of principal eigenvalue of the controlled diffusion operatorGivj
, v
j∈
Ajin the whole space Rd and establish their relations with the risk-sensitive ergodic optimal control problem. The generalized principal eigenvalue of Givj in the whole space is defined byλ
+i (v
j)=
inf{ λ ∈
R|
for someϕ ∈
Wloc2,p(Rd)∩
C(Rd),
p>
d, ϕ >
0,
Givjϕ ≤ λϕ
a.
e. } .
(4.1)In order to study our game problem we enforce following Foster–Lyapunov condition on the dynamics.
4
M.K. Ghosh, K.S. Kumar, C. Pal et al. Systems & Control Letters 172 (2023) 105443
Assumption 2.
(i) In bounded cost case:There existV
∈
C2(Rd) with infRdV≥
1, constantsδ, α > ˜
0 and a compact setK˜
such thatsup
ui∈Ui,i=1,2
Lu1,u2V
≤ ˜ α
IK˜− δ
V.
(4.2) and maxi=1,2∥
ri∥
∞< δ
.Or,
(ii) In unbounded cost case: There exist V
∈
C2(Rd) with infRdV≥
1, an inf-compact positiveℓ ∈
C(Rd) (i.e., the sublevel sets{ ℓ ≤ κ }
are compact, or empty, inRdfor eachκ ∈
R), a constantα > ˜
0 and a compact setK˜
such thatsup
ui∈Ui,i=1,2
Lu1,u2V
≤ ˜ α
IK˜− ℓ
V,
(4.3) and fori=
1,
2ℓ
(x)−
supui∈Ui,i=1,2
r
¯
i(x,
u1,
u2) is inf-compact.
(4.4) As noted in [7,9], if a and b are bounded, it might not be possible to find an unbounded functionℓ
which satisfies (4.3).In view of this, we are assuming(4.2).
Fori
̸=
j, it is easy to see that underAssumption 2(i)vsup1∈A1 vsup2∈A2
lim sup
T→∞
1
T logEvx1,v2
[ e
∫T
0ri(X(t),v1(t),v2(t))dt]
≤ ∥ri∥∞<∞. Also, underAssumption 2(ii), applying Itô–Krylov formula, it fol- lows that
vsup1∈A1 vsup2∈A2
lim sup
T→∞
1
T logEvx1,v2
[ e
∫T 0ℓ(X(t))dt]
≤ α ˜
minK˜V.
From(4.4), it is clear that supuk∈Uk,k=1,2
¯
ri(· ,
u1,
u2)≤ κ
1+ ℓ
(·
), for some positive constantκ
1. Therefore, we obtainsup v1∈A1
sup v2∈A2
lim sup
T→∞
1
TlogEvx1,v2
[
e∫0Tri(X(t),v1(t),v2(t))dt]
≤κ1+ α˜ minK˜V.
(4.5) Now we proceed to prove the existence of the principal eigenpair to certain semi-linear elliptic pdes in the whole spaceRd. Theorem 4.1. LetAssumptions1and2hold. Suppose
v
j∈
Sj, then there exists a uniqueψ ∈
Wloc2,p(Rd)∩
C(Rd),
p>
d, ψ >
0such thatGivj
ψ = λ
+i (v
j)ψ
withψ
(0)=
1.
(4.6) Moreoverλ
+i (v
j)is simple and satisfiesλ
+i (v
j)≤ λ
i(v
j),
for i̸=
j,
i,
j=
1,
2.
(4.7) Proof. Takei=
1,
j=
2. LetD=
Bn,
n≥
1, denote the open ball centered at the origin with radiusn. FromTheorem 3.1, there exists a (unique)ψ
n∈
W2,p(Bn)∩
C(B¯
n), ψ
n>
0 inBn withψ
n(0)=
1 satisfyingG1v2
ψ
n= λ
nψ
nψ
n=
0 on∂
Bn,
(4.8)where
λ
n= λ
+1(v
2,
Bn). Choosev
1∈
A1, sinceψ
n=
0 on∂
Bnapplying Ito–Dynkin’s formula we obtain
ψ
n(x)≤
Evx1,v2[ e
∫T
0(r1(X(t),v1(t),v2(X(t)))−λn)dt
ψ
n(X(T))I{T≤τ}]
≤ ∥ ψ
n∥
∞,BnEvx1,v2[
e∫0T(r1(X(t),v1(t),v2(X(t)))−λn)dt ]
for all (T
,
x)∈
R+×
Bn,
where
τ
is the first exit time of the process X(t) fromBn and∥ ψ
n∥
∞,Bn:=
supx∈Bnψ
n(x). Thus, taking logarithm on both sides of the inequality, dividing byT and lettingT→ ∞
, it follows thatλ
n≤
lim supT→∞
1
T logEvx1,v2
[
e∫0Tr1(X(t),v1(t),v2(X(t)))dt]
< ∞ .
(4.9) Sinceλ
nis nondecreasing inn(see,(3.1)), it follows that limnλ
n= λ
exists.Now using Harnack inequality (see [31, Corollary 8.21, p.199]) and the interior estimates [31, Theorem 9.11, p.235], we get for each bounded domainD, there existsn0such that
sup
n≥n0
∥ ψ
n∥
W2,p(D)< ∞ .
(4.10)Hence, by a standard diagonalization procedure and Banach–
Alaoglu theorem, we can extract a subsequence
{ ψ
nk}
such that for someψ ∈
Wloc2,p(Rd)∩
C(Rd),
p≥
2{ψnk → ψ in Wloc2,p(Rd) (weakly)
ψnk → ψ in C1,α(K) (strongly) for all compact setK⊂Rd, (4.11) where 0
< α <
1−
dp. Now multiplying both sides of(4.8)by a test function
ϕ ∈
Cc∞(Rd), integrating, and then lettingn→ ∞
, we deduce thatψ ∈
Wloc2,p(Rd)∩
C(Rd),
p≥
2 satisfiesG1v2
ψ = λψ
inRd.
(4.12)From(4.9), it follows that
λ ≤ λ
1(v
2).
Since for eachn
∈
Nwe haveψ
n>
0 it is clear thatψ ≥
0 inRd and sinceψ
n(0)=
1 for alln, we haveψ
(0)=
1. Thus, applying Harnack’s inequality we deduce thatψ >
0 inRd.Next from the definition of the generalized principal eigen- value, it is immediate that
λ ≥ λ
+1(v
2).
(4.13)Also from the definition of the generalized principle eigenvalue (see Eq.(3.1)), it follows that
λ
n= λ
+1(v
2,
Bn)≤ λ
+1(v
2).
(4.14) Thus, combining(4.13)and(4.14)we getλ = λ
+1(v
2).
Next we show that any eigenvalue of G1v2 corresponding to a positive eigenfunction in the class Wloc2,p(Rd)
∩
C(Rd) is simple.This, in particular, would impliy the simplicity of the generalized principal eigenvalue
λ
+1(v
2).Let
ψ
k∈
Wloc2,p(Rd)∩
C(Rd),
k=
1,
2 be positive eigenfunctions corresponding to an eigenvalueλ
(in particular, we are interested inλ = λ
+1(v
2)) satisfyingψ
k(0)=
1. Let t0>
0 be such thatψ
1−
t0ψ
2≥
0 inB¯
R.
Let
v
1be a minimizing selector ofG1v2ψ
1. Thus Lv11,v2ψ
1+
r1(x, v
1(x), v
2(x))ψ
1=
Gv12ψ
1= λψ
1Lv11,v2
ψ
2+
r1(x, v
1(x), v
2(x))ψ
2≥
Gv12ψ
2= λψ
2.
This gives us the following inequalityLv11,v2(
ψ
1−
t0ψ
2)+
r1(x, v
1(x), v
2(x))(ψ
1−
t0ψ
2)≤ λ
(ψ
1−
t0ψ
2).
Sinceψ
1−
t0ψ
2≥
0 inB¯
R, it follows that
Lv11,v2(