LEARNING-BASED PREDICTIVE CONTROL: REGRET ANALYSIS
5.5 Results
Causally Invariant Safety Constraints
Motivated by the lower bound in the previous section, we now introduce a particular class of safety constraints where it is possible to have better performance. The class is defined by a form of causal invariance that is intuitive and general. Specifically, we state a condition under which the sets of subsequent feasible actions do not change too much if the measures of the sets are close.
We define the following specific sequences of actions. Let uβ€π‘ = (π’1, . . . , π’π‘) be a subsequence of optimal actions that maximizes the volume of the set of feasible actions, defined as uβ€π‘ := arg supuβUπ‘ π(S(u)). With a slight abuse of notation, given uβ€π‘, define the length-π maximizing subsequence of actions as uπ‘+1:π‘+π :=arg supuβUπ π(Sπ(uβ€π‘,u)).
Definition 5.5.1((π , πΏ, π)-causal invariance). The safety sets are(π , πΏ, π)-causally invariant if there exist constantsπΏ, π > 0such that the following holds:
1. For allπ‘ β [π] and sequences of actionsuβ€π‘ andvβ€π‘, πH(Sπ(uβ€π‘),Sπ(vβ€π‘)) β€πΏ
|π(S(uβ€π‘)) βπ(S(vβ€π‘)) | π(B)
1/( (πβπ‘)π)
(5.10) whereBdenotes the unit ball inRπΓ(πβπ‘).
2. For allπ‘ β [π]and sequences of actionsuβ€π‘, π(S(uβ€π‘))
π(S(uβ€π‘)) β€π
π(S( (uβ€π‘,uπ‘+1:π‘+π))) π(S(uβ€π‘+π))
πβπ‘ πβπ‘βπ
. (5.11)
Note that the definition of causal invariance is independent of the costs. Condition (1) is an inverse of the isodiametric inequality. It says that if the two subsequences of actionsuβ€π‘andvβ€π‘ do not affect the measure of the space of feasible sequences of actions too much, then the Hausdorff distance between the two setsSπ(uβ€π‘)and Sπ(vβ€π‘) is also small. Condition (2) states that the ratio of the measure of a space of feasible sequences of actions, given any previous actionsuβ€π‘, does not differ too much from the ratio of the measure of another space by fixing the next π actions to be those preserving the most future flexibility.
Definition 5.5.1 is general and practically applicable, as the following examples show.
Example4(Inventory constraints). Consider the following set of inventory constraints, which is a common form of constraints in energy storage and demand response problems [118, 127, 142]. The summation of the squares of theβ2norms of the actions is bounded from above byπΎ > 0, representing the limited resources of the system:
Γπ
π‘=1||π’π‘||2
2 β€ πΎ whereπ’π‘ βRπ. These inventory constraints are (π ,1,1)-causally invariant for all1 β€ π β€π.
Example5(Tracking constraints). The following form of tracking constraints are common in situations where the system has a target "optimalβ resource configuration that varies depending on the environment [116, 137, 143]. The agent seeks to track a fixed sequence of nominal actions y := (π¦1, . . . , π¦π). The constraints are Γπ
π‘=1|π’π‘ βπ¦π‘|π β€ πwith π β₯ 2whereπ’π‘, π¦π‘ βRandπ >0represents the systemβs adjusting ability. These tracking constraints are(π ,β2
π
,1)-causally invariant for all 1β€ π β€π.
Additionally, to highlight the structures that arenotcausally invariant, we return to the construction of constraints used in the proof of the lower bound in Theorem 5.5.1.
Example6(Constraints in the lower bound (Theorem 5.5.1)). Consider the following set of constraints defined in the proof of Theorem 5.5.1 for obtaining a lower bound Ξ©(π(π βπ€)) on the dynamic regret: S:={uβUπ :π’π‘ β Uπ‘,βπ‘ β [π]}where
Uπ‘ :=


ο£²

ο£³
B (π/2), if||π’π‘β1βπ£π‘β1||2 β€ π U\B (π), if||π’π‘β1βπ£π‘β1||2> π
(5.12) for someπ > 0. Suppose the action spaceU β Rπ is closed and bounded. These constraints are not causally invariant for allπ = Ξ©(π). The reason why the causal invariance criterion is violated for the safety constraints in(5.12)is that, in order to satisfy Condition (1) in(5.10), it is necessary to haveπΏ = Ξ©(π) = Ξ©(π), which is not a constant.
Motivated by the example above, and mimicking the adversarial construction in Theorem 5.5.1, we can derive the following theorem and corollary, whose proofs are in Appendix 5.B.
Theorem 5.5.2 (Lower bound on Regret). If there exist a sequence of actions vβ€π βUπ ,1β€ π β€π βπ,π β₯ π€with π = Ξ©(π)and some constantπΌ >0such that πH(Sπ(uβ€π ),Sπ(vβ€π )) β₯ πΌ π π for anyuβ€π βUπ , then for any sequence of actions u β Sgenerated by a deterministic online policy that can access the safety sets
{Uπ‘ : π‘ β [π]} and {Xπ‘ : π‘ β [π]}, Regret(u) = Ξ©(π(π βπ€)) where π(π) is a parameter that may depend onπ;π is the diameter of the action spaceU;π€is the prediction window size andπ is the total number of time slots.
The theorem above states that, for any online policy knowing the safety sets {Uπ‘ : π‘ β [π]} and {Xπ‘ : π‘ β [π]} in advance, if there are two sets of length-π subsequences of actionsSπ(uβ€π‘)andSπ(vβ€π‘)that are far from each other in terms of the Hausdorff distance, then a sub-linear regret is impossible. This highlights the necessity of the causal invariance condition. We make this explicit by further restricting the power of the online policy and assuming that it can only access the MEF ππ‘, . . . , ππ‘+πβ1at timeπ‘, which yields the following impossibility result as a corollary of Theorem 5.5.2.
Corollary 5.5.2 (Lower bound on Regret). If there exist a sequence of actions vβ€π‘ β Uπ‘,1 β€ π β€ π β π, π β₯ π€ with π = Ξ©(π), a constantπ > 0 and πΌ = Ξ©(π) such that
π(S(uβ€π )) >π π(S(uβ€π )), πH(Sπ(uβ€π ),Sπ(vβ€π )) β₯πΌ
|π(S(uβ€π )) βπ π(S(vβ€π )) | π(B)
1/( (πβπ )π)
for anyuβ€π βUπ , then for any sequence of actionsuβSgenerated by a determin- istic online policy that can access the MEF ππ‘, . . . , ππ‘+π€β1 at each time π‘ β [π], Regret(u) = Ξ©(π(π βπ€))whereπ(π) is a parameter that may depend onπ;π is the diameter of the action spaceU;π€is the prediction window size andπ is the total number of time slots.
The corollary above indicates that it is necessary to have an upper bound on the Hausdorff distance betweenSπ(uβ€π‘) andSπ(vβ€π‘) (such as Condition (1) and (2) in the causal invariance criterion) in order to make the dynamic regret sub-linear. In the next section, we show that, however, if the causal invariance criterion holds such that bothπΏandπare constants, then the dynamic regret can be made sub-linear with sufficiently many predictions, even for arbitrary Lipschitz continuous cost functions.
Bounding the Dynamic Regret of PPC
We are now ready to present our main result, which bounds the dynamic regret by a decreasing function of the prediction window size under the assumption that the safety sets are causally invariant.
Theorem 5.5.3 (Upper bound on Regret). Suppose the safety sets are (π€ , πΏ, π)- causally invariant. The dynamic regret for the sequence of actionsugiven by PPC is bounded from above by
Regret(u) =π ππ
πΏlogπ
β π€
+
β πΏ π€1/4
! !
whereπdenotes the diameter of the action spaceU,π€is the prediction window size andπ is the total number of time slots.
This theorem implies that, with additional assumptions on the safety constraints, a sub-linear(inπ) dynamic regret is achievable, given a sufficiently large prediction window sizeπ€ =π(1)(inπ). Notably, Theorem 5.5.1 implies that for the worst-case costs and constraints, a deterministic online controller that has full information of the constraints suffers from a linear regret. As a comparison, Theorem 5.5.3 shows that, under additional assumptions, even if only aggregated information is available, PPC achieves a sub-linear regret. This does not contradict to the lower bound on the dynamic regret in Theorem 5.5.1, since if there is no regulation assumptions on the safety sets, in the worst-case, the Hausdorff distanceπH(Sπ€(uβ€π‘),Sπ€(vβ€π‘)) =π(π π€).
This implies that a trivial upper bound ofπ(ππ)holds. Additionally, note that there is a trade-off betweenflexibilityandoptimalitywhen selecting the tuning parameter π½ > 0. On the one hand, if the tuning parameter is too small, the algorithm is greedy and may suffer losses in the future; on the other hand, if the tuning parameter is too large, the algorithm is penalized by the feedback and therefore is far from being optimal.
A proof is presented in Appendix 5.B. Briefly, Theorem 5.5.1 is proven by optimizing π½ in an upper bound Regret(u) = π π π πΏlogπ/π€+π πΏ
β
π€/π½+π½/π π€ , which holds for any sequence of actions ugiven by the PCC with any tuning parameter π½ > 0.
APPENDIX
5.A Explanation of Definition 5.5.1 and Two Examples
Figure 5.A.1: Graphical illustration of (5.10) in Definition 5.5.1.
Example 1: Inventory constraints.
Recall that the feasible set of actions for the inventory constraints is S=
(
uβRπΓπ :
π
βοΈ
π‘=1
||π’π‘||2
2β€ πΎ )
.
The sequence of actionsuβ€π‘ maximizing the size of the set of admissible actions is the all-zero vector. Hence,
π(S(uβ€π‘)) π(S(uβ€π‘))
πβπ‘1
= 1β Γπ‘
π=1||π’π||2
2
πΎ
!π2
=
π(S( (uβ€π‘,uπ‘+1:π‘+π))) π(S(u<π‘+π))
πβπ‘βπ1 . Therefore settingπ=1, (5.11) is satisfied.
It remains to check the bound on the Hausdorff distance. Figure 5.A.1 shows the idea behind the definition. If the two setsS(vβ€π‘) andS(uβ€π‘) are close to each other, the Hausdorff distance of the projected sub-spacesSπ(uβ€π‘) and Sπ(vβ€π‘) can also be bounded. For inventory constraints, this is indeed the case. For all 1 β€ π β€ π, sequences of actionsuβ€π‘andvβ€π‘,
πH(Sπ(uβ€π‘),Sπ(vβ€π‘)) =
π‘
βοΈ
π=1
||π’π||22β
π‘
βοΈ
π=1
||π£π||22
!1/2
(5.13)
and
π(B)(πβ2π‘)π
π‘
βοΈ
π=1
||π’π||22β
π‘
βοΈ
π=1
||π£π||22
=
π(S(vβ€π‘))(πβ2π‘)π β π(S(uβ€π‘))(πβ2π‘)π
β€ |π(S(vβ€π‘)) βπ(S(uβ€π‘)) |(πβπ‘)π2 . (5.14) Therefore settingπΏ =1, (5.13) and (5.14) imply (5.10). We validate that the inventory constraints in Example 4 are(π ,1,1)-causally invariantfor all 1 β€ π β€π.
Example 2: Tracking constraints.
Recall that the feasible set of actions for the tracking constraints is (withπ β₯2):
S=
u βRπ : ||uβy||ππ β€ π .
The sequence of actions maximizing the size of the set of admissible actions is uβ€π‘ = yβ€π‘. Similar to Example 4, according to the formula of the volume of an βπ-ball [144], we have
π(S(uβ€π‘)) π(S(uβ€π‘))
πβπ‘1
= 1β ||uβ€π‘βyβ€π‘||ππ π
!1/π
=
π(S( (uβ€π‘,uπ‘+1:π‘+π))) π(S(u<π‘+π))
πβπ‘βπ1 . Therefore settingπ=1, (5.11) is satisfied. Next,we give a bound on the Hausdorff distance. For all 1β€ π β€π, sequences of actionsuβ€π‘andvβ€π‘,
(2Ξ(1+1/π))πβπ‘ Ξ( (πβπ‘)/π+1)
πβπ‘1
πH(Sπ(uβ€π‘),Sπ(vβ€π‘))
β€2Ξ(1+1/π)
β π
π(πβπ‘)/2 Ξ( (π βπ‘)/2+1)
πβπ‘1
π‘
βοΈ
π=1
||π’πβπ¦π||πβ
π‘
βοΈ
π=1
||π£πβπ¦π||π
=2Ξ(1+1/π)
β π
π(B)πβπ‘1
π‘
βοΈ
π=1
||π’πβ π¦π||πβ
π‘
βοΈ
π=1
||π£πβπ¦π||π
β€2Ξ(1+1/π)
β π
|π(S(vβ€π‘)) βπ(S(uβ€π‘)) |π1βπ‘
whereΞ(Β·)is Eulerβs gamma function. Therefore settingπΏ= 2Ξ(1+1/β π)
π β€ β2
π for all π β₯2, (5.10) holds. The tracking constraints in Example 5 are (π ,2/β
π,1)-causally invariantfor all 1 β€ π β€ π.
Example 3: Constraints in the Proof of Theorem 5.5.1.
Denote by A1 := B (π/2) and A2 := U\B (π/2). The feasible length-π sub- sequences of actions are either in the Cartesian product of setsAπ1 :=A1Γ Β· Β· Β· ΓA1
orA2π :=A2Γ Β· Β· Β· ΓA2. If the two sequences u<π‘ andv<π‘ are in the same set, then πΏ = π = 1; otherwise, the RHS of (5.11) becomes a constant term. Assuming π‘+π β€π, the Hausdorff distance betweenSπ(u<π‘) =A1π andSπ(v<π‘) =A2π isΞ©(π).
Therefore, a non-scalar parameterπΏ = Ξ©(π)is necessary for (5.11) to hold.
5.B Proofs
Proofs of Corollary 5.5.1
Proof of Corollary 5.5.1. The explicit expression in Lemma 17 ensures that when- ever πβ
π‘(π’π‘|u<π‘) > 0, then there is always a feasible sequence of actions inS(u<π‘). Now, if the tuning parameter π½ > 0, then the optimization (5.8)-(5.9) guaran- tees that πβ
π‘(π’π‘|u<π‘) > 0 for all π‘ β [π]; otherwise, the objective value in (5.8)- (5.9) is unbounded. Corollary 4.4.1 guarantees that for any sequence of actions u = (π’1, . . . , π’π), if πβ
π‘(π’π‘|u<π‘) > 0 for all π‘ β [π], thenu β S. Therefore, the sequence of actionsugiven by the PPC is always feasible. β‘ Proof of Theorem 5.5.1
The actions space U is closed and bounded by Assumption 5. Therefore, there exists a closed ball B (π) of radius π > 0 centered atv = (π£1, . . . , π£π) β U such thatB (π) β U. Consider the following safety constraints for allπ‘ β [π]\{1}with memory size one:
Uπ‘ :=


ο£²

ο£³
B (π/2), if||π’π‘β1βπ£π‘β1||2 β€ π U\B (π), if||π’π‘β1βπ£π‘β1||2 > π
(5.15) whereB (π/2)is a closed balls around the same center asB (π)with radiusπ/2. Let the state safety setXπ‘ =Rπfor allπ‘ β [π](i.e., no constraints on states). Whenever the first actionπ’1at timeπ‘ = 1 is taken, the future actions have to stay in the ball B (π/2), or stay outsideB (π). Any deterministic policy at timeπ‘ β [π]\{1}has to take either||π’π‘βπ£π‘||2 β€ π/2 or ||π’π‘βπ£π‘||2 > π.
Consider the following Lipschitz continuous cost functions that can be chosen adversarially:
ππ‘(π’π‘) =



ο£²



ο£³
0, ifπ‘ β€ π€
||π’π‘βπ£π‘||2, if||π’π‘β1βπ£π‘β1||2> π, π‘ > π€ (π β ||π’π‘βπ£π‘||2), if||π’π‘β1βπ£π‘β1||2β€ π/2, π‘ > π€
where π :=supπ’βU||π’||2. The construction ofc= (π1, . . . , ππ)guarantees that the dynamic regret for any sequence of actionsugiven by a deterministic online policy
is bounded from below by
Regret(u) β₯(π βπ€)minn π,
π β π
2 o
.
Take π = π/2 and note that π = Ξ©(π) where π := diam(U) := sup{||π’βπ£||2 : π’, π£ β U} is the diameter of the action spaceU. Therefore the dynamic regret is bounded from below as
Regret(u)= Ξ©(π(π βπ€))
for any sequence of actionsugiven by a deterministic online policy.
Proofs of Theorem 5.5.2 and Corollary 5.5.2
Proof of Theorem 5.5.2. Suppose there exists a sequence of actionsvβ€π β Uπ , 1 β€ π β€ πβ π, π β₯ π€withπ = Ξ©(π)and some constantπΌ >0 such that
πH(Sπ(uβ€π ),Sπ(vβ€π )) β₯πΌ π π (5.16) for anyuβ€π βUπ . For any safety sets satisfying (5.16), whenever the firstπ actions uβ€π are taken, the future actions have to stay inSπ(uβ€π ). The same argument also holds for choosingvβ€π . This means that there exist two subsequencesuπ +1:π +π and vπ +1:π +π such that
||uπ +π€:π +πβvπ +π€:π +π||2β₯ ||uπ +1:π +π βvπ +1:π +π||2β ||uπ +1:π +π€β1βvπ +1:π +π||2
β₯ ||uπ +1:π +π||2β
π +π€β1
βοΈ
π‘=π +1
||π’π‘βπ£π‘||2
β₯ πΌ π πβπ(π€β1)
=π(πΌ π βπ€+1).
Then the adversary can construct Lipschitz continuous cost functions similar to the one we used in the proof of Theorem 5.5.1 so that the costsππ +π€, . . . , ππ +π satisfy Γπ +π
π‘=π +π€π(ππ‘) = Ξ©( (πβπ€+1)π) = Ξ©(π(πβπ€))(sinceπ = Ξ©(π)) for the case when ππ‘ =π£π‘ forπ < π‘ β€ π +π andππ‘ =π£π‘ forπ‘ β€ π ; butππ +π€ =Β· Β· Β·=ππ +π =0 for the case whenππ‘ =π’π‘ forπ < π‘ β€ π +πandππ‘ =π£π‘ forπ‘ β€ π and vice versa. This βswitching patternβ attack is possible, because the online agent knows nothing about the costs ππ +π€, . . . , ππ +π at time π and the adversary can design costs freely as long as they satisfy Assumption 5. The bound on the dynamic regret follows since for the optimal actionsuβ = (π’1, . . . , π’π),Γπ +π
π‘=π +π€π(π’β
π‘) = 0 but Γπ +π
π‘=π +π€π(ππ‘) = Ξ©(π(π βπ€)) for any actionsagenerated by a deterministic online policy. β‘
Proof of Corollary 5.5.2. Applying Theorem 5.5.2 and noting thatπΌ = Ξ©(π) and at each timeπ‘ the received MEF functions πβ
π‘, . . . , πβ
π‘+π€β1equivalently form a joint density function onuπ‘+1:π‘+π€, which has less information than an online policy that knows the offline safety sets. Sinceumaximizesπ(S(Β·)),
π(S(uβ€π ))
π(S(vβ€π )) β₯ π(S(uβ€π )) π(S(uβ€π )) > π , it follows that
|π(S(uβ€π )) βπ π(S(vβ€π )) | π(B)
1/( (πβπ )π)
= Ξ©(1).
Hence, the same argument in the proof of Theorem 5.5.2 follows and we obtain the desired regret lower bound.
β‘
Proof of Theorem 5.5.3
In this appendix, we prove Theorem 5.5.3 in four steps. First, in Lemma 20, we bound the deviation of the density function (feedback) evaluated at the actions selected by the PPC from the largest density function value, given the previous actions selected by the PPC. The deviation decreases with the tuning parameter π½ >0. Next, using the bound obtained from the first step, in Lemma 21 we show that the ratio of the volume of the set of feasible actions, given the previous actions selected by the PPC and the volume of the largest feasible set is bounded from above by an exponential function, that is decreasing withπ½as well. Next, we bound the partial cost difference, of the cost induced by a subsequence of feasible actions and the offline optimal cost.
Finally, combing the partial costs, we bound the total cost and this leads to an upper bound on the dynamic regret.
Step 1. Bound the feedback deviation
Recall that π‘β² := min{π‘ +π€ β1, π}. For every π‘ β I, let uπ‘:π‘β² = π’π‘, . . . , π’π‘β² be a subsequence of optimal actions that maximizes the penalty term:
uπ‘:π‘β² := arg sup
uβUπ‘β²βπ‘+1
πβ
π‘:π‘β²(u|u<π‘) = arg sup
uβUπ‘β²βπ‘+1
π(S( (u<π‘,u))).
Before proceeding to the proof of Theorem 5.5.3, we first show the following lemma, which gives a lower bound on the feedback given the sequence of actions selected by the PPC. Note that whenπ‘β1 <0, the density functions become unconditional.
Lemma 20. For anyπ‘ β I, the sequence of actionsu= (uβ€π‘β²)selected by the PPC satisfies
πβ
π‘:π‘β²(uπ‘:π‘β²|u<π‘) πβ
π‘:π‘β²(uπ‘:π‘β²|u<π‘) β₯ exp(βπΏcπ€ π/π½)
whereπΏcis the Lipschitz constant,π :=diam(U) =sup{||π’βπ£||2 :π’, π£ βU} is the diameter of the action spaceU.
Proof of Lemma 20. First, we note that the actions uπ‘:π‘β² chosen by the PPC must satisfy
π‘β²
βοΈ
π=π‘
π½ logπβ
π(π’π|u<π) βlogπβ
π(π’π|u<π)
β€
π‘β²
βοΈ
π=π‘
(ππ(π’π) βππ(π’π))
. (5.17) To see this, suppose (5.17) does not hold. Then choosing uπ‘:π‘β² gives a smaller objective value in (5.8)-(5.9), which violates the definition of PPC-generated actions uπ‘:π‘β². Using the chain rule, (5.17) becomes
logπβ
π‘:π‘β²(uπ‘:π‘β²|u<π‘) β€logπβ
π‘:π‘β²(uπ‘:π‘β²|u<π‘) + 1 π½
π‘β²
βοΈ
π=π‘
(ππ(π’π) βππ(π’π)) . Therefore, since the cost functions are Lipschitz continuous,
πβ
π‘:π‘β²(uπ‘:π‘β²|u<π‘) β€exp 1 π½
π‘β²
βοΈ
π=π‘
(ππ(π’π) βππ(π’π))
! πβ
π‘:π‘β²(uπ‘:π‘β²|u<π‘)
β€ exp
πΏc π€ π
π½
πβ
π‘:π‘β²(uπ‘:π‘β²|u<π‘).
β‘ Step 2. Bound the Lebesgue measure deviation
Based on Lemma 20, the following lemma holds. For every π‘ β I, let u<π‘β² = π’1, . . . , π’π‘β² be a subsequence of optimal actions maximizing the volume of the set of feasible actions:
u<π‘β² :=arg sup
uβUπ‘β²β1
π(S(u)).
Lemma 21. Suppose the safety sets are(π€ , πΏ, π)-causally invariant. For anyπ‘ β I, the actions selected by the PPC satisfy that
π(S(uβ€π‘β²))
π(S(u<π‘β²)) β₯ exp(βπΏcπ‘ π/π½)πββπ‘/π€β
whereπΏπ > 0is the Lipschitz constant,π:=sup{||π’βπ£||2 :π’, π£ βU}is the diameter of the action spaceU.
Proof. For any π‘ β I, since the safety sets are (π€ , πΏ, π)-causally invariant and π(S(u<π‘β²)) =π(S(uβ€π‘β²)),
π(S(uβ€π‘β²)) π(S(u<π‘β²)) β₯ 1
π
π(S(u<π‘)) π(S(u<π‘))
πβπ‘πβπ‘β²+1+1
Β· π(S(uβ€π‘β²))
π(S( (u<π‘,uπ‘:π‘β²))). (5.18) Applying the explicit expression in Lemma 17, Lemma 20 implies that
π(S(uβ€π‘β²)) π(S( (u<π‘,uπ‘:π‘β²))) =
πβ
π‘:π‘β²(uπ‘:π‘β²|u<π‘) πβ
π‘:π‘β²(uπ‘:π‘β²|u<π‘) β₯exp(βπ€ πΏcπ/π½). Therefore, applying (5.18) recursively leads to that for anyπ‘ β I,
π(S(uβ€π‘β²))
π(S(u<π‘β²)) β₯ exp(βπΏcπ‘ π/π½)πββπ‘/π€β.
β‘
Step 3. Bound the partial cost deviation
Lemma 21 implies that for the offline optimal solutionuβπ‘β²:π‘β²+π€ :=(π’β
π‘β², . . . , π’β
π‘β²+π€), π(S(uβ€π‘β²)) β₯exp(βπΏcπ‘ π/π½)πβ2π‘/π€π(S(u<π‘β²))
β₯exp(βπΏcπ‘ π/π½)πβ2π‘/π€π S uβ<π‘β² , which leads to
π S uβ<π‘β² βπ(S(uβ€π‘β²)) β€
1+exp 2π‘
π€
logπβ πΏc π‘ π
π½
π S uβ<π‘β²
β€
πΏc π‘ π
π½ + 2π‘
π€ logπ
π(S(uβ<π‘β²))
β€
πΏc π‘ π
π½ + 2π‘
π€ logπ
π(B)
π 2
(πβπ‘β²+1)π
where we have used the inequality ππ₯ β₯ 1+π₯ for allπ₯ β R. The last inequality follows from the isodiametric inequality, withBdenoting the unit ball inR(πβπ‘
β²+1)π
. Since the safety sets are(π€ , πΏ, π)-causally invariant, it follows that for anyπ‘ β I,
||uβπ‘:π‘β² β
buπ‘:π‘β²||2 β€ πH Sπ€ uβ<π‘
,Sπ€(u<π‘)
β€πΏ π
πΏc π‘ π
π½ + 2π‘
π€ logπ
(πβπ‘+11)π
for somebuπ‘:π‘β² βSπ€(u<π‘) satisfyingπβ
π‘:π‘β²(buπ‘:π‘β²|u<π‘) >0.
Next, we consider the objective for the PPC. For anyπ‘ β I, denote byπΆ(uπ‘:π‘β²) :=
Γπ‘β²
π=π‘ππ(π’π) the cost for the times slots betweenπ‘ andπ‘β²:=min{π‘+π€ , π}. Since the costs are Lipschitz continuous,
πΆ(
buπ‘:π‘β²) βπΆ(uβπ‘:π‘β²) β€
π‘β²
βοΈ
π=π‘
ππ(π’π) βππ(π’β
π)
β€
π‘β²
βοΈ
π=π‘
πΏc||π’β
πβ bπ’π||2
β€β
π€ πΏc||uβπ‘:π‘β²βbuπ‘:π‘β²||2
β€π πΏ πΏc
πΏc π‘ π
β π€ π½
+ 2π‘
β π€
logπ
(πβπ‘+1)1 π
. (5.19) Sinceuπ‘:π‘β² is a minimizer of (5.8)-(5.9), we obtain
πΆ(uπ‘:π‘β²) β
π‘β²
βοΈ
π=π‘
π½logππ(π’π|u<π) β€πΆ(buπ‘:π‘β²) β
π‘β²
βοΈ
π=π‘
π½logππ(bπ’π|u<π), which implies
πΆ(uπ‘:π‘β²) β€πΆ(buπ‘:π‘β²) +π½log πβ
π‘:π‘β²(uπ‘:π‘β²|u<π‘) πβ
π‘:π‘β²(buπ‘:π‘β²|u<π‘). (5.20) Step 4. Bound the dynamic regretRegret(u)
Proof of Theorem 5.5.3. For the total cost, it follows that πΆπ(u) =βοΈ
π‘βI
ππ‘(uπ‘:π‘β²) β€βοΈ
π‘βI
πΆ(buπ‘:π‘β²) +π½log πβ
π‘:π‘β²(uπ‘:π‘β²|u<π‘) πβ
π‘:π‘β²(buπ‘:π‘β²|u<π‘)
β€βοΈ
π‘βI
πΆ( buπ‘:π‘β²)
| {z }
:=(π)
+π½
βοΈ
π‘βI
logπ(S(u<π‘,uπ‘:π‘β²)) π S(u<π‘,
buπ‘:π‘β²)
| {z }
:=(π)
.
For (a), plugging in (5.19), the total cost for the sequencebuis bounded from above by
βοΈ
π‘βI
πΆ(
buπ‘:π‘β²) β€πΆβ
π +βοΈ
π‘βI
πΆ(
buπ‘:π‘β²) βπΆ(uβπ‘:π‘β²)
β€πΆβ
π +
2π/π€
βοΈ
π=1
π πΏ πΏc
πΏc π π
β π€ π½
+ π
β π€
logπ
(πβπ π€1+1)π
=πΆβ
π +π
π
β π€
π πΏlogπ+ π2πΏ π½
π
β π€
whereπΆβ
π :=Γπ π‘=1πβ
π‘(π’β
π‘) is the optimal total cost. Finally, (b) becomesπ(π π½/π€) since the safety sets are atomic andUis bounded by Assumption 6 and 7. Rearranging the terms,
Regret(u)=π
π π
πΏlogπ
β π€
+ π πΏ
β π€ π½
+ π½ π π€
. Settingπ½ =π
β
πΏπ€3/4immediately implies Regret(u)=π
π π
(πΏlogπ)/β π€+
β
πΏ/π€1/4 .
β‘