LEARNING-BASED PREDICTIVE CONTROL: FORMULATION
4.7 Application
are strictly convex and differentiable, which is common in practice. Further, let ๐(ยท)denote the Lebesgue measure. Note that the set of subsequent feasible action trajectories S(๐ขโค๐ก) is Borel-measurable for all ๐ก โ [๐], implied by the proof of Corollary 16. We also assume that the measure of the set of feasible actions ๐(S(๐ขโค๐ก))is differentiable and strictly logarithmically-concave with respect to the subsequence of actions๐ข๐ก = (๐ข1, . . . , ๐ข๐ก) for all๐ก โ [๐], which is also common in practice, e.g., it holds in the case of inventory constraintsร๐
๐ก=1โฅ๐ข๐กโฅ2 โค ๐ต with a budget๐ต > 0. Finally, recall the definition of the set of subsequent feasible action trajectories:
S(๐ขโค๐ก) :=n
๐ฃ>๐ก โU๐โ๐ก :๐ฃsatisfies (4.2b)โ(4.2d), ๐ฃโค๐ก =๐ขโค๐ก o
. Putting the above together, we can state our assumption formally as follows.
Assumption 4. The cost functions๐๐ก(๐ข) : R โ R+ are differentiable and strictly convex. The mappings ๐(S(๐ขโค๐ก)) : R๐ก โ R+ are differentiable and strictly logarithmically-concave.
Given regularity of the cost functions and time-varying constraints, we can prove optimality of PPC.
Theorem 4.6.1(Existence of optimal actions). LetU =R. Under Assumption 3 and 4, there exists a sequence๐1, . . . , ๐๐ such that implementing(4.13)with๐ฝ๐ก = ๐๐ก and ๐๐ก = ๐โ
๐ก at each time๐ก โ [๐] ensures ๐ข = (๐ขโ
1, . . . , ๐ขโ
๐), i.e., the generated actions are optimal.
Crucially, Theorem 4.6.1 shows that there exists a sequence of โgoodโ tuning parameters so that the PPC scheme is able to generate optimal actions under reasonable assumptions. However, note that the assumption ofU=Ris fundamental.
When the action spaceUis discrete orUis a high-dimensional space, it is impossible to generate the optimal actions because, in general, fixing๐ก, the differential equations in the proof of Theorem 4.6.1 (see Appendix 4.F) do not have the same solution for all๐ฝ๐ก > 0. Therefore a detailed regret analysis is necessary in such cases, which is a challenging task for future work.
Experimental Setups
In the following, we present settings of parameters and useful metrics in our experiments.
Dataset and hardware. We use real EV charging data from ACN-Data [1], which is a dataset collected from adaptive EV charging networks (ACNs) at Caltech and JPL.
The detailed hardware setup for that EV charging network structure can be found in [101].
Error metrics. Recall the EV charging example in optimization (4.5a)-(4.5b). We first introduce two error metrics to measure the EV charging constraint violations.
Note that the constraints (4.4a), (4.4b) and (4.4e) are hard constraints depending only on the scheduling policy, but not the actions and energy demands. Therefore they can be automatically satisfied in our experiments by fixing a scheduling policy satisfying them such as least laxity first. Violations may happen on constraint (4.4c) and (4.4d). To measure the violation of (4.4c), we use the (normalized) mean squared error (MSE) as the tracking error:
MSE:=
๐ฟ
โ๏ธ
๐=1 ๐
โ๏ธ
๐ก=1
๐
โ๏ธ
๐=1
๐ (
๐)
๐ก (๐) โ๐ข(
๐) ๐ก
2
/(๐ฟร๐ ร๐), (4.14)
where๐ข(
๐)
๐ก is the๐ก-th power signal for the๐-th test and๐ (
๐)
๐ก (๐)is the energy scheduled to the ๐-th charging session at time๐กfor the๐-th test. To better approximate real-world cases, we consider an additional operational constraintsfor the operator (central controller) and require that๐ข๐ก โค ๐(kWh) for every๐ก โ [๐]. The total number of tests is ๐ฟand the total number of charging sessions is๐. Additionally, define the mean percentage error with respect to the undelivered energy corresponding to (4.4d) as
MPE:=1โ
๐ฟ
โ๏ธ
๐=1 ๐
โ๏ธ
๐ก=1 ๐
โ๏ธ
๐=1
๐ (๐)
๐ก (๐)
(๐ฟร๐) ยท
๐
โ๏ธ
๐=1
๐๐
, (4.15)
where๐๐is the energy request for each charging session ๐ โ [๐];๐ (๐)
๐ก (๐)is the energy scheduled to the ๐-th charging session at time๐กfor the๐-th test.
Hyper-parameters.
The detailed parameters used in our experiments are shown in Table 4.7.1.
Control spaces. For the experimental results presented in this section, the control state space isX=R2+ร๐ where๐ is the total number of charging stations and a state vector for each charging station is(๐๐ก,[๐(๐) โ๐ก]+), i.e., the remaining energy to be charged
Table 4.7.1: Hyper-parameters in the experiments.
Parameter Value
System Operator Number of power levels|U| 10
Cost functions๐1, . . . , ๐๐ Average LMPs
Operator function๐ Penalized Predictive Control Tuning parameter ๐ฝ 1ร103- 1ร106
EV Charging Aggregator
Number of Chargers๐ 54
State spaceX R108+
Action space [0,1]10
Time intervalฮ 12 minutes
Private vector (๐(๐), ๐(๐), ๐(๐), ๐(๐)) ACN-Data [1]
Power rating 150 kW
Scheduling algorithm๐ Least Laxity First (LLF)
Laxity ๐๐ก(๐) โ๐๐ก(๐)/๐(๐)
RL algorithm Soft Actor-Critic (SAC) [73]
Optimizer Adam [102]
Learning rate 3ยท10โ4
Discount factor 0.5
Relay buffer size 106
Number of hidden layers 2
Number of hidden units per layer 256 Number of samples per minibatch 256
Non-linearity ReLU
Reward function ๐1=0.1,๐2=0.2,๐3 =2
Temperature parameter 0.5
and the remaining charging time if it is being used (see Section 4.3); otherwise the vector is an all-zero vector. The control action space isU = {0,15,30, . . . ,150}
(unit: kW) with |U| =10, unless explicitly stated. The scheduling policy๐ is fixed to be least-laxity-first (LLF).
RL spaces. The RL action space3 of the Markov decision process used in the RL algorithm is [0,1]10. The outputs of the neural networks are clipped into the probability simplex (space of MEF)Pafterwards.
3Note that the RL action space (consisting of๐๐กโs) and state space (consisting of๐ฅ๐กโs) referred here are the standard definitions in the context of RL and they are different from the โcontrol action spaceโUand โcontrol state spaceโXdefined in Section 4.3.
RL rewards. We use the following specific reward function for our EV charging scenario, as a concrete example of (4.11):
๐EV(๐ฅ๐ก, ๐๐ก) =H(๐๐ก) +๐1
๐โฒ
โ๏ธ
๐=1
โฅ๐ข๐ก(๐) โฅ2
โ๐2
๐โฒ
โ๏ธ
๐=1
I(๐(๐๐) โค๐ก โค ๐(๐๐) +ฮ)h ๐(๐) โ
๐
โ๏ธ
๐ก=1
๐ข๐ก(๐)i
+
!
โ๐3
๐๐ก(๐๐ก) โ
๐โฒ
โ๏ธ
๐=1
๐ข๐ก(๐)
(4.16) where๐1, ๐2and๐3are positive constants;๐โฒis the number of EVs being charged;
๐๐ก is the operator function, which is specified by (4.13);I(ยท)denotes an indicator function and๐(๐๐)is the arrival time of the๐โ๐ก โEV in the charging station with ๐๐ being the index of this EV in the total accepted charging sessions [๐]. The entropy functionH(๐๐ก)in the first term is a greedy approximation of the definition of MEF (see Definition 4.4.1). The second term is to further enhance charging performance and the last two terms are realizations of the last term in (4.11) for constraints (4.4c) and (4.4d). Note that The other constraints in the example shown in Section 4.3 can automatically be satisfied by enforcing the constraints in the fixed scheduling algorithm๐. With the settings described above, in Figure 4.5.2 we show a typical training curve of the reward function in (4.16). We observe policy convergence with respect to a wide range of choices of the hyper-parameters๐1, ๐2and ๐3. In our experiments, we do not optimize them but fix the constants in (4.16) as๐1 =0.1, ๐2 =0.2 and๐3 =2.
Cost functions. We consider the specific form of costs in (4.5a). In the RL training process, we train an aggregator function๐ using linear price functions๐๐ก =1โ๐ก/24 where๐ก โ [0,24](unit: Hrs) is the time index and we test the trained system with real price functions๐1, . . . , ๐๐ being the average locational marginal prices (LMPs) on the CAISO (California Independent System Operator) day-ahead market in 2016 (depicted at the bottom of Figure 4.7.3).
Tuning parameters. In PPC defined in Algorithm 6, there is a sequence of tuning parameters (๐ฝ1, . . . , ๐ฝ๐). In our experiments, we fix ๐ฝ๐ก = ๐ฝfor all๐ก โ [๐] where ๐ฝ > 0 is a universal tuning parameter that can be varied in our experiments.
Figure 4.7.1: Trade-offs of cost and charging performance. The dashed curve in the left figure corresponds to offline optimal cost. The tested days are selected (with no less than 30 charging sessions, i.e.,๐ โฅ 30) from Dec. 2, 2019 to Jan. 1, 2020.
Figure 4.7.2: Charging results of EVs controlled by PPC with tuning parameters ๐ฝ =2ร103 (top), 4ร103(mid) and 6ร103 (bot) for selected days (with no less than 30 charging sessions, i.e.,๐ โฅ 30) from Dec. 2, 2019 to Jan. 1, 2020. Each bar represents a charging session.
Experimental Results.
Sensitivity of๐ฝ. We first show how the changes of the tuning parameter ๐ฝaffect the total cost and feasibility. Figure 4.7.1 compares the results by varying๐ฝ. The agents are trained on data collected from Nov. 1, 2018 to Dec. 1, 2019 and the tests are performed on data from Dec. 2, 2019 to Jan. 1, 2020 . Weekends and days with less than 30 charging sessions are removed from both training and testing data. For
0 20 40 60 80 100 120
Power(kW)
Offline MPC PPC
0 3 6 9 12 15 18 21
Hrs 25
50
Price($)
Figure 4.7.3: Substation charging rates generated by the PPC (orange) in the closed-loop control shown in Algorithm 5, together with the MPC generated (blue) and global optimal (dashed black) charging rates.
charging performance, we show in Figure 4.7.2 the battery states of each session after the charging cycle ends, tested with tuning parameters ๐ฝ = 2ร103,4ร103 and 6ร103, respectively. The results indicate that with a sufficiently large tuning parameter, the charging actions given by the PPC is able to satisfy EVsโ charging demands and in practice, there is a trade-off between costs and feasibility depending on the choice of tuning parameters.
Charging curves. In Figure 4.7.3, substation charging rates (in kW) are shown. The charging rates generated by the PPC correspond to a trajectory
โ๏ธ
๐
๐ 1(๐)/ฮ, . . . ,
โ๏ธ
๐
๐ ๐(๐))ฮ
! ,
which is the aggregate charging power given by the PPC for all EVs at each time ๐ก =1, . . . , ๐. The agent is trained on data collected at Caltech from Nov. 1, 2018 to Dec. 1, 2019 and tested on Dec. 16, 2019 at Caltech using real LMPs on the CAISO day-ahead market in 2016. We use a tuning parameter๐ฝ =4ร103for both training and testing. The figure highlights that, with a suitable choice of tuning parameter, the operator is able to schedule charging at time slots where prices are lower and avoid charging at the peak of prices, as desired. In particular, it achieves a lower cost compared with the commonly used MPC scheme described in (2.3)-(4.17f).The offline optimal charging rates are also provided.
0.0 0.2 0.4 0.6 0.8 1.0 MPE
0 10 20 30
Costs(k)
Offline MPC PPC
Figure 4.7.4: Cost-energy curves for the offline optimization in (4.2a)-(4.2d) (for the example in Section 4.3), MPC (defined in (4.17a)-(4.17f)) and PPC (introduced in Section 4.6).
Comparison of PPC and MPC. In Figure 4.7.4, we show the changes of the cumulative costs by varying the mean percentage error (MPE) with respect to the undelivered energy defined in (4.15). There are in total๐พ =14 episodes tested for days selected from Dec. 2, 2019 to Jan. 1, 2020 (days with less than 30 charging sessions are removed, i.e. we require,๐ โฅ 30). Note that 0โค MPEโค 1 and the larger MPE is, the higher level of constraint violations we observe. We allow constraint violations and modify parameters in the MPC and PPC to obtain varying MPE values.
For the PPC, we vary the tuning parameter ๐ฝto obtain the corresponding costs and MPE. For the MPC in our tests, we solve the following optimization at each time for obtaining the charging decisions๐ ๐ก = (๐ ๐ก(1), . . . , ๐ ๐ก(๐โฒ)):
๐ ๐ก =arg min
๐ ๐ก
๐กโฒ
โ๏ธ
๐=๐ก
๐๐
๐โฒ
โ๏ธ
๐=1
๐ ๐(๐)
subject to : (4.17a)
๐ ๐(๐) =0, ๐ < ๐(๐), ๐=1, . . . , ๐โฒ, (4.17b) ๐ ๐(๐) =0, ๐ > ๐(๐), ๐=1, . . . , ๐โฒ, (4.17c)
๐โฒ
โ๏ธ
๐=1
๐ ๐(๐) =๐ข๐ก, ๐ =๐ก , . . . , ๐กโฒ, (4.17d)
๐
โ๏ธ
๐=1
๐ ๐(๐) =๐พยท๐(๐), ๐=1, . . . , ๐โฒ, (4.17e) 0 โค ๐ ๐(๐) โค๐(๐), ๐ =๐ก , . . . , ๐กโฒ (4.17f)
where at time ๐ก, the integer ๐โฒ denotes the number of EVs being charged at the charging station and the time horizon of the online optimization is from๐ =๐ก to๐กโฒ, which is the latest departure time of the present charging sessions; ๐(๐) and ๐(๐) are the arrival time and departure time of the๐-th session;๐พ >0 relaxes the energy demand constraints and therefore changes the MPE region for MPC. The offline cost-energy curve is obtained by varying the energy demand constraints in (4.4d) in a similar way. We assume there is no admission control and an arriving EV will take a charger whenever it is idle for both MPC and PPC. Note that this MPC framework is widely studied [103] and used in EV charging applications [96]. It requires the precise knowledge of a 108-dimensional state vector of 54 chargers at each time step. We observe that withonlyfeasibility information, PPC outperforms MPC for all 0โค MPE โค1. The main reason that PPC outperforms vanilla MPC is that PPC utilizes MEF as its input, which is generated by a pre-trained aggregator function.
Therefore the MEF may contain useful future feasibility information that vanilla MPC does not know, despite that it is trained and tested on separate datasets.
APPENDIX
4.A Proof of Lemma 16
Proof. We first define a set ๐โ1(X๐ก) denoting the inverse image of the setX๐ก for actions: ๐โ1(X๐ก) (๐ข<๐ก) := {๐ข โU: ๐(๐ฅ๐กโ1, ๐ข) โ X๐ก}. The inverse image ๐โ1(X๐ก) depends only on the past actions๐ข<๐ก since the states๐ฅ<๐ก are determined by๐ข<๐ก and a pre-fixed initial state๐ฅ1via the dynamics in (3.3). Note thatX๐ก and the dynamic ๐ are Borel measurable. Therefore the inverse image ๐โ1(X๐ก) is also a Borel set, implying that the intersection U๐ก
ร ๐โ1(X๐ก) is also Borel measurable. The set of feasible action trajectoriesScan be reprised as
S:=n
๐ข โU๐ :๐ข๐ก โU๐ก
ร
๐โ1(X๐ก) (๐ข<๐ก),โ๐ก โ [๐]o ,
which is a Borel measurable set of allfeasiblesequences of actions. โก 4.B Proof of Lemma 17
Proof of Lemma 17. We prove the statement by induction. It is straightforward to verify the results hold when๐ = 1. We suppose the lemma is true when๐ = ๐. Suppose๐ =๐+1. Let
๐ญ(๐ข):= max
๐2,..., ๐๐ ๐
โ๏ธ
๐ก=2
H(๐๐ก|U2:๐กโ1;๐1=๐ข)
denote the optimal value corresponding to the time horizon๐ก โ [๐], given the first action๐1=๐ข. By the definition of conditional entropy, we have
๐ญ=max
๐1
โซ
๐ขโU
๐1(๐ข)๐ญ(๐ข)d๐ข+H(๐1). By the induction hypothesis,๐ญ(๐ข) =๐(S(๐ข)). Therefore,
๐ญ=max
๐1
โซ
๐ขโU
๐1(๐ข)log๐(S(๐ข))d๐ข+H(๐1)
=max
๐1
โซ
๐ขโU
๐1(๐ข)log
๐(S(๐ข)) ๐1(๐ข)
d๐ข whose optimizer ๐โ
1 satisfies (4.8) and we get ๐ญ = ๐(S). The lemma follows by finding the optimal conditional distributions ๐โ
1, . . . , ๐โ
๐ inductively. โก
4.C Proof of Corollary 4.4.1
Proof of Corollary 4.4.1. Lemma 17 shows that the value of the density function corresponding to choosing ๐ข๐ก = ๐ข in the MEF is proportional to the measure of
S( (๐ข<๐ก, ๐ข)), completing the proof of interpretability. According to the explicit expres- sion in (4.8) of the MEF, the selected action๐ขalways ensures that๐( (S( (๐ข<๐ก, ๐ข))) > 0 and therefore the setS( (๐ข<๐ก, ๐ข) is non-empty. This guarantees that the generated
sequence๐ขis always inS. โก
4.D Proof of Corollary 4.6.1
Proof of Corollary 4.6.1. The explicit expression in Lemma 17 ensures that when- ever ๐โ
๐ก(๐ข๐ก|๐ข<๐ก) > 0, then there is always a feasible sequence of actions inS(๐ข<๐ก).
Now, if the tuning parameter ๐ฝ๐ก > 0, then the optimization (4.13) guarantees that ๐โ
๐ก(๐ข๐ก|๐ข<๐ก) > 0 for all๐ก โ [๐]; otherwise, the objective value in (4.13) is unbounded.
Corollary 4.4.1 guarantees that for any sequence of actions ๐ข = (๐ข1, . . . , ๐ข๐), if ๐โ
๐ก(๐ข๐ก|๐ข<๐ก) > 0 for all๐ก โ [๐], then๐ข โ S. Therefore, the sequence of actions๐ข
given by the PPC is always feasible. โก
4.E Proof of Theorem 18
Proof of Lemma 18. We note that the offline optimization (4.2a)โ(4.2d) is equivalent to
inf
๐ขโU๐ ๐
โ๏ธ
๐ก=1
๐๐ก(๐ข๐ก) โ๐ฝlog๐(๐ข) (4.18) for any ๐ฝ >0 and๐(๐ข)is a uniform distribution onS:
๐(๐ข) :=
๏ฃฑ๏ฃด
๏ฃด
๏ฃฒ
๏ฃด๏ฃด
๏ฃณ
1/๐(S) if๐ข โS
0 otherwise
where ๐(ยท)is the Lebesgue measure. Further, decomposing the joint distribution ๐(๐ข) =ร๐
๐ก=1๐โ
๐ก(๐ข๐ก|๐ข<๐ก)into the conditional distributions given by (4.7a)-(4.7c), the objective function (4.18) becomes
๐
โ๏ธ
๐ก=1
๐๐ก(๐ข๐ก) โ ๐ฝlog
๐
ร
๐ก=1
๐โ
๐ก(๐ข๐ก|๐ข<๐ก)
!
=
๐
โ๏ธ
๐ก=1
๐๐ก(๐ข๐ก) โ ๐ฝlog๐โ
๐ก(๐ข๐ก|๐ข<๐ก) ,
which implies the lemma. โก
4.F Proof of Theorem 4.6.1
Proof. Define the following optimalcost-to-go function, which computes the minimal cost given a subsequence of actions:
๐OPT
๐ก (๐ข<๐ก) := min
๐ข๐ก:๐โU๐โ๐ก+1 ๐
โ๏ธ
๐=๐ก
๐(๐ข๐) โ๐ฝlog๐โ
๐ก(๐ข๐ก:๐|๐ขโ
๐กโ๐:๐กโ1)
!
=min
๐ข๐กโU
๐(๐ข๐ก) โlog๐โ
๐ก(๐ข๐ก|๐ขโ
๐กโ๐:๐กโ1)+
min
๐ข๐ก+1:๐โU๐โ๐ก
๐
โ๏ธ
๐=๐ก+1
๐(๐ข๐) โlog๐โ
๐ก+1(๐ข๐ก+1:๐|๐ขโ
๐กโ๐:๐ก)
!
=min
๐ข๐กโU
๐(๐ข๐ก) โlog๐โ
๐ก(๐ข๐ก|๐ขโ
๐กโ๐:๐กโ1) +๐OPT
๐ก+1 (๐ขโค๐ก) .
Let ๐(ยท) denote the Lebesgue measure. Based on the definition of the optimal cost-to-go functions defined above and applying Lemma 18, we obtain the following expression of the optimal action๐ขโ
๐ก at each time๐ก โ [๐]:
๐ขโ
๐ก =arg min
๐ข๐กโU
๐๐ก(๐ข๐ก) โ๐ฝlog๐โ
๐ก(๐ข๐ก|๐ขโ
๐กโ๐:๐กโ1) +๐OPT
๐ก+1 (๐ข๐กโ๐+1:๐ก)
=arg min
๐ข๐กโU
๐๐ก(๐ข๐ก) โ๐ฝlog
๐ S(๐ขโ
๐กโ๐:๐กโ1, ๐ข๐ก) ๐
S(๐ขโ
๐กโ๐:๐กโ1) + min
๐ข๐ก+1:๐
๐
โ๏ธ
๐=๐ก+1
๐(๐ข๐) โ๐ฝlog
๐ S(๐ขโ
๐กโ๐:๐กโ1, ๐ข๐ก:๐) ๐
S(๐ขโ
๐กโ๐:๐กโ1, ๐ข๐ก:๐โ1)
!
=arg min
๐ข๐กโU
๐๐ก(๐ข๐ก) โ๐ฝlog
๐ S(๐ขโ
๐กโ๐:๐กโ1, ๐ข๐ก) ๐
S(๐ขโ
๐กโ๐:๐กโ1) + min
๐ข๐ก+1:๐
๐
โ๏ธ
๐=๐ก+1
๐(๐ข๐) +๐ฝlog
๐ S(๐ขโ
๐กโ๐:๐กโ1, ๐ข๐ก) ๐
S(๐ขโ
๐กโ๐:๐กโ1, ๐ข๐ก:๐)
! , which implies (4.19) below
๐ขโ
๐ก =arg min
๐ข๐กโU
๐๐ก(๐ข๐ก) + min
๐ข๐ก+1:๐ ๐
โ๏ธ
๐=๐ก+1
๐(๐ข๐) โlog๐ S(๐ขโ
๐กโ๐:๐กโ1, ๐ข๐ก:๐)
!
| {z }
๐(๐ข๐ก)
!
(4.19)
and when๐ข<๐ก =๐ขโ
<๐ก, the solution of the PPC in Algorithm 6 satisfies ๐ข๐ก =arg min
๐ข๐กโU
๐๐ก(๐ข๐ก) โ๐ฝlog๐๐ก(๐ข๐ก|๐ขโ
๐กโ๐:๐กโ1)
=arg min
๐ข๐กโU
ยฉ
ยญ
ยญ
ยซ
๐๐ก(๐ข๐ก) โ๐ฝlog
๐ S(๐ขโ
๐กโ๐:๐กโ1, ๐ข๐ก) ๐
S(๐ขโ
๐กโ๐:๐กโ1)
ยช
ยฎ
ยฎ
ยฌ
=arg min
๐ข๐กโU
๐๐ก(๐ข๐ก) +๐ฝlog 1/๐ S(๐ขโ
๐กโ๐:๐กโ1, ๐ข๐ก)
| {z }
๐(๐ข๐ก)
!
. (4.20)
Since the cost functions๐๐ก(๐ข)and the measure log(1/๐(S(๐ข)))are strictly convex, the inner minimization in (4.19) is a convex minimization and hence ๐(๐ข๐ก)is convex.
Therefore,๐ขโ
๐ก in (4.19) is unique. Denoting by๐โฒand ๐โฒthe corresponding derivatives of a given cost function๐and the function ๐ defined in (4.19), we have
๐โฒ
๐ก(๐ขโ
๐ก) + ๐โฒ(๐ขโ
๐ก) =0.
Furthermore, the unique solution of the PPC scheme satisfies ๐โฒ
๐ก(๐ข๐ก) +๐ฝ๐โฒ(๐ข๐ก) =0
where๐โฒis the derivative of the function๐ defined in (4.20). Choosing๐ฝ = ๐๐ก = ๐โฒ(๐ขโ
๐ก)/๐โฒ(๐ขโ
๐ก) implies that๐ข๐ก =๐ขโ
๐ก for all๐ก โ [๐]. โก
C h a p t e r 5