

Next:7.
Extensions and Concluding RemarksUp:6.
Risk-Sensitive Hierarchical ControlsPrevious:6.1
Risk-sensitive hierarchical controls with
6.2 Risk-sensitive hierarchical controls with long-run average costs
In this section we consider a manufacturing system with the objective of
minimizing the risk-sensitive average cost criterion over an infinite horizon.
The risk-sensitive approach has been applied to the so-called disturbance
attenuation problem; see, for example,
Whittle
(1990), Fleming and McEneaney (1995),
and references therein.
Let us consider a single product, parallel-machine manufacturing system
with stochastic production capacity and constant demand for its production
over time. For
,
let
,
,
and
z denote the surplus level (the state variable), the production
rate (the control variable), and the constant demand rate, respectively.
We assume
,
,
,
and z a positive constant. They satisfy the differential equation
|
|
|
(6.1) |
where a>0 is a constant, representing the deterioration rate (or
spoilage rate) of the finished product.
We let
represent the maximum production capacity of the system at time t,
where
is given in Section 3.1. The production constraints
is given by the inequalities
Definition 6.1 A production
control process
is admissible if
-
(i)
-
is
-progressively
measurable;
-
(ii)
-
for all
.
Let
denote the class of admissible controls with
.
Let H(x,u) denote a cost function of surplus and production.
The objective of the problem is to choose
to minimize
|
|
|
(6.2) |
where
is the surplus process corresponding to the production process
.
Let
.
A motivation for choosing such an exponential cost criterion is that such
criteria are sensitive to large values of the exponent which occur with
small probability, for example, rare sequences of unusually many machine
failures resulting in shortages (
).
We assume that the cost function H(x,u) and the production
capacity process
satisfy the following:
Assumption 6.1
is continuous, bounded, and uniformly Lipschitz in x.
Remark 6.1 In manufacturing
systems, the running cost function
H(x,u) is usually
chosen to be of the formH(x,u)=h(x)+c(u)
with piecewise linear h(x) and c(u). Note that
piecewise linear functions are not bounded as required in (A.6.1). However,
this is not important, in view of the uniform bounds on
and on
for the initial state
in any bounded set.
Assumption 6.2 Q is
irreducible
in the following sense: The equations
have a unique solution
with
.
The vector
is called the
equilibrium distribution of the Markov chain
.Formally,
we can write the associated HJB equation as follows:
where
is the potential function,
denotes the partial derivative of
with respect to x, and
for a function f on
.
Zhang
(1995) proves the following theorem.
Theorem 6.5 (i) The HJB
equation (6.3)
has a viscosity
solution
.
(ii) The pair
satisfies
the following conditions:For some constant C1 independent of
,
-
(a)
-
,
and
-
(b)
-
.
(iii) Assume that
to
be Lipschitz continuous in x. Then,
This theorem implies that
in the viscosity solution
is unique. Furthermore, Zhang (1995) gives
the following verification theorem.
Theorem 6.6 Let
be
a viscosity solution to the HJB equation in (6.3).
Assume
that
to
be Lipschitz continuous in x. Let
.Suppose
that there are
,
,
andr*(t)
such that
satisfying
|
|
 |
|
|
|
 |
(6.4) |
a.e. in t and w.p. 1. Then,
.
We next discuss the asymptotic property of the HJB equation (6.5)
as
.
First of all, note that this HJB equation is similar to that for an ordinary
long-run average cost problem except for the term involving the exponential
factor. In order to get rid of such a term, we make use of the logarithmic
transformation as in Fleming and Soner (1992).
Let
.
Define
With the logarithmic transformation, we have for each
,
The supremum is obtained at
.
The logarithmic transformation suggests that the HJB equation is equivalent
to an Isaacs equation for a two-player zero-sum dynamic stochastic game.
The Isaacs equation is given as follows:
|
|
|
(6.5) |
where
for
;
see Evans and Souganidis (1984) and Fleming
and Souganidis (1989).
We consider the limit of the problem as
.
In order to define a limiting problem, we first define the control sets
for the limiting problem. Let
and
For each
,
let
be such that
and let
denote the equilibrium distribution of
.
The next theorem says that
is irreducible. Therefore, there exists a unique positive
for each
.
Moreover,
depends continuously on V. It can be shown that for each
,
is irreducible.
Theorem 6.7 Let
be
a sequence such that
and
.
Then,
-
(i)
-
w0(x,k) is independent of k,
i.e.,w0(x,k)=w0(x);
-
(ii)
-
w0(x) is Lipschitz; and
-
(iii)
-
(k0,w0(x)) is a viscosity
solution to the following Isaacs equation:
k0= |
|
 |
|
|
|
 |
(6.6) |
Let
Note that
,
where
is the sup norm. Moreover, since
,
where V=1 means
vi(j)=1 for all i,j.
Then the equation in (6.6) is an Isaacs
equation associated with a two-player, zero-sum dynamic game with the objective
subject to
where
and
are Borel measurable functions and
and
for
;
see Barron and Jensen (1989).
One can show that
which implies the uniqueness of
.
Finally, in order to use the solution to the limiting problem to obtain
a control for the original problem, a numerical scheme must be used to
obtain an approximate solution. The advantage of the limiting problem is
its dimensionality, which is much smaller than that of the original problem
if the number of states in
is large.
Let (U*(x),V*(x))
denote a solution to the limiting problem. As in Section 3.1,
it is expected that the constructed control
is nearly optimal for the original problem (6.2).
See Zhang (1995) for the proof of this result.


Next:7.
Extensions and Concluding RemarksUp:6.
Risk-Sensitive Hierarchical ControlsPrevious:6.1
Risk-sensitive hierarchical controls with