Volume 2013, Issue 1 368096
Research Article
Open Access

Optimal Oil-Owner Behavior in Piecewise Deterministic Models

Atle Seierstad

Corresponding Author

Atle Seierstad

Department of Economics, University of Oslo, P.O. Box 1095, Blindern, 0317 Oslo, Norway uio.no

Search for more papers by this author
First published: 28 November 2013
Academic Editor: Alain Piétrus

Abstract

Six simple piecewise deterministic oil production models from economics are discussed by using solution tools that are available in the theory of piecewise deterministic optimal control problems.

1. Introduction

Six simple piecewise deterministic models for optimal oil-owner behavior are presented. Their central property is sudden jumps in states. The aim of this paper is to show in admittedly exceedingly simple models how available tools for piecewise deterministic models, namely, the HJB equation and the maximum principle, can be used to solve these models analytically. We are looking for solutions given by explicit formulas. That can only be obtained if the models are simple enough. The models may be too simple to be of much interest in themselves, but they can provide some intuition about features optimal solutions may have in more complicated models.

Piecewise deterministic models have been used a number of times in economic problems in the literature; some few scattered references are given that contain such applications [14]. I have not been able to find references directly concerned with piecewise deterministic oil production problems. For different probability structures, and for discrete time, a host of related problems has been discussed in the literature; references to such literature have been left out, with one exception. Problems of control of jump diffusions, see [5], encompass piecewise deterministic problems, and some problems appearing in [5] are related to the ones discussed below. A classic reference to piecewise deterministic control problems is [2].

In all models below, an unbounded number of jumps in the state can occur at times 0 < τ1 < τ2 < τ3 < ⋯, and, when τj is given, τj+1 is exponentially distributed in [τj, ) (all τiτi−1 independent). Sometimes, the size of the jumps is influenced by stochastic variables Vj. Let ω = (τ1, V1, τ2, V2, …). At time t, we imagine that the control values chosen can be allowed to depend on what has happened, that is, on τi, Vi for τi < t, but not on future events, that is, τi, Vi for which τi > t. Such controls (written u(t, ω)) are called nonanticipative. Corresponding state solutions denoted by x(t, ω) are then also nonanticipative. (A general set-up, with further explanations, is given in Appendix.) Frequently below, x(t, ω) will be the wealth of the oil owner. In infinite horizon economic models, the weakest terminal condition that is natural to apply is an expectational no-Ponzi-game condition; namely, liminf TE[ekTx(T, ω)] ≥ 0, where k is the discount factor. Note that some stronger conditions will be used in some of the models presented in the sequel.

2. Model 1

Consider the optimization problem
()
where u is the control, subject to the state dynamics,
()
where k and x0 are fixed positive constants,
()
αi fixed. (Formally, only sup iαi < is needed.) Let τ0 = 0. There is a fixed intensity λ of jumps; that is, for i = 1,2, …, given , τiτi−1 (all τiτi−1 independent), and we assume δ > k.

The interpretation of the model is that u is consumption rate, αi is the size (in dollars) of an oil field found at time τi, x is wealth, and kx is interest. Oil fields are sold immediately after discovery.

Let us solve problem (1)–(3) by using the extremal method (see the appendix). Now, with H = eδtuγ/γ + p(kxu), solving the first-order condition for maximum of H, we get that u = eδt/(γ−1)p1/(γ−1) maximizes the Hamiltonian.

The adjoint equation (see (A.8) in Appendix) is
()
we guess that we do not need the fact that p(t, j) depends on arbitrary initial points (s, y), as in appendix. Equation (4) is satisfied by , (we try the possibility that p(t, j) is even independent of j).

Now, for , ρ : = (kδ)/(γ − 1) > 0, we get u(t) = eδt/(γ−1)(p(t)) 1/(γ−1) = beρt. Because H is concave in (x, u), ax is linear, and αi is independent of x, sufficient conditions based on concavity (see Theorem 3 in Appendix) then give us that the control u = beρt is an optimal control for problem (1)–(3). The optimal control is independent of λ and the αi’s.

Write β : = b/(k + ρ). For u = beρt, we have that . The solution of this equation equals, for t ∈ (τi, τi+1),
()
To see this, find and simply check that the differential equation is satisfied. Moreover, evidently x*(·) satisfies x*(τi+) = x*(τi−) + αi. (To ensure that x*(t)≥0 a.s., we could have postulated at the outset that a is so large that x0β.)

Write x*(t) = ektΦi(t, ω), for t ∈ (τi, τi+1), where .

We are now going to replace the bequest function ax(T) by the terminal constraint Ex(T, ω) ≥ 0. For this purpose, we are now going to vary a and hence also and β. For t = T, consider the expectation of Φi(t, ω). We will not give an explicit formula for this expectation, but we mention that it can be calculated in two steps, first given that i jumps have happened in [0, T] and then the expectation with i being stochastic (the rule of double expectations is then used). The expectation of the sum containing αj as well as x0 is positive, and the term in front of β is negative. There thus exists a unique positive value βT of β and, hence, of a (denoted by aT) such that EΦi(T, ω) = 0; that is, Ex*(T, ω) = 0.

If we drop the term ax(T) in the criterion but add the terminal condition Ex(T, ω) ≥ 0, the free end optimality obtained in the original problem for a = aT evidently means optimality of in the end constrained problem (the found value aT appears in u, though now not in the criterion). (Alternatively, we could use the sufficient conditions for end constrained problems in Theorem 3 in Appendix to obtain optimality in the present end constrained problem. This would require us to check condition (A.13), which is easily done.)

Let us discuss a little more the value of β for large values of T. Now,
()
And, continuing, we get .

Then, .

Now, with probability 1 an infinite number of jumps occur in [0, ), and
()
Denote by β the value of β for which EΦ(, ω) = 0, and define b : = β(ρ + k). For T being large, EΦi(T, ω) ≈ EΦi(, ω) = EΦ(, ω), so βT is close to β and bT : = βT(ρ + k) is close to b. So for T being large, the optimal control is approximately beρt. Note that ρ belongs to (δk, ) and both ρ and b are increasing in γ ∈ (0,1), so a larger γ means we consume more in the beginning. In the problem where ax(T) is replaced by the terminal constraint Ex(T, ω) ≥ 0, the optimal control depends on αi, λ. It is immediately seen that β is increasing in λ and in each αi, so also b has these properties, indicating, for T being large, that even βT and bT have these properties. This conclusion actually follows for any T, because is evidently increasing in λ and in each αi.

We may assume that the jumps are stochastic, that is, that x(τi+) − x(τi−) = Vi, where take values in a common bounded set and are independent, and are independent of the τi’s. If we then assume that ∑E(Vi) < , the solution in this problem is the same as the one above for αi = E(Vi).

Note that we must assume that we have a deal with the bank in which our wealth is placed that it accepts the above behavior. That is, before time 0, we have got an acceptance for the possibility of operating with this type of admissible solutions, which means that only in expectation we leave a wealth in the bank ≥0. In actual runs, sometimes we leave in the bank a positive wealth (that the bank gets), and for other runs a negative wealth (debt) that the bank has to cover.

Consider now the case where T = , a = 0 in (1). Then, for ,
()
when β = β. In this case, u = beρt is optimal in the problem where liminf TE[p(T)x(T, ω)] ≥ 0 is added as an end constraint. (See the appendix, (A.14).) The latter condition is equivalent to the so-called no-Ponzi-game condition.

3. Model 2

(This model Is related to exercise 4.1 in Øksendal and Sulem [5].)

Consider the following problem:
()
where u is the control, subject to
()
where α is a given positive number, δ > γk, τi, i = 1,2, …, is exponentially distributed in (τi−1, ) with intensity λi and λi is decreasing towards zero (τ0 = 0, all τiτi−1 independent).

In contrast to problem (1)–(3), now the jumps are linearly dependent on x(τi−). To defend such a feature, one might argue that the richer we are, the more we are able to generate large jumps (the jump may actually represent a collection of oil finds). On the other side, we will assume that such jumps occur with smaller and smaller intensities.

It is easy to see that the current value function is of the form when we replace x0 by an arbitrary start value y. For any pair (x(t, ω), u(t, ω)) satisfying (10), we can write u(t, ω) = v(t, ω)x(t, ω), v(t, ω) ≥ 0. Then, for t ∈ (τi, τi+1),
()
so u(t) = y times some function w(t, ω), t ∈ (τi, τi+1). Then, the value of the criterion for this v(·), denoted by , is proportional to yγ and so is also sup v(,·,). A similar argument works for , the optimal value of the criterion when j ≥ 1 jumps have already occurred at time t = 0. So , j = 0,1, …. Writing J and Aj instead of and , the current value function J satisfies the following HJB equation:
()
The first order condition for maximum is γuγ−1γAjxγ−1 = 0. So u = Bjx, for Bj = (Aj) 1/(γ−1). Then,
()
or dividing by xγ and rearranging,
()
Dividing by Aj gives
()
Let A(λ) satisfy
()
Evidently, relationship (16) would be obtained from (15) if all λj = λ(⇒Aj = Aj+1); in that case, the optimal value function would be A(λ)yγ.

Now, assume first, for some k*, that . Then, is the optimal value in the problem with no jumps, so . Given , (14) determines , . It must be the case that Aj is decreasing. Given the same start point x, the optimal value when j + 1 jumps have occurred already at time 0 must be smaller than the optimal value when j jumps have occurred at time 0; in the former problem, prospective jumps have smaller probabilities for happening. (Let us show by backwards induction that Aj, j = 0,1, 2, …, k, is nonincreasing. Evidently, Ak = Ak+1. Assume by induction that AjAj+1. If Aj−1 < Aj, then, by (15), + λj(1 + α) γAj/Aj−1>−δ + γkλj+1 + λj+1(1 + α) γAj/Aj−1 > −δ + γrλj+1+λj+1(1 + α) γAj+1/Aj = . Because γ − 1 < 0 and z → (z) 1/(γ−1) is decreasing, a contradiction is obtained.) As we will let k* vary, denote the sequence defined by (14) for by .

Now, , so . In fact, as is easily seen, this holds for all j: . (Compare the optimal values in the case where jump j occurs at (before) t = 0 in the two problems where λm = 0, mk* + 1, and where λm = 0, mk* + 2.) Assume now that λj > 0 for all j and let . (It is shown below that, for some , , for all k*, j.) Then, lim jAj = A(0), and the Aj’s satisfy (14).

Let x*(t, ω) be the solution for u ≡ 0. By (11), , where . Choose the smallest j* such that . If j* jumps occur at once at 0  (λj = ,  jj*), while further jumps occur with intensity , then EΓ would equal . Hence, . For any admissible solution x(t, ω), x(t, ω) ≤ x*(t, ω), for all (t, ω); hence, for t = T, 0 ≤ E[eδTAj(x(T, ω)) γ] ≤ when T. So (A.7) in the appendix is satisfied. Hence, sufficient conditions hold (see Theorem 2 in Appendix) and , j = 0,1, …, are optimal. That is, if and x*(t, ω) = x0(1 + for t ∈ (τi, τi+1) (see (11)), then u*(t, ω) = v*(t, ω)x*(t, ω) is optimal (x*(t, ω) evidently satisfies x*(t, ω) > 0, for all t). Hence, . (In the appendix, for the sufficiency of the HJB equation to hold, it is required that u*(t, ω) is bounded if T < (for T = , we need boundedness for t in all bounded intervals). This is not the case here. But we could have replaced u by vx, v being the control. Assume that we require v(t, ω) to be bounded in the above manner. Now, v*(t, ω) is so bounded, and it is then optimal in the set of such bounded v(·, ·)’s.)

We can show that the Aj’s are increasing in α and in each λi, i > j. The simplest argument is that this must be so, when we now know that Ajyγ is the optimal value function after j jumps at (before) t = 0.

If all λj = λ < (δkγ)/α, then, using (16), when t ∈ (τi, τi+1), where 0 < ρ* = ρ + [λλ(1 + α) γ]/(1 − γ) < ρ = (δk)/(1 − γ). (The inequality αλ < δkγ implies δkγ + λλ(1 + α) γ > δkγ + λλ(1 + α) > 0, so A(λ) > 0. In particular, . The above calculations show that in the problem where , for j > j*, the optimal value functions are , where for jj*, given by backward induction using (14) for j < j*, again . Evidently, . Now, for jj* and, from (16), it follows that when j.) Now, relate the present case to what happened in Model 1. As and ρ* < ρ, in each interval (τi, τi+1), (d/dt)u*(t) > d/dt(beρt) (beρt is the optimal control in Model 1); moreover, when t passes each τi  u*(t) changes by a factor (1 + α).

4. Model 3

In this model, the physical volume of oil production is constant = 1, but the oil price z jumps up and down at Poisson distributed time points τi, i = 1,2, …. Let τ0 = 0. Consider the problem
()
where u is the control, subject to the differential equations below and
()
where Vi is taking a finite number of values with given probabilities pj, all Vi are identically distributed, and τi is exponentially distributed in (τi−1, ) with intensity λ (all the random variables τiτi−1, Vi independent). Let v = EV. The two states y and z are governed by (18) (no jumps in y) and
()
y(0) = y0, z(0) = z0, where y0 and z0 are given constants; in fact, it will be convenient to assume that z0 = v. The end constraint E[y(T)] ≥ 0 is required to hold. Here, y is wealth and ay is interest earned.
Let us use the extremal method (see the appendix) for solving this problem. In what follows, it is guessed that adjoint functions do not depend on arbitrarily given initial points (which in the appendix are denoted by (s, y)). The adjoint function corresponding to y is hence denoted by p(t, i). We make the guess that p(t, i) simply equals Keat (independent of i), K being an unknown. It does satisfy the adjoint equation
()
(Epi(t), the expectation with respect to Vi, simply equals p(t, i), since p(t, i) does not depend on Vi.) The adjoint function pz(t, i) corresponding to the state variable z satisfies , with pz(T, i) = 0 (z(T) is free), we do not need the formula for pz(t, i).

Next, maximum of the Hamiltonian H = eδtuγ/γ + p(ay + zu) is obtained for u satisfying the first order condition uγ−1eδt = p, so u = (eδtp) 1/(γ−1) = Kβeκt, where κ = (aδ)/(γ − 1) > 0, β = 1/(γ − 1).

Now, , so
()
where C = y0Kβ/(a + κ).
Note that, for any s, Ez(s) = v. We want y(t) to satisfy the condition Ey(T) = 0; that is,
()
which determines K. From now on, denote the solutions y(t) and z(t) by y*(t, ω) and z*(t, ω), with u*(t) = Kβeκt. Sufficient conditions (see Theorem 3 in the appendix) are easily seen to be satisfied. We need to check the end condition E[p(T, ω)(y(T, ω) − y*(T, ω))] ≥ 0  (p  (·, ω) = p(·)) for all admissible solutions y(t, ω) (see (A.13) in the appendix). The terminal condition is Ey(T, ω) ≥ 0, for any admissible y(t, ω), and by construction y*(t, ω) satisfies Ep(T)y*(T, ω) = 0, so the just-mentioned end condition is satisfied. We hence have optimality of u*(t), y*(t, ω), and z*(t, ω) among admissible triples (u(t, ω), y(t, ω), z(t, ω)) for which Ey(T, ω) ≥ 0  (z(t, ω) = z*(t, ω)).

If T = , the end condition required for using the sufficient condition in Theorem 2 in the appendix is liminf TE[p(T, ω)(y(T, ω) − y*(T, ω))] ≥ 0. The condition lim TE[p(T)y*(T, ω)] = 0 determines K; it now reduces to the condition y0 + v/a = Kβ/(a + κ). Moreover, we now have optimality among all triples u(t, ω), y(t, ω), z(t, ω) for which liminf TE[p(T)y(T, ω)] ≥ 0. (This inequality means that the next-to-last inequality is satisfied.) Note that Kβ and so the control, quite expectably, are increasing in v = EV.

5. Model 4

In this example, y is the rate of production (production per unit of time) of oil. The oil price z jumps up or down between two values at Poisson distributed time points. The income the owner does not consume, that is, zyu, can be used to increase the rate of production, with the change being proportional to zyu. (If zyu is negative, it means that he takes money out of his business and runs it down.) For simplicity, the proportionality factor is put equal to 1, and, simply, z jumps between the values 1 and 2. So consider
()
δ > 2γ, u ≥ 0 the control, subject to
()
where τi is exponentially distributed in (τi−1, ) with intensity λ (all Vi, τiτi−1 independent V0 = 1, z0 = 0). There are no jumps in y and y(0) = y0 > 0, y0 being fixed.
Let J(y, j), j = 1,2, be the current optimal value function sought, where j = z ∈ {1,2}. The current value HJB equation is as follows
()
Let us try J(y, 1) = A1yγ, J(y, 2) = A2yγ. Then, the maximizing u equals u = Bjy, Bj = (Aj) 1/(γ−1). In the two cases j = 1 and j = 2, we get the two equations
()
()
The equations can be solved for positive A1 and A2 (shown in a moment). Automatically, the solution y*(t, ω) corresponding to u(y, z) = u(y, j) = Bjy satisfies y*(t, ω) ≥ 0 a.s., for all t.

To show the existence of a solution of (26) and (27), note that if we put A2 = A1 in (26), then by dividing by A1, we get , and κ1(A1) = 0 yields . Similarly, if we put A1 = A2 in (27), we get , and κ2(A2) = 0 yields . Now, . Denote the right-hand sides in (26) and (27) by ϕ(A1, A2) and ψ(A2, A1), respectively. Both are increasing in the second place and decreasing in the first place. When , , and (the only difference in ψ(A, A) and ϕ(A, A) is the third term in the formulas). Thus, ψ(A2, A1) = 0 can be uniquely solved with respect to A2 for A1 in , denote the solution A2(A1), it is evidently increasing in A1 and , . Moreover, = ψ(A1, A1)⇒A2(A1) > A1. Consider the function β(A1): = ϕ(A1, A2(A1)) > ϕ(A1, A1). Now, and , so exists for which 0 = β(A1) = ϕ(A1, A2(A1)). So this and , A2 > A1, satisfy both (27) and (26).

For any admissible y(t, ω), , where y*(t) is the solution of (i.e., y*(T) = y0e2t), so, for all admissible y(t, ω), 0 ≤ lim TE[eδTAj(y(T, ω)) γ] ≤ . Hence, (A.7) in the appendix holds; the sufficient conditions in Theorem 2 are satisfied, and if u = y(Aj) 1/(γ−1) when Vi = j ∈ {1,2}, then u is optimal. (Hence, if + and y*(t, ω) is the solution of , then u*(t, ω) = v(t, ω)y*(t, ω) is optimal.)

Define A* = [(δ − 3γ/2)/(1 − γ)] γ−1 (i.e., A* satisfies 0 = −δA + Aγ/(γ−1) + 3γA/2 − γAA1/(γ−1)). When λ = 0 then , and one may show that when λ > 0 decreases, A2 increases towards while A1 decreases towards . When λ increases to infinity, A1 increases to A*, and A2 decreases to A*. In fact, and . To prove convergence to A*, note first that A1, A2 are bounded uniformly in λ. If A1 and A2 did not converge to a common number, say A*, when λ, the two last terms in (27) and (26) would blow up to infinity for (certain) large values of λ, while the remaining terms would be bounded, a contradiction. Summing the two equations in (27) and (26) gives . Hence, A* satisfies ; that is, A* = A*.

Two very simple models discussed below contain the feature that the owner can influence the chance of discovery but that it is costly to do so. In the first one, the intensity of discoveries λ is influenced by how much money is put into search at any moment in time; in the second one, it is costly buildup of expertise that matters for the intensity of discoveries.

6. Model 5

Consider the problem
()
and λ = μu1−κ/(1 − κ), where μ > 0, κ are given constants, κ ∈ (0,1), u ≥ 0 is the control, and λ is the intensity of jumps. All fields found are of the same size α. We imagine that fields are sold immediately when they are found or that they are produced over a fixed period of time, with a fixed production profile. In both cases, we let income from a field discounted back to the time of discovery be equal to g0 : = βα, for a given β > 0. The oil owner wants to maximize the sum of discounted incomes earned over [0, ).
Let us try the proposal that the current value optimal value function J = a, where a, an unknown. The current value HJB equation is
()
which implies that the maximizing u satisfies μg0uκ = 1, so u = (μg0) 1/κ, −u + μg0u1−κ/(1 − κ) = , and . Trivially, (A.7) in the appendix holds, so the sufficient conditions in Theorem 2 are satisfied and u = (μg0) 1/κ is optimal. Because a > 0, in expectation (but not in all runs), over [0, ), a positive sum of discounted incomes ( = a) has been earned.
Often, it is the case that the best fields are exploited first hence, x(τi+) − x(τi−) = αi, where αi is decreasing and even αi ↓ 0. Then, when j jumps are imagined to have happened already at time 0, we guess J(j) = aj. Define . Then, the HJB equation is
()
and yields maximum. Postulate for the moment that, for some j, . Then, = νj(aj+1). Thus, if we try aj = aj+1, we get δaj < νj(aj). If we try , we get δaj > νj(aj) = 0. So a correct value of exists such that δaj = ν(aj). This continues backwards and gives a0 > a1 > ⋯>aj+1. If, for some j*, αj = 0, for jj* + 1, then , so in this case, (30) gives . If all αj > 0, letting j*, it is easily seen that there exists an infinite decreasing sequence aj ↓ 0 when j satisfying (30) (compare a similar argument in Model 2).

Still, trivially condition (A.7) below holds, sufficient conditions in Theorem 2 in the appendix are satisfied and on [τj, τj+1), j = 0,1, 2, …, are optimal. (Uniqueness of the aj’s follows from optimality.)

7. Model 6

Consider the problem
()
no jumps in y, λ = μy, μ, κ are fixed positive numbers, κ < 1, λ is the intensity of jumps in x.

All fields found are of the same size α. We again imagine that fields are sold immediately when they are found or that they are produced over a fixed period of time, with a fixed production profile. Let income from a field discounted back to the time of discovery be equal to g0 : = βα, for a given β > 0. The state y(t) is the amount of expertise available for finding new fields, built up over time according to , where u is money per unit of time spent on building expertise.

Let us try the proposal that the current value function J equals a + ky, a and k being unknowns. The current value HJB equation is
()
Then, kδ = μg0, so k = μg0/δ. Moreover, the maximizing u equals u = k1/κ, and −u + ku1−κ/(1 − κ) = , . The optimal solution y*(t) equals tb, where b = k(1−κ)/κ/(1 − κ). Evidently, (A.7) below is satisfied, because J = eδT(a + ky(T)) ≥ 0 for any admissible x(·), y(·) and eδT(a + ky*(T)) → 0 when T. Hence, u = k1/κ = (μg0/δ) 1/κ is optimal.

8. Comparisons

In Models 1, 2, 5, and 6, oil finds are made at stochastic points in time; in Models 3 and 4, it is the price of oil that changes at stochastic points in time. In Model 1, we operate with the constraint Ex(, ω) ≥ 0 (for T = , a = 0), where x is the oil-owners’ wealth. Here, for some runs, x(, ω) can be negative (β > x0) and, for other runs, positive. In Model 2, we required x(t, ω) > 0 for all t, all ω. (The results in that model would have been the same if we had required only x(, ω) ≥ 0 for all ω.) In Model 2, the optimal control comes out as stochastic and not deterministic as in Model 1. Moreover, as a comment in Model 2 says, as a function of time, the optimal control decreases more rapidly in Model 1 as compared to Model 2. The latter feature stems from the fact that, in Model 2, we enhance future income prospects by not decreasing x too fast, because the jump term (the right-hand side of the jump equation) depends positively on x, which is not the case in Model 1.

In Models 3 and 4, the oil price exhibits sudden stochastic jumps. In Model 3, the rate of oil production is constant, but income earned (as well as interest) is placed in a bank after subtraction of consumption. In Model 4, income earned, after subtraction of consumption, is reinvested in the oil firm to increase production. In Model 4, the optimal control is stochastic; it depends on whether the current price is high or low. In Model 3, the control is deterministic, and it depends only on the expectation v of the stochastic price. Consider the case where, in Model 3, the expected price v is zero and a = 3/2. Then, u = Kβeκt, where κ = (3/2 − δ)/(γ − 1) and Kβ = y0(3/2 + κ). We saw in Model 4 that when the intensity of jumps λ is very high, A1A2A*, where A* = [δ − 3γ/2)/(1 − γ)] γ−1; we hardly pay attention to what the current price is, because it changes so frequently. Now u = y(Aj) 1/(γ−1) when the current price is j( = z), or uy(A*) 1/(γ−1) when λ is large. When λ is large (so z switches very frequently between 1 and 2), the stochastic path of the equation most often is very close to the deterministic path of = (3/2)y − [(δ − 3γ/2)/(1 − γ)]y = −κy. The latter equation has the solution y = y0eκt, and the corresponding u equals = (3/2)y0eκt + κy0eκt = κy0eκt = y0(3/2   +   κ)eκt, the same control as that obtained in Model 3 in the case a = 3/2, v = 0.

In the extremely simple Models 5 and 6, the frequency of oil finds is not fixed but influenced by a control. In Model 5, the current frequency (or intensity) is determined by how much money is put into search at that moment in time. In the simplest case considered in Model 5, a find today does not influence the possibility of making equally sized discoveries tomorrow. Then it, is not unreasonable that the optimal control (which equals (μg0) 1/κ) is independent of the discount rate δ but dependent on the fixed value of the finds g0 = βα. (Here, actually α could be the expected size of a find, in both Model 5 and Model 6; we could have had the sizes of the finds being independently stochastic, with α being the expected value of the sizes.) In Model 6, it is the ability to discover oil that is built up over time, so with greater impatience (higher δ), we should expect less willingness of devoting money to increase this ability, and this shows up in the formula u = (μg0/δ) 1/κ.

Acknowledgment

The author is very grateful for useful comments received from a referee that made it possible to improve the exposition and remove some typographical errors.

    Appendix

    Consider the problem
    ()
    ()
    x0, U, T > 0 fixed, u is the control, with criterion
    ()
    Here, τj and Vj, j = 1,2, …, are random variables, with all τjτj−1 and Vj being independent (τ0 = 0). Given τj−1, the intensity of the jump τj in [τj−1, ) is λj(t, x, u). (So for given nonanticipative function x(t) = x(t, ω), u(t) = u(t, ω), for ττj−1, the probability distribution of τ = τj is .)

    Assume that the five given functions f0, f, g0, g, and h0 are continuous and Lipschitz continuous with respect to x with a common rank κn for uUB(0, n), independently of (t, u, V), and also that these functions satisfy an inequality of the form |ϕ(t, x, u, V)| ≤ αn + κn|x| for all (t, x, u), uU ∩  B(0, n), all V. Also, λj(·, ·, ·) is continuous and for all n. Define u(t, ω) to be admissible if u(t, ω) ∈ U, for all (t, ω), sup t,ω|u(t, ω)| < , and if u(t, ω) is nonanticipative and is separately piecewise continuous in each real variable on which it depends. Then, corresponding to u(t, ω), for each ω, there exists a nonanticipative function x(t, ω) (also called admissible), which is piecewise continuous in t, satisfying (A.2) and (A.1), for t ∈ (τj, τj+1), j = 0,1, … (x(t, ω) becomes piecewise continuous in each real variable on which it depends). If there are terminal conditions in the problem, u(t, ω) and x(t, ω) are called admissible if also the terminal conditions are satisfied. For pairs u(t, ω) and x(t, ω) to be called admissible, we can allow additional restrictions; namely, for given sets Qj(t) ⊂ n, t ∈ (0, T), j = 0,1, …, it can be required that x(t, ω) has to belong to Qj(t) a.s. for t ∈ (τj, τj+1).

    HJB Equations. The following condition is sufficient for optimality. (Below, j = 0,1, 2, … indicates the number of jumps that have already happened).

    Theorem 1. Assume that there exist functions J(s, y, j) and controls uj(t, x), j = 0,1, 2, …, with J(T, y, j) = h0(y), C1 in (s, y) ∈ [0, T] × Rn, j = 0,1, 2, …, such that the functions J(s, y, j), for all (s, y), satisfy the HJB equation

    ()
    and assume that each uj(s, y) yields maximum in this equation. Moreover, assume that a solution x*(t, ω) of , with jumps given by (A.2), exists such that . Assume also that, for some constants aJ and bJ, |J(s, y, j)| ≤ aJ + bJ|y| for all (s, y, j). Then, is optimal.

    See pages 147 and 155 (and for a proof, see page 168) in [6]. (Formally, we only need to assume that the nonanticipative function u*(t, ω) is measurable and bounded.)

    In case of restrictions of the form x(t, ω) ∈ Qj(t), we must assume that x*(t, ω) satisfies these restrictions, in this case J(s, y, j) needs to only be C1 in a neighborhood of each point in , with J(s, y, j) C0 in cl (cl = closure). Then, (A.4) needs to hold only at each point in .

    If there are terminal conditions in the problem, replace J(T, x, j) = h0(x) by
    ()
    for all admissible x(·, ·).

    If h0 ≡ 0, [0, T] is replaced by [0, ) and if f0 = eβtf0(x, u), g0 = eβtg0(x, V), β > 0, and g, λj, and f are independent of t, and given a sequence uj(x), then a sufficient condition for optimality of u*(t, ω) (catching up optimality in case infinite values of the criterion appear) is as follows.

    Theorem 2. For j = 0,1, 2, …, assume that some functions uj(y) and (C1 in y) exist, such that satisfy the current value HJB equation

    ()
    for j = 0,1, 2, …, with uj(y) yielding maximum. The functions g0, f0, f, g, and are postulated to be continuous and Lipschitz continuous in x with rank κn for uUB(0, n) and all V and to satisfy a growth conditions of the form |ϕ(x, u, V)| ≤ αn + κn|x| for all V, x, uUB(0, n). Finally, for each n. (It suffices to assume that, for any admissible pair (x(t, ω), u(t, ω), for t in bounded intervals, sup t,ωλj(x(t, ω), u(t, ω)) < .) Assume that a solution x*(t, ω) of , with jumps given by (A.2), exists. Assume also that, for some constants aJ and bJ, for all (y, j). Define u*(t, ω): = and assume that sup t,ω|u*(t, ω)| < for t in bounded intervals. (If restrictions x(t, ω) ∈ Qj, t ∈ (τj, τj+1) are present, one must have x*(t, ω) ∈ Qj; in this case, it suffices that is C1 in some open set containing Qj, with in cl Qj, and (A.6) needs to hold only in Qj.) Then, u*(t, ω) is optimal, provided
    ()
    for all admissible x(·, ·).

    (This is the infinite horizon current value form of (A.4); compare (3.71) and (3.69), page 150 in [6].)

    The Extremal Method. The so-called extremal method (see page 117 and pages 126–130 in [6]) now to be described yields solutions satisfying a maximum principle (necessary condition). Now, it is assumed that sup jλj < and that λj does not depend on (t, x, u). To the assumptions above on f0, f, g0, g, and h0, we now add the assumption that these five functions have derivatives with respect to x that are continuous in (t, x, u, V). Let I be the identity matrix and let H = f0 + pf. For arbitrary (s, y), we seek controls u(t; s, y, j) and solutions x(t; s, y, j), j = 0,1, …, u(t; s, y, j) yielding maximum of H(t, x(t; s, y, j), u, p(t; s, y, j)), where each p(t) = p(t; s, y, j) is a solution on [s, T] of
    ()
    with end condition and where each x(t) = x(t; s, y, j) is a solution of with initial condition x(s; s, y, j) = y (j = the  numberofjumpsthathashappened).
    Frequently, to find x(t; s, y, j), u(t; s, y, j), and p(t; s, y, j), as in deterministic problems, one first seeks a control that maximizes H(t, x, u, p), and then one solves the equations and
    ()
    with side conditions
    ()
    Evidently, to solve the equations and (A.9) to obtain x(t) = x(t; s, y, j) and p(t) = p(t; s, y, j), knowledge of p(t; s, y, j + 1) is needed. If and only if λN+1 = λN+2 = ⋯ = 0 for some N, a backward recursion is possible, finding first x(·, ·, N) and p(·, ·, N), then x(·, ·, N − 1), p(·, ·, N − 1), x(·, ·, N − 2) and p(·, ·, N − 2), and so on. When x(t; s, y, j) and p(t; s, y, j) are known, .
    The entities x(t; s, y, j) and u(t; s, y, j) are used to piece together a nonanticipative pair x*(t, ω), u*(t, ω) defined for t∈[0, T]. On [0, τ1], define u*(t, ω) = u(t): = u(t; 0, x0, 0) and x*(t, ω) = x(t): = x(t; 0, x0, 0). On (τ1, τ2],
    ()
    Generally, when x*(t, ω) = x(t, τ1, V1, …, τj−1, Vj−1) has been defined on (τj−1, τj], then on (τj, τj+1]
    ()
    So and (where, for j = 0, the u- and x-functions in the sums are simply u(t) and x(t)). Let . (These entities in fact satisfy a maximum principle not stated here: see page 126 and page 131 in [6].) (If there are restrictions of the form x(t − , ω) ∈ Qj(t) in the problem, one only needs to construct solutions for (s, y) ∈ Qj(s), for j > 0, for j = 0, only for (s, y) = (0, x0). Eventually, one has to test if x*(t − , ω) ∈ Qj for t ∈ (τj, τj+1.)

    So far, we have rendered conditions pertaining to a free terminal state problem. If there are hard terminal conditions (hard = holding a.s.) or soft terminal conditions (holding in expectation), then certain transversality conditions have to be satisfied by the p(t, ω)-functions; we may have to allow H-functions p0f0 + pf, where p0 ∈ {0,1}, and for hard terminal conditions, certain controllability conditions have to be satisfied. Such conditions are omitted here, since our main interest is sufficient conditions; what we then need is a sort of transversality condition at the terminal time as given by (A.13), working for all types of terminal restrictions.

    Assume now that entities x(t; s, y, j), u(t; s, y, j), p(t; s, y, j), x*(t, ω), and u*(t, ω) have been constructed by means of the extremal method described above, and define . Then, the following sufficient condition holds (here p0 = 1).

    Theorem 3. If all the entities h0(x), g0(t, x, V), , and p(t; s, y, j)g(t, x, V) are concave functions of x, then the control u*(t, ω) is optimal in the free end case. (In case of restrictions of the form x(t − , ω) ∈ Qj(t), again x*(t, ω) must of course satisfy them, Qj(t) must be convex, and, for each j, the just-mentioned concavity needs to hold only on Qj(t). If U is convex, then concavity of (x, u) → H(t, x, u, p(t; s, y, j)) in Qj(t) × U suffices for concavity of in Qj(t).) In case one has terminal conditions on the solutions, then u*(t, ω) is optimal, provided also that the following condition holds:

    ()
    that is, for all x(·, ω) satisfying the terminal conditions.

    In the end constrained case, p(T, ω) is perhaps different from h0x(x*(T, ω)). See proof of Theorem  2, page 30 in [7]. Note that (A.13) is implied by standard transversality conditions for soft end constraints that are omitted here; see (3.62), page 146 in [6] for the case h0 ≡ 0.

    When T = , these concavity conditions again ensure optimality (perhaps catching up optimality) provided (see [7], page 32)
    ()
    Boundedness, Lipschitz—and growth—conditions on u*, g0, f0, f, and g are required, as in Theorem 2.

      The full text of this article hosted at iucr.org is unavailable due to technical difficulties.