Optimal Oil-Owner Behavior in Piecewise Deterministic Models
Abstract
Six simple piecewise deterministic oil production models from economics are discussed by using solution tools that are available in the theory of piecewise deterministic optimal control problems.
1. Introduction
Six simple piecewise deterministic models for optimal oil-owner behavior are presented. Their central property is sudden jumps in states. The aim of this paper is to show in admittedly exceedingly simple models how available tools for piecewise deterministic models, namely, the HJB equation and the maximum principle, can be used to solve these models analytically. We are looking for solutions given by explicit formulas. That can only be obtained if the models are simple enough. The models may be too simple to be of much interest in themselves, but they can provide some intuition about features optimal solutions may have in more complicated models.
Piecewise deterministic models have been used a number of times in economic problems in the literature; some few scattered references are given that contain such applications [1–4]. I have not been able to find references directly concerned with piecewise deterministic oil production problems. For different probability structures, and for discrete time, a host of related problems has been discussed in the literature; references to such literature have been left out, with one exception. Problems of control of jump diffusions, see [5], encompass piecewise deterministic problems, and some problems appearing in [5] are related to the ones discussed below. A classic reference to piecewise deterministic control problems is [2].
In all models below, an unbounded number of jumps in the state can occur at times 0 < τ1 < τ2 < τ3 < ⋯, and, when τj is given, τj+1 is exponentially distributed in [τj, ∞) (all τi − τi−1 independent). Sometimes, the size of the jumps is influenced by stochastic variables Vj. Let ω = (τ1, V1, τ2, V2, …). At time t, we imagine that the control values chosen can be allowed to depend on what has happened, that is, on τi, Vi for τi < t, but not on future events, that is, τi, Vi for which τi > t. Such controls (written u(t, ω)) are called nonanticipative. Corresponding state solutions denoted by x(t, ω) are then also nonanticipative. (A general set-up, with further explanations, is given in Appendix.) Frequently below, x(t, ω) will be the wealth of the oil owner. In infinite horizon economic models, the weakest terminal condition that is natural to apply is an expectational no-Ponzi-game condition; namely, liminf T→∞E[e−kTx(T, ω)] ≥ 0, where k is the discount factor. Note that some stronger conditions will be used in some of the models presented in the sequel.
2. Model 1
The interpretation of the model is that u is consumption rate, αi is the size (in dollars) of an oil field found at time τi, x is wealth, and kx is interest. Oil fields are sold immediately after discovery.
Let us solve problem (1)–(3) by using the extremal method (see the appendix). Now, with H = e−δtuγ/γ + p(kx − u), solving the first-order condition for maximum of H, we get that u = eδt/(γ−1)p1/(γ−1) maximizes the Hamiltonian.
Now, for , ρ : = (k − δ)/(γ − 1) > 0, we get u(t) = eδt/(γ−1)(p(t)) 1/(γ−1) = be−ρt. Because H is concave in (x, u), ax is linear, and αi is independent of x, sufficient conditions based on concavity (see Theorem 3 in Appendix) then give us that the control u = be−ρt is an optimal control for problem (1)–(3). The optimal control is independent of λ and the αi’s.
Write x*(t) = ektΦi(t, ω), for t ∈ (τi, τi+1), where .
We are now going to replace the bequest function ax(T) by the terminal constraint Ex(T, ω) ≥ 0. For this purpose, we are now going to vary a and hence also and β. For t = T, consider the expectation of Φi(t, ω). We will not give an explicit formula for this expectation, but we mention that it can be calculated in two steps, first given that i jumps have happened in [0, T] and then the expectation with i being stochastic (the rule of double expectations is then used). The expectation of the sum containing αj as well as x0 is positive, and the term in front of β is negative. There thus exists a unique positive value βT of β and, hence, of a (denoted by aT) such that EΦi(T, ω) = 0; that is, Ex*(T, ω) = 0.
If we drop the term ax(T) in the criterion but add the terminal condition Ex(T, ω) ≥ 0, the free end optimality obtained in the original problem for a = aT evidently means optimality of in the end constrained problem (the found value aT appears in u, though now not in the criterion). (Alternatively, we could use the sufficient conditions for end constrained problems in Theorem 3 in Appendix to obtain optimality in the present end constrained problem. This would require us to check condition (A.13), which is easily done.)
Then, .
We may assume that the jumps are stochastic, that is, that x(τi+) − x(τi−) = Vi, where take values in a common bounded set and are independent, and are independent of the τi’s. If we then assume that ∑E(Vi) < ∞, the solution in this problem is the same as the one above for αi = E(Vi).
Note that we must assume that we have a deal with the bank in which our wealth is placed that it accepts the above behavior. That is, before time 0, we have got an acceptance for the possibility of operating with this type of admissible solutions, which means that only in expectation we leave a wealth in the bank ≥0. In actual runs, sometimes we leave in the bank a positive wealth (that the bank gets), and for other runs a negative wealth (debt) that the bank has to cover.
3. Model 2
(This model Is related to exercise 4.1 in Øksendal and Sulem [5].)
In contrast to problem (1)–(3), now the jumps are linearly dependent on x(τi−). To defend such a feature, one might argue that the richer we are, the more we are able to generate large jumps (the jump may actually represent a collection of oil finds). On the other side, we will assume that such jumps occur with smaller and smaller intensities.
Now, assume first, for some k*, that . Then, is the optimal value in the problem with no jumps, so . Given , (14) determines , . It must be the case that Aj is decreasing. Given the same start point x, the optimal value when j + 1 jumps have occurred already at time 0 must be smaller than the optimal value when j jumps have occurred at time 0; in the former problem, prospective jumps have smaller probabilities for happening. (Let us show by backwards induction that Aj, j = 0,1, 2, …, k, is nonincreasing. Evidently, Ak = Ak+1. Assume by induction that Aj ≥ Aj+1. If Aj−1 < Aj, then, by (15), + λj(1 + α) γAj/Aj−1>−δ + γk − λj+1 + λj+1(1 + α) γAj/Aj−1 > −δ + γr − λj+1+λj+1(1 + α) γAj+1/Aj = . Because γ − 1 < 0 and z → (z) 1/(γ−1) is decreasing, a contradiction is obtained.) As we will let k* vary, denote the sequence defined by (14) for by .
Now, , so . In fact, as is easily seen, this holds for all j: . (Compare the optimal values in the case where jump j occurs at (before) t = 0 in the two problems where λm = 0, m ≥ k* + 1, and where λm = 0, m ≥ k* + 2.) Assume now that λj > 0 for all j and let . (It is shown below that, for some , , for all k*, j.) Then, lim j→∞Aj = A(0), and the Aj’s satisfy (14).
Let x*(t, ω) be the solution for u ≡ 0. By (11), , where . Choose the smallest j* such that . If j* jumps occur at once at 0 (λj = ∞, j ≤ j*), while further jumps occur with intensity , then EΓ would equal . Hence, . For any admissible solution x(t, ω), x(t, ω) ≤ x*(t, ω), for all (t, ω); hence, for t = T, 0 ≤ E[e−δTAj(x(T, ω)) γ] ≤ when T → ∞. So (A.7) in the appendix is satisfied. Hence, sufficient conditions hold (see Theorem 2 in Appendix) and , j = 0,1, …, are optimal. That is, if and x*(t, ω) = x0(1 + for t ∈ (τi, τi+1) (see (11)), then u*(t, ω) = v*(t, ω)x*(t, ω) is optimal (x*(t, ω) evidently satisfies x*(t, ω) > 0, for all t). Hence, . (In the appendix, for the sufficiency of the HJB equation to hold, it is required that u*(t, ω) is bounded if T < ∞ (for T = ∞, we need boundedness for t in all bounded intervals). This is not the case here. But we could have replaced u by vx, v being the control. Assume that we require v(t, ω) to be bounded in the above manner. Now, v*(t, ω) is so bounded, and it is then optimal in the set of such bounded v(·, ·)’s.)
We can show that the Aj’s are increasing in α and in each λi, i > j. The simplest argument is that this must be so, when we now know that Ajyγ is the optimal value function after j jumps at (before) t = 0.
If all λj = λ < (δ − kγ)/α, then, using (16), when t ∈ (τi, τi+1), where 0 < ρ* = ρ + [λ − λ(1 + α) γ]/(1 − γ) < ρ = (δ − k)/(1 − γ). (The inequality αλ < δ − kγ implies δ − kγ + λ − λ(1 + α) γ > δ − kγ + λ − λ(1 + α) > 0, so A(λ) > 0. In particular, . The above calculations show that in the problem where , for j > j*, the optimal value functions are , where for j ≥ j*, given by backward induction using (14) for j < j*, again . Evidently, . Now, for j ≥ j* and, from (16), it follows that when j → ∞.) Now, relate the present case to what happened in Model 1. As and ρ* < ρ, in each interval (τi, τi+1), (d/dt)u*(t) > d/dt(be−ρt) (be−ρt is the optimal control in Model 1); moreover, when t passes each τi u*(t) changes by a factor (1 + α).
4. Model 3
Next, maximum of the Hamiltonian H = e−δtuγ/γ + p(ay + z − u) is obtained for u satisfying the first order condition uγ−1e−δt = p, so u = (eδtp) 1/(γ−1) = Kβe−κt, where κ = (a − δ)/(γ − 1) > 0, β = 1/(γ − 1).
If T = ∞, the end condition required for using the sufficient condition in Theorem 2 in the appendix is liminf T→∞E[p(T, ω)(y(T, ω) − y*(T, ω))] ≥ 0. The condition lim TE[p(T)y*(T, ω)] = 0 determines K; it now reduces to the condition y0 + v/a = Kβ/(a + κ). Moreover, we now have optimality among all triples u(t, ω), y(t, ω), z(t, ω) for which liminf T→∞E[p(T)y(T, ω)] ≥ 0. (This inequality means that the next-to-last inequality is satisfied.) Note that Kβ and so the control, quite expectably, are increasing in v = EV.
5. Model 4
To show the existence of a solution of (26) and (27), note that if we put A2 = A1 in (26), then by dividing by A1, we get , and κ1(A1) = 0 yields . Similarly, if we put A1 = A2 in (27), we get , and κ2(A2) = 0 yields . Now, . Denote the right-hand sides in (26) and (27) by ϕ(A1, A2) and ψ(A2, A1), respectively. Both are increasing in the second place and decreasing in the first place. When , , and (the only difference in ψ(A, A) and ϕ(A, A) is the third term in the formulas). Thus, ψ(A2, A1) = 0 can be uniquely solved with respect to A2 for A1 in , denote the solution A2(A1), it is evidently increasing in A1 and , . Moreover, = ψ(A1, A1)⇒A2(A1) > A1. Consider the function β(A1): = ϕ(A1, A2(A1)) > ϕ(A1, A1). Now, and , so exists for which 0 = β(A1) = ϕ(A1, A2(A1)). So this and , A2 > A1, satisfy both (27) and (26).
For any admissible y(t, ω), , where y*(t) is the solution of (i.e., y*(T) = y0e2t), so, for all admissible y(t, ω), 0 ≤ lim T→∞E[e−δTAj(y(T, ω)) γ] ≤ . Hence, (A.7) in the appendix holds; the sufficient conditions in Theorem 2 are satisfied, and if u = y(Aj) 1/(γ−1) when Vi = j ∈ {1,2}, then u is optimal. (Hence, if + and y*(t, ω) is the solution of , then u*(t, ω) = v(t, ω)y*(t, ω) is optimal.)
Define A* = [(δ − 3γ/2)/(1 − γ)] γ−1 (i.e., A* satisfies 0 = −δA + Aγ/(γ−1) + 3γA/2 − γAA1/(γ−1)). When λ = 0 then , and one may show that when λ > 0 decreases, A2 increases towards while A1 decreases towards . When λ increases to infinity, A1 increases to A*, and A2 decreases to A*. In fact, and . To prove convergence to A*, note first that A1, A2 are bounded uniformly in λ. If A1 and A2 did not converge to a common number, say A*, when λ → ∞, the two last terms in (27) and (26) would blow up to infinity for (certain) large values of λ, while the remaining terms would be bounded, a contradiction. Summing the two equations in (27) and (26) gives . Hence, A* satisfies ; that is, A* = A*.
Two very simple models discussed below contain the feature that the owner can influence the chance of discovery but that it is costly to do so. In the first one, the intensity of discoveries λ is influenced by how much money is put into search at any moment in time; in the second one, it is costly buildup of expertise that matters for the intensity of discoveries.
6. Model 5
Still, trivially condition (A.7) below holds, sufficient conditions in Theorem 2 in the appendix are satisfied and on [τj, τj+1), j = 0,1, 2, …, are optimal. (Uniqueness of the aj’s follows from optimality.)
7. Model 6
All fields found are of the same size α. We again imagine that fields are sold immediately when they are found or that they are produced over a fixed period of time, with a fixed production profile. Let income from a field discounted back to the time of discovery be equal to g0 : = βα, for a given β > 0. The state y(t) is the amount of expertise available for finding new fields, built up over time according to , where u is money per unit of time spent on building expertise.
8. Comparisons
In Models 1, 2, 5, and 6, oil finds are made at stochastic points in time; in Models 3 and 4, it is the price of oil that changes at stochastic points in time. In Model 1, we operate with the constraint Ex(∞, ω) ≥ 0 (for T = ∞, a = 0), where x is the oil-owners’ wealth. Here, for some runs, x(∞, ω) can be negative (β > x0) and, for other runs, positive. In Model 2, we required x(t, ω) > 0 for all t, all ω. (The results in that model would have been the same if we had required only x(∞, ω) ≥ 0 for all ω.) In Model 2, the optimal control comes out as stochastic and not deterministic as in Model 1. Moreover, as a comment in Model 2 says, as a function of time, the optimal control decreases more rapidly in Model 1 as compared to Model 2. The latter feature stems from the fact that, in Model 2, we enhance future income prospects by not decreasing x too fast, because the jump term (the right-hand side of the jump equation) depends positively on x, which is not the case in Model 1.
In Models 3 and 4, the oil price exhibits sudden stochastic jumps. In Model 3, the rate of oil production is constant, but income earned (as well as interest) is placed in a bank after subtraction of consumption. In Model 4, income earned, after subtraction of consumption, is reinvested in the oil firm to increase production. In Model 4, the optimal control is stochastic; it depends on whether the current price is high or low. In Model 3, the control is deterministic, and it depends only on the expectation v of the stochastic price. Consider the case where, in Model 3, the expected price v is zero and a = 3/2. Then, u = Kβe−κt, where κ = (3/2 − δ)/(γ − 1) and Kβ = y0(3/2 + κ). We saw in Model 4 that when the intensity of jumps λ is very high, A1≃A2≃A*, where A* = [δ − 3γ/2)/(1 − γ)] γ−1; we hardly pay attention to what the current price is, because it changes so frequently. Now u = y(Aj) 1/(γ−1) when the current price is j( = z), or u≃y(A*) 1/(γ−1) when λ is large. When λ is large (so z switches very frequently between 1 and 2), the stochastic path of the equation most often is very close to the deterministic path of = (3/2)y − [(δ − 3γ/2)/(1 − γ)]y = −κy. The latter equation has the solution y = y0e−κt, and the corresponding u equals = (3/2)y0e−κt + κy0e−κt = κy0e−κt = y0(3/2 + κ)e−κt, the same control as that obtained in Model 3 in the case a = 3/2, v = 0.
In the extremely simple Models 5 and 6, the frequency of oil finds is not fixed but influenced by a control. In Model 5, the current frequency (or intensity) is determined by how much money is put into search at that moment in time. In the simplest case considered in Model 5, a find today does not influence the possibility of making equally sized discoveries tomorrow. Then it, is not unreasonable that the optimal control (which equals (μg0) 1/κ) is independent of the discount rate δ but dependent on the fixed value of the finds g0 = βα. (Here, actually α could be the expected size of a find, in both Model 5 and Model 6; we could have had the sizes of the finds being independently stochastic, with α being the expected value of the sizes.) In Model 6, it is the ability to discover oil that is built up over time, so with greater impatience (higher δ), we should expect less willingness of devoting money to increase this ability, and this shows up in the formula u = (μg0/δ) 1/κ.
Acknowledgment
The author is very grateful for useful comments received from a referee that made it possible to improve the exposition and remove some typographical errors.
Appendix
Assume that the five given functions f0, f, g0, g, and h0 are continuous and Lipschitz continuous with respect to x with a common rank κn for u ∈ U∩B(0, n), independently of (t, u, V), and also that these functions satisfy an inequality of the form |ϕ(t, x, u, V)| ≤ αn + κn|x| for all (t, x, u), u ∈ U ∩ B(0, n), all V. Also, λj(·, ·, ·) is continuous and for all n′. Define u(t, ω) to be admissible if u(t, ω) ∈ U, for all (t, ω), sup t,ω|u(t, ω)| < ∞, and if u(t, ω) is nonanticipative and is separately piecewise continuous in each real variable on which it depends. Then, corresponding to u(t, ω), for each ω, there exists a nonanticipative function x(t, ω) (also called admissible), which is piecewise continuous in t, satisfying (A.2) and (A.1), for t ∈ (τj, τj+1), j = 0,1, … (x(t, ω) becomes piecewise continuous in each real variable on which it depends). If there are terminal conditions in the problem, u(t, ω) and x(t, ω) are called admissible if also the terminal conditions are satisfied. For pairs u(t, ω) and x(t, ω) to be called admissible, we can allow additional restrictions; namely, for given sets Qj(t) ⊂ ℝn, t ∈ (0, T), j = 0,1, …, it can be required that x(t, ω) has to belong to Qj(t) a.s. for t ∈ (τj, τj+1).
HJB Equations. The following condition is sufficient for optimality. (Below, j = 0,1, 2, … indicates the number of jumps that have already happened).
Theorem 1. Assume that there exist functions J(s, y, j) and controls uj(t, x), j = 0,1, 2, …, with J(T, y, j) = h0(y), C1 in (s, y) ∈ [0, T] × Rn, j = 0,1, 2, …, such that the functions J(s, y, j), for all (s, y), satisfy the HJB equation
See pages 147 and 155 (and for a proof, see page 168) in [6]. (Formally, we only need to assume that the nonanticipative function u*(t, ω) is measurable and bounded.)
In case of restrictions of the form x(t, ω) ∈ Qj(t), we must assume that x*(t, ω) satisfies these restrictions, in this case J(s, y, j) needs to only be C1 in a neighborhood of each point in , with J(s, y, j) C0 in cl (cl = closure). Then, (A.4) needs to hold only at each point in .
If h0 ≡ 0, [0, T] is replaced by [0, ∞) and if f0 = e−βtf0(x, u), g0 = e−βtg0(x, V), β > 0, and g, λj, and f are independent of t, and given a sequence uj(x), then a sufficient condition for optimality of u*(t, ω) (catching up optimality in case infinite values of the criterion appear) is as follows.
Theorem 2. For j = 0,1, 2, …, assume that some functions uj(y) and (C1 in y) exist, such that satisfy the current value HJB equation
(This is the infinite horizon current value form of (A.4); compare (3.71) and (3.69), page 150 in [6].)
So far, we have rendered conditions pertaining to a free terminal state problem. If there are hard terminal conditions (hard = holding a.s.) or soft terminal conditions (holding in expectation), then certain transversality conditions have to be satisfied by the p(t, ω)-functions; we may have to allow H-functions p0f0 + pf, where p0 ∈ {0,1}, and for hard terminal conditions, certain controllability conditions have to be satisfied. Such conditions are omitted here, since our main interest is sufficient conditions; what we then need is a sort of transversality condition at the terminal time as given by (A.13), working for all types of terminal restrictions.
Assume now that entities x(t; s, y, j), u(t; s, y, j), p(t; s, y, j), x*(t, ω), and u*(t, ω) have been constructed by means of the extremal method described above, and define . Then, the following sufficient condition holds (here p0 = 1).
Theorem 3. If all the entities h0(x), g0(t, x, V), , and p(t; s, y, j)g(t, x, V) are concave functions of x, then the control u*(t, ω) is optimal in the free end case. (In case of restrictions of the form x(t − , ω) ∈ Qj(t), again x*(t, ω) must of course satisfy them, Qj(t) must be convex, and, for each j, the just-mentioned concavity needs to hold only on Qj(t). If U is convex, then concavity of (x, u) → H(t, x, u, p(t; s, y, j)) in Qj(t) × U suffices for concavity of in Qj(t).) In case one has terminal conditions on the solutions, then u*(t, ω) is optimal, provided also that the following condition holds:
In the end constrained case, p(T, ω) is perhaps different from h0x(x*(T, ω)). See proof of Theorem 2, page 30 in [7]. Note that (A.13) is implied by standard transversality conditions for soft end constraints that are omitted here; see (3.62), page 146 in [6] for the case h0 ≡ 0.