Article ID: | iaor20091253 |
Country: | United States |
Volume: | 2005 |
Issue: | 2 |
Start Page Number: | 97 |
End Page Number: | 118 |
Publication Date: | May 2005 |
Journal: | Journal of Applied Mathematics and Stochastic Analysis |
Authors: | Benkherouf Lakdere, Pitts Susan |
Keywords: | energy, markov processes, programming: dynamic |
This paper is concerned with optimal strategies for drilling in an oil exploration model. An exploration area contains n1 large and n2 small oilfields, where n1 and n2 are unknown, and represented by a two-dimensional prior distribution n. A single exploration well discovers at most one oilfield, and the discovery process is governed by some probabilistic law. Drilling a single well costs c, and the values of a large and small oilfield are v1 and v2 respectively, v1>v2>c>0. At each time t= 1,2,…, the operator is faced with the option of stopping drilling and retiring with no reward, or continuing drilling. In the event of drilling, the operator has to choose the number k, 0≤k≤m (m fixed), of wells to be drilled. Rewards are additive and discounted geometrically. Based on the entire history of the process and potentially on future prospects, the operator seeks the optimal strategy for drilling that maximizes the total expected return over the infinite horizon. We show that when n≻n′ in monotone likelihood ratio, then the optimal expected return under prior n is greater than or equal to the optimal expected return under n′. Finally, special cases where explicit calculations can be done are presented.