Many optimization problems in mathematics and statistics are concerned with a quantity, or procedure, that yields some optimum value rather than the value itself. Frequently, they are referred to as searching problems. In this paper we consider a problem of this nature, which is described as follows. An object to be found is located in one of R locations. A prior probability pi that the object is in i is given (∑ip i = 1), along with an overlook probability αi that if the object is in i, it is not found there on a given inspection of i (0 α_i 1, i = 1, \cdots, R). The R locations--or boxes--can be searched one at a time, and it is assumed that all outcomes are independent, conditional on the location of the object and the inspection procedure used. Consequently, if the object is in i, the realization of the `first success'--the detection of the object--follows a geometric distribution with parameter 1 - αi. A search procedure δ = (δ1, δ2, ⋯) is a sequence indicating the location to be inspected at each stage of the search, and one is interested in a δ that in some sense is optimal. In his notes on dynamic programming, Blackwell (see [3]) has shown that if, in addition, an inspection of i costs ci, the procedure minimizing the expected searching cost is a one-stage procedure, which instructs the searcher to inspect at each stage that box for which the ratio of the current detection probability, and the cost of an inspection there, is greatest. It is assumed here that ci = 1, i = 1, ⋯, R. If δ* = (δ* 1, δ* 2, ⋯) denotes an optimal one-stage procedure and δ* n = j, then the detection probability for j on the nth inspection is \begin{equation*}\tag(1.1) p_jα_j^{m(j,n-1,δ^\ast)} (1 - α_j) = \max_i {p_iα^{m(i,n-1,δ^\ast}_i (1 - α_i)}\end{equation*} where m(i, n, δ) is the number of inspections of i among the first n inspections made using δ. In the next section δ* is shown to be strongly optimal in the sense that \begin{equation*}\tag{1.2}P\lbrack N n |δ^\ast\rbrack \leqq P\lbrack N n | δ\rbrack,\end{equation*} for each n = 0, 1, 2, ⋯, and every search procedure δ. (Here N denotes the (random) number of inspections required to find the object.) The use of procedure δ* thus ensures the greatest chance of finding the object within any fixed number of inspections, or, according to B. O. Koopman [2], it "optimally allocates the available search effort Φ" for any number Φ of inspections. As a consequence of (1.2), δ* also minimizes the expected `cost' of the search (since E\lbrack N \mid δ\rbrack = \sum^\infty_{n = 0} P\lbrack N n \mid δ\rbrack), which is Blackwell's result for this case. Among the many interesting problems which arise in connection with this searching problem is the question of periodic features of the optimal procedure. Staroverov [5] and Matula [3] concerned themselves with this question, which was virtually answered by Matula who found necessary and sufficient conditions ensuring ultimate periodicity. We attempt to answer a different question in Section 3, where the same search problem is considered with the modification that \sum_i p_i = 1 - q 1. That is, the object is in one of R + 1 locations, but searching is permitted only among the first R. (For instance, the (R + 1)st location could be the rest of the world.) In this problem every search procedure has positive probability of never terminating, making it necessary to couple a stopping rule s (integer-valued) with any procedure δ. A loss function is defined by imposing on the searcher a penalty cost c(0), payable when searching stops if the object has not been found. Thus, one either pays the cost of unsuccessful inspection plus the penalty cost, or simply the cost of inspection if the object is found prior to stopping. A procedure (δ, s) is sought which minimizes the expected cost to the searcher, i.e. which yields the Bayes risk. Such a procedure exists and is shown to be (δ*, s*), where δ* is a strongly optimal procedure satisfying (1.2). The determination of S* is the more difficult problem, however, and the main part of Section 3 is devoted to it. Instead of s* we consider the problem of finding the Bayes stopping region SB, that set of posterior probabilities p = (p1, ⋯ pR) for which the Bayes risk equals the penalty cost c. Regions SL SU are given which bound SB in the sense that S_L \subset S_B \subset S_U. In Section 4, it is shown that for sufficiently large C, SL and SU `differ' by at most R inspections taken according to δ*, and no more than one additional inspection is required in each of the R boxes. If (δ*, sU) is a procedure which searches according to δ*, but stops once the posterior distribution p ε SU, a constant independent of c, is exhibited which bounds the difference between the Bayes risk and the expected risk using (δ*, sU).
CITATION STYLE
Chew, M. C. (1967). A Sequential Search Procedure. The Annals of Mathematical Statistics, 38(2), 494–502. https://doi.org/10.1214/aoms/1177698965
Mendeley helps you to discover research relevant for your work.