Positivity

DOI 10.1007/s11117-014-0295-3

Positivity

A prophet inequality for Lp -bounded dependent random variables

Adam Osekowski

Received: 21 March 2014 / Accepted: 21 May 2014

© The Author(s) 2014. This article is published with open access at Springerlink.com

Abstract Let X = (Xn)n>1 be a sequence of arbitrarily dependent nonnegative random variables satisfying the boundedness condition

where t > 0, 1 < p < ro are fixed numbers and the supremum is taken over all finite stopping times of X. Let M = E supn Xn and V = supT EXT denote the expected supremum and the optimal expected return of the sequence X, respectively. We establish the prophet inequality

and show that the bound on the right is the best possible. The proof of the inequality rests on Burkholder's method and exploits properties of certain special functions. The proof of the sharpness is somewhat indirect, but we also provide an indication how the extremal sequences can be constructed.

Keywords Prophet inequality • Optimal stopping • Bellman function • Best constants

Research partially supported by the NCN grant DEC-2012/05/B/ST1/00412. A. Osekowski (B)

Department of Mathematics, Informatics and Mechanics, University of Warsaw, Banacha 2, 02-097 Warsaw, Poland e-mail: ados@mimuw.edu.pl

Published online: 11 June 2014 Birkhäuser

sup E X p < t,

Mathematics Subject Classification (2010) Primary 60G40 ■ 60E15; Secondary 62L15

1 Introduction

The purpose of the paper is to establish a sharp estimate between the expected supre-mum of a sequence X = (Xn)n>1 of Lp-bounded random variables and the optimal expected return (i.e., optimal stopping value) of X. Such comparisons are called "prophet inequalities" in the literature and play a distinguished role in the theory of optimal stopping, as evidenced in the papers of Allaart [1], Allaart and Monticino [2], Assaf et al. [4,5], Boshuizen [3,6], Hill [8,9], Hill and Kertz [10-12], Kennedy [13,14], Kertz [15,16], Krengel, Sucheston and Garling [17-19], Tanaka [24,25] and many others.

We start with the necessary background and notation. Assume that X = (Xn)n>1 is a sequence of (possibly dependent) random variables defined on the probability space F, P). With no loss of generality, we may assume that this probability space is the interval [0, 1] equipped with its Borel subsets and Lebesgue measure. Let (Fn )n>1 = (o(xX1, X2,..., X n))n>1 be the natural filtration of X. The problem can be generally stated as follows: under some boundedness condition on X, find universal inequalities which compare M = E supn Xn, the expected supremum of the sequence, with V = supr EXT, the optimal stopping value of the sequence; here t runs over the class T of all finite stopping times adapted to (Fn)n>1. The term "prophet inequality" arises from the optimal-stopping interpretation of M, which is the optimal expected return of a player endowed with complete foresight; this player observes the sequence X and may stop whenever he wants, incurring a reward equal to the variable at the time of stopping. With complete foresight, such a player obviously stops always when the largest value is observed, and on the average, his reward is equal to M. On the other hand, the quantity V corresponds to the optimal return of the non-prophet player.

Let us mention here several classical results in this direction; for an excellent exposition on the subject, we refer the interested reader to the survey by Hill and Kertz [12]. The first universal prophet inequality is due to Krengel, Sucheston and Garling [17,18]: if Xi, X2,... are independent and nonnegative, then

M < 2V

and the constant 2 is the best possible. The next result, coming from [8] and [10], states that if X1, X2,... are independent and take values in [0, 1], then

M - V < -4

M < V - V2.

Both estimates are sharp: equalities may hold for some non-trivial sequences X. Analogous inequalities for other types of variables X1, X2,... (e.g., arbitrarily-dependent

and uniformly bounded, i.i.d., averages of independent r.v.'s, exchangeable r.v.'s, etc.), as well as for other stopping options (for instance, stopping with partial recall, stopping several times, using only threshold stopping rules, etc.) have been studied intensively in the literature and have found many interesting applications. We refer the reader to the papers cited at the beginning.

The motivation for the results obtained in this paper comes from the following statement proved by Hill and Kertz [11]: if X1, X2, ... are arbitrarily dependent and take values in [0, 1], then we have the sharp bound

M < V - V log V. (1.1)

There is a very natural problem concerning the Lp-version of this result, where p is a fixed number between 1 and infinity. For example, consider the following interesting question. Suppose that X1, X2, ... are nonnegative random variables satisfying supn EXVp < 1. What is the analogue of (1.1)? Unfortunately, as we will see in Sect. 5 below, there is no non-trivial prophet inequality in this setting. More precisely, for any K > 0 one can construct a sequence X bounded in Lp which satisfies V = 1 and M > K.

We will work under the more restrictive assumption

X1, X2, ... are nonnegative and satisfy sup EXp < t, (1.2)

where t is a given positive number. For instance, this holds true, if the sequence X possesses a majorant $ which satisfies Efp < t. The main result of the paper is the following.

Theorem 1.1 Let t > 0, 1 < p < to be fixed and suppose that X1, X2, ... satisfy (1.2). Then

M < V +-- log — (1.3)

and the inequality is sharp: the bound on the right cannot be replaced by a smaller number.

Note that this statement generalizes the inequality (1.1) of Hill and Kertz: it suffices to take t = 1 and let p go to to to recover the bound. On the other hand, the expression on the right of (1.3) explodes as p I 1, which indicates that there is no prophet inequality in the limit case p = 1.

A few words about the proof. Our approach is based on the following two-step procedure: first we show that it suffices to establish (1.3) under the additional assumption that X is a nonnegative supermartingale; second, we prove that in the supermartingale setting, the validity of (1.3) is equivalent to the existence of a certain special function which enjoys appropriate majorization and convexity properties. In the literature this equivalence is often referred to as Burkholder's method or Bellman function method, and it has turned out to be extremely efficient in numerous problems in probability and analysis: consult e.g. [7,20,21,26] and references therein.

We have organized the paper as follows. In the next section we reduce the problem to the supermartingale setting. Section 3 contains the description of Burkholder's method (or rather its variant which is needed in the study of (1.3) for supermartingales). In Sect. 4 we apply the method and provide the proof of Theorem 1.1. In the final part of the paper we show that there are no interesting prophet inequalities in the case when the variables X 1; X2,... are only assumed to be bounded in Lp.

2 A reduction

In this section we show how to relate (1.3) to a certain inequality for nonnegative supermartingales. Throughout, we use the notation X* = supn>1 Xn and Xm — sup1<n<m Xn for the maximal and the truncated maximal function of X. Recall that V — supTst EXT. We start with the observation that it is enough to deal with finite sequences only (in this paper we say that X is finite, if it is of the form (X1, X2,.XN-1, XN, XN, XN,...) for some deterministic N). This is straightforward: suppose we have successfully established the prophet inequality in this special case, and pick an arbitrary, possibly non-finite X. Then for any fixed N, the truncated sequence XN — (X1, X 2,..., XN-1, XN, XN, XN,...) is finite, inherits (1.2) an dits optimal expected return does not exceed V. Since the function V ^ V+p—j log {yp) is nondecreasing on (0, t1/p], an application of (1.3) gives

E(XN)*< V + log(Ve) .

p - 1 \Vp J

It remains to let N ^ro and use Lebesgue's monotone convergence theorem.

Lemma 2.1 Suppose that X — (Xn )n>1 is an arbitrarily dependent finite sequence of random variables satisfying X1 = 0 and (1.2). Then there is a finite supermartingale Y = (Yn )n>1 adapted to the filtration ofX, which satisfies

Yn > Xn almost surely for all n, (2.1)

P (Y1 = V) = 1 (.2)

sup EYp < t, sup EYt < V. (2.3)

TsT TsT

Note that the additional assumption X1 = 0 is not restrictive at all: we can always replace the initial sequence X1, X2,... with 0, X1, X2,..., and the prophet inequality remains the same. In the proof of the above lemma we will need the notion of essential supremum, a well-known object in the optimal stopping theory. Let us briefly recall its definition, for details and properties we refer the reader to the monographs of Peskir and Shiryaev [22] and Shiryaev [23].

Definition 2.1 Let (Za)asj be a family of random variables. Then there is a countable subset J of J such that the random variable Z — supas j Za satisfies the following two properties:

(i) P(Za < Z) = 1 for each a e I, _ _

(ii) if Z is another random variable satisfying (i) in theplace of Z,then P(Z < Z) = 1.

The random variable Z is called the essential supremum of (Za)aeI.

Proof of Lemma 2.1 We will use some basic facts from the optimal stopping theory; for details, we refer the reader to Chapter I in Peskir and Shiryaev [22]. Suppose that X = (X1, X2,..., XN, XN, XN,...) is a finite sequence and let Y be the Snell envelope of X, i.e., the smallest adapted supermartingale majorizing the sequence (X n)n>i.It is a well-known fact that for each n the variable Yn is given by the formula

Yn = ess supIE(X„ \Fn) : o e 7;J

where Tn denotes the class of all finite adapted stopping times not smaller than n. Since a = n belongs to Tn, the inequality (2.1) is given for free. To show (2.2), observe that Y1 is a constant random variable, since the a-algebra F1 = a(X 1) is trivial. Thus the bound Y1 > V follows directly from the above formula for Y1 and the definition of an essential supremum. On the other hand, for any finite stopping time a we have V > EXa = E(Xa\F1) almost surely, which implies P(V > Y1) = 1, again from the definition of an essential supremum. This gives (2.2). Since the sequence X stabilizes after N steps, so does Y and therefore the second estimate in (2.3) holds true, directly from (2.2) and the supermartingale property. To prove the first bound in (2.3), recall that Y can be alternatively defined by the backward induction

Yn = Xn, Yn = max {Xn, E(Yn+i\Fn)}, n = 1, 2,..., N - 1. This implies that Yn = E(Xtn\Fn), where the stopping time Tn is given by

Tn = inf {k e{n, n + 1,...,N }:Yk = Xk}. Thus, if we fix an arbitrary t e T, then

Yt = Yt an = Yn 1 {t an =n} = 1{t a N=n}E( X tn\Fn),

n=1 n=1

which gives

EYP = e[e( 1{ran =n}Xrn\Fn)P] E[ 1{TAN=n}Xpn

n=1 n=1

= E X p,

where a = ^N=1 1{TAN =n}Tn. We easily check that a is a stopping time: for any 1 < k < N, the event

{o = k}= U ({T A N = n} n {Tn = k}) ^ J ({T A N = n} n {Tn = k})

n=1 n=1

belongs to . Therefore, the boundedness assumption (1.2) implies EYT < t, as desired.

Therefore, it suffices to establish the inequality (1.3) under the additional assumption that the process X is a finite supermartingale and the variable X1 is constant almost surely. By some standard approximation arguments, we may further restrict ourselves to the class of simple supermartingales; recall that the sequence X = (Xn )n>1 is called simple, if for each n the random variable Xn takes only a finite number of values. We are ready to apply Burkholder's method, which is introduced in the next section.

3 Burkholder's method

Now we will describe the main tool which will be used to establish the inequality (1.3). Distinguish the set

D = {(x, y, t) e [0, œ) x [0, œ) x [0, œ) : xp < t, x < y}

and, for each (x, y, t ) e D, let S (x, y, t ) denote the class of all simple, finite and nonnegative supermartingales X = (Xn)n>1 satisfying X1 = x and supT EXp < t, where the supremum is taken over all stopping times adapted to the natural filtration of X. Suppose that we are interested in the explicit formula for the function

where the supremum is taken over all positive integers n and all X e S(x, y, t). The key idea in the study of this problem is to introduce the class C which consists of all functions B : D ^ R satisfying

and the following concavity-type property: if a e [0, 1], (x, y, t) e D and (x±, x± v y, t±) e D satisfy ax- + (1 - a)x+ < x and at- + (1 - a)t+ < t, then

B(x, y, t) > aB(x-, x- v y, t-) + (1 - a)B(x+, x+ v y, t+). (3.2)

We turn to the main result of this section. Recall that the probability space is the interval [0, 1] with its Borel subsets and Lebesgue measure.

Theorem 3.1 The function B is the least element of the class C.

Proof It is convenient to split the reasoning into two parts.

Step 1. First we will show that B belongs to the class C. The majorization (3.1) is immediate, since X*v y > y. The main difficulty lies in proving the concavity property (3.2). Fix the parameters a, x, t, x±, t± as in the statement and pick arbitrary supermartingales X- e S(x-, y, t-), X + e S(x+, y, t+). We splice these two processes into one sequence X = (Xn)n>\ by setting Xi = x and, for n > 2,

B(x, y, t) > y for any (x, y, t) e D,

Xn (v)

Xrl—i(v/a) x+—i«v -

((v — a)/(i — a)) if v e (a, 1].

if v e [0, a],

Then X is a nonnegative supermartingale (with respect to its natural filtration), because the processes X± have this property and ax- + (1 - a)x+ < x. Furthermore, for any stopping time t of X we have EXp < t. To see this, we consider two cases.

1. If P(t = 1) > 0, then the set {t < 1} is nonempty; combining this with the facts that X1 is constant and t is a stopping time of X, we see that {t < 1} = or t = 1. Then EXp = xp < t by the definition of D.

2. Suppose that {t = 1} = 0, or t > 2 almost surely. Then we easily verify that the variables t±, given by

are stopping times of X and X +. Therefore,

EXp = aE(X--)p + (1 - a)E(X ++)p < at- + (1 - a)t+ < t. Hence X e S(x, y, t). Since x < y, we have X*v y = sup2<k<n Xk v y and thus B(x, y, t) > E(Xl v y) = aE((X--) v y) + (1 - a)E((X+-) v y).

Take the supremum over all n and X± as above to obtain the desired bound (3.2).

Step 2. Now suppose that B is an arbitrary element of C; we must prove that B < B. To do this, rephrase the condition (3.2) as follows. Suppose that (X, T) is an arbitrary random variable with two-point distribution, such that P(Xp < T) = 1. Then for any (x, y, t) e D such that EX < x and ET < t, we have

Note that the set {(x, t) : xp < t} is convex. Therefore, by straightforward induction, the above inequality extends to the case when (X, T) is an arbitrary simple random variable satisfying Xp < T with probability 1. Now, pick X e S (x, y, t) and consider the sequence (X, Y, T), where Yn = X*v y and Tn = esssuprejn E(Xp\Fn) (here Tn denotes the class of all stopping times of X not smaller than n). Then the process B(X, Y, T) is a supermartingale: to see this, fix n > 1 and apply (3.3) conditionally with respect to Fn, with X := Xn, y = Yn, t = Tn, X = Xn+1 and T = Tn+1. Let us verify the assumptions: the inequalities Xp < t, X < y and Xp < T are evident; the inequalities E(X\Fn) < X and E(T\Fn) < t follow from the supermartingale property of X and T (T is a supermartingale, since it is the Snell envelope of the sequence (Xp)n>1). Thus, (3.3) yields

t (v) = T(av) — 1, t + = t(a + (1 — a)v) — 1,

B(x, y, t) > EB(X, y, T).

B( Xn, Yn, Tn ) > E B(Xn+i, Xn+i v Yn, Tn+i)\Fn = 4B(Xn+i, Yn+i, Tn+i)\Fn .

Combining this with (3.1) yields

E(X* v y) < EB(Xn, Yn, Tn) < EB(X1, Y1, T1) — B(x, y, sup EXp) < B(x, y, t).

Here in the last inequality we have used the fact that the function t ^ B(x, y, t) is nondecreasing; this follows immediately from (3.2), applied to x+ — x- — x and t+ — t- < t. Taking the supremum over all n and all X s S(x, y, t), we obtain the bound B(x, y, t) < B(x, y, t). This finishes the proof. □

4 Proof of Theorem 1.1

We will prove that the function B admits the following explicit formula:

B(x, y, t) — jy + p—1 log (#1) ify <(t/x)ll{p-1), ly + ty1-p/(p - 1) if y > (tlx)1|(p 1).

This will clearly yield the assertion of Theorem 1.1, which is nothing else but the explicit formula for B(V, V, t). Denote expression on the right above by B(x, y, t).

4.1 Proof of B < B

In the light of Theorem 3.1, it suffices to verify that B s C. The condition (3.1) is obvious, and the main problem is to establish (3.2). First, we easily check that the functions x ^ B(x, x v y, t) and t ^ B(x, y, t) are nondecreasing. Consequently, in the proof of (3.2) we may restrict ourselves to the case x — ax- + (1 - a)x+ and t — at— + (1 - a)t+. Since the region {(x, t) : xp < t} is convex, it is enough to prove the following. For any h s R and any (x, y, t) s D, the function

cp(s) — B(x + sh, (x + sh) v y, t + s)

(defined for those s, for which (x + sh, (x + sh) v y, t + s) s D) is concave. We start from observing that y is of class C1; this follows immediately from the fact that B is of class C1 and By (x, x, t) — 0 (the latter condition guarantees that the one-sided derivatives y'(s-) and y'(s+) will match for x + sh — y). To deal with the concavity of y on the set {s : x + sh < y}, we must prove that the matrix

Bxx(x + sh, y, t + s) Bxt(x + sh, y, t + s) Bxt(x + sh, y, t + s) Btt(x + sh, y, t + s)

(defined for y — (t + s)/(x + sh)) is nonpositive-definite. Substituting x — x + sh and t — t + s if necessary, we may assume that s — 0. Now, if y < (t |x)1/(p-1), then the matrix equals

- ((p - 1)*) 1 ((p - 1)t) 1 ((p - 1)t)~l -X/((p - 1)t2)

which clearly has the required property; if y > (t|x)1/(p-1), the situation is even simpler, since all the entries of the matrix are 0. Finally, it remains to show the concavity of y on {s : x + sh > y}. Because (x + sh)p < t + s (see the definition of D), we have y < ((t + s)l(x + sh))p 1) and hence

x + sh ( (t + s)e p(s) = x + sh + ——r log f

- logf (t + s)e ) 1 5\(x + sh) pj

A direct differentiation yields

p"(s) = -

(x - th)2

x + sh (p - 1)(t + s)2(x + sh)

and the claim follows.

4.2 Proof of B > B

Now we will use the second half of Theorem 3.1, which states that B s C. We will also exploit the following additional homogeneity property of B: for any (x, y, t) s D and I > 0 we have

B(Xx, ly, Xpt) — XB(x, y, t). (4.1)

This condition follows at once from the very definition of B and the fact that X s S(x, y, t) if and only if lX s S(lx, ly, Xpt).

For the sake of clarity, we have split the reasoning into a few parts.

Step 1. Let 8 be a small positive number. Using (3.2), we can write B(1, 1, 1) > (1 - (1 +18)p^j B(0, 1, 0)+ (1 +18)p B(1+8, 1+8, (1+8)p). (4.2)

Thus, using (3.1) and (4.1), the right-hand side is not smaller than

11 „ 1 B(1, 1, 1) 1--+-B(1 + 8, 1 + 8,(1 + 8)p) — 1--+ ■

(1 + 8)p (1 + 8)P J J (1 + 8)P (1 + 8)P-1'

Combining the above facts, we get

(1 + 8)p - 1

B(1, 1, 1) >

(1 + 8^(1 + 8)p-1 - 1)'

so letting 8 ^ 0 gives

B(1, 1, 1) >—!—. (4.4)

Step 2. Now we provide a lower bound for B(1, 1, t), where t is larger than 1. We argue as in the previous step, applying (3.2) and combining it with (3.1) and (4.1). Precisely, we fix a small positive 8 and write

B(1, 1, t) > 1+8B(0, 1, 0) + 1+8B(1 + 8, 1 + 8, t(1 + 8))

> —^ + B | 1, 1,-t-r ). (4.5)

" 1 + 8 V (1 + 8)p-1)

By induction, this implies

B(1,1, t) + B | 1, 1

(1,1,-Vnr)

V (1 + 8)n(p-1)

1 + 8 V ' ' (1 + 8)n(p-1)

if only (1 + 8)n(p-1) < t. Now we fix a large positive integer n, put 8 = 11/(n(p-1)) - 1 (so that (1 + 8)n(p-1) = t) and let n Then the above bound gives

B(1,1, t) > B(1,1,1) + log t

which combined with (4.4) yields

B(1, 1, t) + . (4.6)

p - 1 p - 1

Step 3. The next move in our analysis is to prove the estimate B(x, y, t) > B(x, y, t) for y < (t/x)1/(p-1). We proceed as previously: first apply (3.2) to obtain

y - x x y, t) >--B(0, y, 0) + -B y, y,

y ^ V x

■(yy?)

(here we use the assumption y < (t/x)1/(p-1); if this inequality does not hold, the point (y, y, ty/x) does not belong to D and B(y, y, ty/x) does not make sense). Next, using (3.1), (4.1) and finally (4.6), we get

y - x x t

, y, t) >---y + -• yB 1, 1

V xyp-1)

( p , log(t/(xyp-1))\

> y - x + ^ -- +--- = B(x, y, t).

p - 1 p - 1

Step 4. Now we will deal with B(1, y, 1) for y > 1. By (3.2), (3.1) and (4.1), we have

B(1, y, 1) > (1 - y-p)B(0, y, 0) + y-pB(y, y, yp) > (1 - y-p)y + y1-pB(1, 1, 1).

Combining this with (4.4), we obtain

B(1, y, 1) > y + . (4.7)

Step 5. This is the final part. Pick (x, y, t) e D such that y > (t/x)l/(p-l) and apply (3.2) and then (3.1) to get

B(x, y, t) > «1(0, y, 0) + (1 - a)B ( ——, y, —— )

\ 1 — a 1 — a )

> ay + B(x, y (1 - a), t (1 - a)p—1),

where a = 1 - (xp/1)1/(p-1). For this choice of a, we have xp = t (1 - a)p-1 and hence, by (4.1) and (4.7),

B(x, y, t) > ay + x1, y(1 - a),

/ (xp\ 1/(p-1)\ r 1 t y/xp\ 1/(p-1)l

>\1 -{t) xp + AT)

= B(x, y, t).

This completes the proof of the inequality B > B on the whole domain. Thus, Theorem 1.1 follows.

4.3 On the construction of extremal examples

The arguments presented in Steps 1-5 can be easily translated into the construction of extremal supermartingales X ("extremizers") corresponding to B(x, y, t), i.e., those for which the supremum defining B(x, y, t) is almost attained. The purpose of this section is to explain how to extract this construction from the above calculations. The reasoning will be a little informal, as our aim is to present the idea of the connection.

First we look at the value B(1, 1, 1), which was studied in Step 1. What about the extremal X e S(1, 1, 1)? For a fixed 8 > 0, consider a Markov process ((Xn, Yn, T„ ))n>o starting from (1, 1, 1), satisfying the following two requirements.

(i) Any point of the form (X, X, Xp) (with some X > 0) leads to (0, X, 0) or to (M1 + 8), X(1 + 8), Xp (1 + 8)p) with probabilities 1 - (1 + 8)-p and (1 + 8)-p, respectively.

(ii) The states of the form (0, X, 0) are absorbing.

Then one can check that the process X is a supermartingale. If we stop it after a large number N of steps, we obtain the desired extremizer: that is, if we take sufficiently small 8 and sufficiently large N, then EX*N can be made arbitrarily close to B(1, 1, 1). One might wonder why we have introduced the more complicated three-dimensional process (X, Y, T) above. The reason is that the moves described in (i) and (ii) are

closely related to the inequality (4.2) and the bound (4.3) on which Step 1 rests. To explain this more precisely, note first that (4.2) encodes the Markov move from (i): to make this more apparent, combine (4.2) with (4.1) to get

,X,Xp) > (1--1-1 B(0, X.

' ' ; - V (1 + 8)pJ y ' '

(1 + 8)p

B(X(1 + 8), X(1 + 8), Xp(1 + 8)p).

Thus, a starting state appears on the left, while the destination states can be found on the right, with the appropriate transition probabilities constituting the appropriate weights. The condition (ii) is connected to the bound (4.3): generally speaking, all the states (x, y, t) at which we use the majorization B(x, y, t) > y in the above considerations, need to be assumed absorbing.

To gain more intuition about this interplay, let us look at Step 2, which concerns the value of B(1, 1, t) with t > 1. We will construct a supermartingale X e S(1, 1, t) for which the supremum defining B(1, 1, t) is almost attained. As previously, we fix 8 > 0 and actually handle an appropriate three-dimensional process ((Xn, Yn, Tn))n>0 starting from (1, 1, t). To do this, combine the first inequality in (4.5) with (4.1) to obtain

B(X, X, Xpt) >J-+8B(0, X, 0) + J+8B(X(1 + 8), X(1 + 8), Xpt(1 + 8)).

This gives us some information about the dynamics of the triple (X, Y, T). Namely, we impose the condition

(i') Any point of the form (X, X, Xpt) (with some X > 0 and t > 1) leads to (0, X, 0) or to (X(1 + 8), X(1 + 8), Xpt (1 + 8)) with probabilities 8/(1 + 8) and 1 /(1 + 8), respectively.

Furthermore, in the second inequality of (4.5), we use the bound B(0, X, 0) > X; this suggests that the requirement (ii) above should be valid.

Now, suppose that 8 is chosen appropriately as in Step 2: 8 = t1/(n(p-1)) - 1 for some large positive integer n. Then, after n steps, the process (X, Y, T) gets to the point ((1+8)n ,(1 + 8)n ,(1+8)np ) with positive probability. Now the procedure described in (i') does not apply since the point is not of appropriate form. In Step 2 we encountered a similar phenomenon: the number B(1, 1, 1) came into play and the arguments of Step 2 did not apply. To overcome this difficulty, we exploited Step 1. Here we do the same and apply the procedure (i) to the point ((1 + 8)n, (1 + 8)n, (1 + 8)np). In other words, the Markov process (X, Y, T) is given by the starting position (1, 1, t) and the conditions (i), (ii) and (i'). It remains to stop the process X after a large number of steps to obtain the extremizer corresponding to B(1, 1, t).

The remaining extremal processes, corresponding to the values of B at remaining points, are found similarly. We leave the details to the reader.

5 Lack of prophet inequalities for Lp bounded variables

In the final part of the paper we show that if the condition (1.2) is replaced by

X1, X2, ... are nonnegative and sup E Xvn < 1, (5.1)

then no non-trivial prophet inequality holds. To prove this, we will exploit the results of Sect. 4. Fix an arbitrary positive number K .We have B(1, 1, t) ^to as t ^ to, and hence there is a positive integer L and a finite nonnegative supermartingale (Yn )N=1 on ([0, 1], B([0, 1]), | ■ \) which satisfies Y1 = 1,

sup EYTp < L and EY* > EY1 + K. (5.2)

Now we construct L "distinct" copies of Y which evolve on pairwise disjoint time intervals. Precisely, consider a sequence X = (Xn)L=N1 defined as follows:

• If m e [0, 1/L), then Xn (m) = Yn (Lm) for n = 1, 2, ..., N and Xn (m) = 0 for other n.

• If m e [1/L, 2/L), then Xn (m) = Yn-N (Lm - 1) for n = N + 1, N + 2, ..., 2N and Xn (m) = 0 for other n.

• If m e [2/L, 3/L), then Xn(m) = Yn-2N(Lm - 2) for n = 2N + 1, 2N + 2, ..., 3N and Xn(m) = 0 for other n.

• ...

• If m e [1 - 1/L, 1), then Xn (m) = Yn-(L-1)N (Lm - (L - 1)) for n = (L - 1)N + 1, (L - 1) N + 2, ..., LN and Xn (m) = 0 for other n.

The variables X1, X2, ..., XLN are nonnegative and enjoy the following properties. First, note that the conditional distribution of XmN+n on [m /L, (m + 1)/L) coincides with the distribution of Yn, so EXimN+n = EyI /L and thus

sup EXpn = sup EYnp /L < 1,

1<n<LN 1<n<N

by the first inequality in (5.2). Next, if m e [m/L, (m + 1)/L), we have X*(m) = Y*(Lm - m) and hence EX* = EY*. Finally, we will prove that supTeT EXT = EY1. The inequality ">" follows by considering the stopping time given by t (m) = mL +1 for m e [m/L, (m + 1)/L). To prove the reverse bound, note that while computing supTeT EXT we may restrict to stopping times which on each [m /L, (m + 1)/L) take values {mL + 1, mL + 2,...,(m + 1)L}. Indeed, if t is an arbitrary stopping time, then t = (t vt)A (t + L - 1) has this property and E X T < E X T. Therefore, since Y is a supermartingale, we have

E(Xt\[m/L, (m + 1)/L)) < E(XmL+1\[m/L, (m + 1)/L)) = EY1

and hence EXT < EY1.

Consequently, by the second estimate in (5.2), the sequence X satisfies

E X * > sup E X r + K.

Since K was arbitrary, no universal prophet inequality holds under (5.1). This completes the proof.

Acknowledgments The author would like to thank an anonymous Referee for the careful reading of the first version of the paper and several helpful suggestions.

Open Access This article is distributed under the terms of the Creative Commons Attribution License which permits any use, distribution, and reproduction in any medium, provided the original author(s) and the source are credited.

References

1. Allaart, P.C.: Prophet inequalities for I.I.D. random variables with random arrival times. Seq. Anal. 26, 403-413 (2007)

2. Allaart, P.C., Monticino, M.G.: Pseudo-prophet inequalities in average-optimal stopping. Seq. Anal. 22, 233-239 (2003)

3. Boshuizen, F.: Multivariate Prophet Inequalities for Negatively Dependent Random Vectors. Strategies for sequential search and selection in real time (Amherst, MA, 1990), Contemporary Mathematics, vol. 125, pp. 183-190. American Mathematical Society, Providence (1992)

4. Assaf, D., Goldstein, L., Samuel-Cahn, E.: Ratio prophet inequalities when the mortal has several choices. Ann. Appl. Probab. 12(3), 972-984 (2002)

5. Assaf, D., Samuel-Cahn, E.: Prophet inequalities for optimal stopping rules with probabilistic recall. Bernoulli 8(1), 39-52 (2002)

6. Boshuizen, F.: Prophet compared to gambler: additive inequalities for transforms of sequences of random variables. Statist. Probab. Lett. 29, 23-32 (1996)

7. Burkholder, D.L.: Explorations in Martingale Theory and its Applications. Ecole d'Ete de Probabilités de Saint-Flour XIX-1989, Lecture Notes in mathematics, vol. 1464, pp. 1-66. Springer, Berlin (1991)

8. Hill, T.P.: Prophet inequalities and order selection in optimal stopping problems. Proc. Am. Math. Soc. 88, 131-137 (1983)

9. Hill, T.P.: Prophet inequalities for averages of independent non-negative random variables. Math. Z. 192, 427-436 (1986)

10. Hill, T.P., Kertz, R.P.: Additive comparisons of stop rule and supremum expectations of uniformly bounded independent random variables. Proc. Am. Math. Soc. 83, 582-585 (1981)

11. Hill, T.P., Kertz, R.P.: Stop rule inequalities for uniformly bounded sequences of random variables. Trans. Am. Math. Soc. 278, 197-207 (1983)

12. Hill, T.P., Kertz, R.P.: A Survey of Prophet Inequalities in Optimal Stopping Theory. Strategies for sequential search and selection in real time (Amherst, MA, 1990), Contemporary Mathematics, vol. 125, pp. 191-207. American Mathematical Society, Providence (1992)

13. Kennedy, R.P.: Optimal stopping of independent random variables and maximizing prophets. Ann. Probab. 13, 566-571 (1985)

14. Kennedy, R.P.: Prophet-type inequalities for multi-choice optimal stopping. Stoch. Proc. Appl. 24, 77-88 (1987)

15. Kennedy, D.P., Kertz, R.P.: A prophet inequality for independent random variables with finite variances. J. Appl. Probab. 34(4), 945-958 (1997)

16. Kertz, R.P. : Comparison of optimal value and constrained maxima expectations for independent random variables. Adv. Appl. Prob. 18, 311-340 (1986)

17. Krengel, U., Sucheston, L.: Semiamarts and finite values. Bull. Am. Math. Soc. 83, 745-747 (1977)

18. Krengel, U., Sucheston, L.: On semiamarts, amarts, and processes with finite value. In: Kuelbs, J. (ed.) Probability on Banach Spaces. Marcel Dekker, New York (1978)

19. Krengel, U., Sucheston, L.: Prophet compared to the gambler: an inequality for transforms of processes. Ann. Probab. 15, 1593-1599 (1987)

20. Nazarov, F., Treil, S.: The hunt for Bellman function: applications to estimates of singular integral operators and to other classical problems in harmonic analysis. Algebra i Analis 8, 32-162 (1997)

21. Osekowski, A.: Sharp Martingale and Semimartingale Inequalities, vol. 72. Monografie Matematyczne, Birkhäuser (2012)

22. Peskir, G., Shiryaev, A.N.: Optimal Stopping and Free-boundary Problems. Lectures in Mathematics ETH Zürich. Birkhäuser Verlag, Basel (2006)

23. Shiryaev, A.N.: Optimal Stopping Rules. Translated from the 1976 Russian second edition by A.B. Aries. Reprint of the 1978 translation. Stochastic Modelling and Applied Probability, vol. 8. SpringerVerlag, Berlin (2008)

24. Tanaka, T.: Prophet inequalities for two-parameter optimal stopping problems. J. Inf. Optim. Sci. 28, 159-172 (2007)

25. Tanaka, T.: Lower semicontinuity property of multiparameter optimal stopping value andits application to multiparameter prophet inequalities. J. Math. Anal. Appl. 359, 240-251 (2009)

26. Vasyunin, V., Volberg, A.: Monge-Ampere Equation and Bellman Optimization of Carleson Embedding Theorems. Linear and complex analysis. American Mathematical Society Translations: Series 2, vol. 226, pp. 195-238. American Mathematical Society, Providence (2009)