🏠
Guest Not signed in

Optimal stopping: the threshold rule

The optimal rule is almost always 'stop when the current value beats this threshold'. The interview question is just figuring out the threshold.

Method · Optimal Stopping
Intro
Three secretary-problem scenarios shown as silhouettes of varying heights β€” illustrates the explore-then-commit 37% rule.
Cmglee, CC BY-SA 4.0 · CC-BY-SA-4.0 · Wikimedia Commons

Optimal stopping problems give you a stochastic sequence one item at a time and force an immediate accept-or-reject on each β€” no going back. The classic example is the secretary problem, and its optimal policy is a single-number threshold: reject the first $r-1$ items to calibrate, then accept the first running-max after that. The work is choosing $r$. Write the success probability as a sum over the position of the prize, $(r-1)/n \cdot \sum_{j=r}^{n} 1/(j-1)$, and maximize. The asymptotic answer $r \approx n/e$ gives $1/e \approx 0.368$ β€” the most-quoted constant in the optimal-stopping literature.

βœ“ Intro Β· expand
More examples

A handful of harder problems on the same theme. Click any prompt to reveal the solution sketch.

Dice-with-bust game. A casino offers a dice game: you may roll as many times as you like, but each roll has rules. If the die shows 1–5, you may pocket the face value and either stop or roll again; if the die shows 6, all winnings accumulated so far are wiped out and the game ends. After each non-6 roll, you decide whether to keep what you have or risk another roll. If you are risk-neutral, how much would you pay to play once?
Optimal stopping threshold: cash out once your accumulated total is at least $\$15$. Expected pre-game value is approximately $\$6.00$. Reasoning: at accumulated total $n$, expected gain from one more roll is $\tfrac{5}{6} \cdot 3 - \tfrac{1}{6} n = \tfrac{5}{2} - \tfrac{n}{6}$, which turns negative for $n > 15$. Stopping rule + Bellman equation gives the entry value.
Dynamic card stopping. A casino deals from a standard 52-card deck face-up, one card at a time. Each red card pays you $1; each black card costs you $1. You may stop the dealer at any moment and pocket your net. What is the optimal stopping rule, and how much would you pay (risk-neutral) to play?
Stop when the running net first reaches its maximum-possible-future value, which by a DP argument is $f(b, r)$ β€” the expected payoff from state with $b$ black and $r$ red cards remaining. The closed-form is $f(b, r) = \max(r - b, 0) + \text{adjustment}$, and $f(26, 26) \approx \$2.62$. So a risk-neutral entry price is about $\$2.62$.
+ More examples Β· expand
Independent · Legal