In sequential analysis we don't have a fixed number of observations.
Instead, observations come in sequence, and we'd like to decide in
favor of
or
as soon as possible. For each
we perform a test:
. There are three outcomes of :
Decide
Decide
Keep testing
NonBayesian Case Let be the stopping time of this
test. We wish to find an optimal tradeoff between:
, the probability:: [
, but
]
, the probability: [
, but
]
, where
or
It turns out, the optimal test again involves monitoring the
likelihood ratio.
This test is called SPRT for ``Sequential
Probability Ratio Test''. It is more insightful to examine this test
in the ``log'' domain. The test involves comparing the log likelihood ratio:
to positive and negative thresholds , . The first time
, we stop the test and decide
The first time
, we stop and declare
. Otherwise we keep on testing.
There is one ``catch''; in the analysis, we ignore overshoots
concerning the threshold boundary. Hence
or .
Properties of SPRT The change (first difference) of is
For an iid process, we drop the conditioning:
The drift of is defined as
. From definitions, it follows that the drifts under
or
are given by the K-L informations:
We can visualize the behavior of , when in fact
undergoes a step transition from
to
:
Again, we have practical issues concerning how we choose thresholds
. By invoking Wald's equation, or some results from martingale
theory, these are easily related to the probabilities of error at the
stopping time of the test. However, the problem arises how to choose
both probabilities of error, since we have a three-way tradeoff with
the average run lengths,
!!
Fortunately, the Bayesian formulation comes to our rescue. We can
again assign costs to the probabilities of false alarm and miss
. We also include a cost proportional to the number of observations prior
to stopping. Let this cost equal the number of observations, which is .
The goal is to minimize expected cost, or sequential Bayes risk.
What is our prior information? Again, we must know
.
It turns out that the optimal Bayesian strategy is again a SPRT. This follows from the theory of optimal stopping. Suppose at time , our
we have yet to make a decision concerning . We must decide among the
following alternatives:
Stop, and declare
or
.
Take one more observation.
We choose to stop only when the minimum additional cost of stopping
is less than the minimum expected additional cost of taking one
more observation.
We compute these costs using the posterior distribution of
, i.e:
which comes by recursively applying Bayes' rule.
If we stopped after observing and declared
, the expected cost due to ``miss'' would be
. Therefore
if we make the decision to stop, the (minimum) additional cost is
The overall minimum cost is:
In the two-hypothesis case, the implied recursion for the minimum
cost can be solved, and the result is a SPRT(!)
Unfortunately, one cannot get a close form
expression for the thresholds in terms of the costs, but the ``Bayes''
formulation allows at least to involve prior information about the
hypotheses.
We will see a much richer extension to the problem of Bayesian
change detection.