Examples
It is intuitively clear that many behaviours can be specified by means
of $FLTL formulae. While there is no simple way in general to
translate between past and future tense expressions,4all of the examples used to illustrate PLTL in
Section 2.2 above are expressible naturally in
$FLTL, as follows.
The classical goal formula
saying that a goal
is rewarded
whenever it happens is easily expressed:
. As already noted,
is the set of finite sequences of
states such that
holds in the last state. If we only care that
is achieved once and get rewarded at each state from then on, we write
. The behaviour that this formula
represents is the set of finite state sequences having at least one
state in which
holds. By contrast, the formula
stipulates only that the first occurrence of
is
rewarded (i.e. it specifies the behaviour in Figure 1).
To reward the occurrence of
at most once every
steps, we write
.
For response formulae, where the achievement of
is triggered by
the command
, we write
to reward every state in which
is true following the
first issue of the command. To reward only the first occurrence
after each command, we write
. As for bounded variants for which we only
reward goal achievement within
steps of the trigger command, we
write for example
to reward all such states in which
holds.
It is also worth noting how to express simple behaviours involving
past tense operators. To stipulate a reward if
has always been
true, we write
. To say that we are rewarded if
has been true since
was, we write
.
Finally, we often find it useful to reward the holding of
until
the occurrence of
. The neatest expression for this is
.
Sylvie Thiebaux
2006-01-20