How do we remember the past in randomised strategies?

How do we remember the past in randomised strategies?

Abstract

Graph games of infinite length are a natural model for open reactive processes: one player represents the controller, trying to ensure a given specification, and the other represents a hostile environment. The evolution of the system depends on the decisions of both players, supplemented by chance.

In this work, we focus on the notion of randomised strategy. More specifically, we show that three natural definitions may lead to very different results: in the most general cases, an almost-surely winning situation may become almost-surely losing if the player is only allowed to use a weaker notion of strategy. In more reasonable settings, translations exist, but they require infinite memory, even in simple cases. Finally, some traditional problems becomes undecidable for the strongest type of strategies.



“You can’t have a strategy against telepaths: you have to act randomly. You have to not know what you’re going to do next. You have to shut your eyes and run blindly. The problem is: how can you randomise your strategy, yet move purposefully towards your goal?”


Solar Lottery

Philip K. Dick

1 Introduction

Since their introduction to verification in the late eighties, graph games have emerged as the model of choice for problems about open systems, where a controller (Eve) must interact with an a priori hostile environment (Adam) [PR89]. In such games, an arenai.e. a graph— models the system and its evolution: at the beginning, a token is laid on one of the vertices, and its moves are determined by the actions of the players, supplemented by chance. The infinite sequence of vertices that ensues constitutes a play of the game, whose winner is defined by some predetermined specification, often given as a regular condition on infinite words [MP92].

This model has been declined in a multiplicity of variants, in terms of both arenas and objectives. However, the questions are nearly always the same: Is there a winning strategy? For which player? How complex is it, in terms of memory and randomisation? Memory is the quantity of information that one is allowed to remember from the past: in general, the whole history is available, but it is often enough to remember a finite quantity of information. In addition, a strategy is pure if it proposes only one possible action after any given sequence of observations. The notion of randomised strategy, and its relation to memory, is the subject of this paper.

In verification, “randomised strategy” usually refers to a function from the history to a probability distribution over the actions. In other domains, such strategies are called “behavioural strategies”, as opposed to two other models of randomised strategies: a mixed strategy is a measure over pure strategies, and a general strategy is a measure over behavioural strategies. These models are also relevant in computer science. Indeed, the IPv6 “Stateless Address Autoconfiguration” protocol, which only uses randomisation at the beginning to generate a new I.P. address [TNJ07], can be accurately described as a mixed strategy. Likewise, the secure shell protocol (ssh) is a general strategy, since a new session key has to be randomly1 generated every hour or gigabyte [YL06].

In this paper, we propose definitions for mixed and general strategies, with or without memory, in the framework of graph games for verification. We expose several situations in which their analyses differ significantly from the behavioural model. In the most general case, the same game can be almost-surely losing or almost-surely winning depending on the type of strategies we consider. In other situations, we conjecture that the values are the same, but we show that memory needs vary (from two to infinity). Altogether, we hope to ask more questions than we give answers: our main objective is to describe these three models for randomised strategies in graph games and to point out that many problems which are solved for behavioural strategies are still open in the mixed and general cases.

The paper is organised as follows. In Section 2, we recall the classical notions about graph games in verification, in a very general framework which subsumes a large part of the literature. Section 3 presents our definitions for behavioural, mixed, and general strategies in graph games, and stresses the fundamental differences between the three notions. Section 4 focuses on memory-related issues: it exhibits variations in the elementary cases of concurrent safety games and simple Muller games. In Section 5, we sum up our observations and results, and propose some open problems.

2 Definitions

Notation

For a finite or countable set , we denote by the set of probability distributions over , i.e. the set of functions from to positive real numbers that sums up to one.

Arenas and plays

An arena is a tuple where is the set of vertices in the graph, is the set of actions of Eve, is the set of actions of Adam, is the transition function, is the set of colours, is the colouring function, is the set of signals of Eve, is her observation function, is the set of signals of Adam, and is his observation function. Many results about graph games for verification consider only restricted arenas, such as:

Synchronous: an arena is synchronous if and are total.

Observable actions: a synchronous arena has observable actions if the restrictions of to and to are one-to-one.

Perfect information: a synchronous arena has perfect information (or is concurrent) if the restrictions of and to are one-to-one.

Simple: a concurrent arena is simple (or turn-based) if for each vertex , depends either on or on , but not both.

A play on the arena is a (possibly infinite) sequence of states such that . The set of plays is usually denoted , and the set of plays starting with the vertex by .

Pure strategies and measures

A pure strategy for Eve (resp. for Adam) on the arena associates an action to each finite sequence of observations: (resp. ) . A play is consistent with a strategy for Eve (resp. for Adam) if and only if at each step , there is an action for Adam (resp. ) such that (resp. ). Notice that, in the case of an asynchronous arena, actions can only change with new observations: otherwise, the same argument leads to the same result over and over. The set of plays consistent with (resp. ; and ) is denoted by (resp. ; ). Once an initial vertex and two strategies and have been fixed, can naturally be made into a measurable space , where is the -field generated by the cones : if and only if is a prefix of . The probability measure is recursively defined by and:

Carathéodory’s extension theorem allows us to extend to the Borel sets of  [Wil91].

Winning conditions and values

A winning condition on a set of colours is a Borel subset of . A play in an arena on is winning for Eve in the game if , and winning for Adam otherwise. In a game , the pure value of a state under the strategies and , denoted , is the measure of under . The value for Eve of a state is the supremum of the values that she can ensure from against any strategy of Adam. Symmetrically, the value for Adam of a state is the infimum of the values he can defend against any strategy of Eve. In simple stochastic games, these two values coincide and are usually called the value of  [Mar98, MS98]2.

Winning criteria

Following de Alfaro and Henzinger [dAH00], we consider several notions of winning strategies and winning regions, depending on the chances Eve has to win. In decreasing order of difficulty, and from an initial vertex , a strategy for Eve:

  • is sure if any play consistent with is winning for Eve;

  • is almost-sure if for any strategy for Adam, ;

  • ensures if for any strategy for Adam, ;

  • is positive if for any strategy for Adam, ;

  • is heroic if for any strategy , there is a play consistent with and which is winning for Eve.

The sure region (resp. almost-sure region of Eve, positive region, heroic region) of Eve is the set of vertices from which she has a sure (resp. almost-sure, positive, heroic) strategy. Furthermore, the bounded region is the set of vertices from which Eve has a strategy ensuring a positive and the limit-one region is the set of vertices from which Eve has a strategy ensuring for any . The same concepts are defined accordingly for Adam, except that we say that a strategy for Adam defends if it guarantees that, for any strategy for Eve, .

3 Behavioural, mixed, and general strategies

As soon as we deal with concurrent arenas, we cannot rely only on pure strategies to make meaningful analyses. In the classical game of “Janken”, any pure strategy is surely beaten by the appropriate counter-strategy ( against , against , and against ), but a strategy which plays each action with probability eventually wins with probability . The main point of this paper is that there are several possible definitions for the notion of “randomized strategy”.

  • A behavioural strategy returns at each step a distribution over the actions: ;

  • a mixed strategy is a measure over pure strategies: ;

  • a general strategy is a measure over behavioural strategies: .

As we show in this paper the expressive powers of these models are quite different. Intuitively, a behavioural strategy does not know in advance what it will play next, so its actions can change when its decisions do not (even when there are no observations). Mixed strategies use randomization to get hidden information at the beginning of the play, which can later be used to correlate undistinguishable actions, e.g. playing or with probability . General strategies subsume both, so they can, in particular, generate hidden information on the fly. These distinctions have mostly been overlooked in verification (apart from a few remarks, e.g. [dAHK98, DHR08]). One reason is that the games we consider are usually synchronous, with observable actions. On synchronous arenas, mixed strategies can simulate behavioural strategies: as each action can be uniquely identified beforehand by its position in the play, it is possible to define a measure which somehow makes all the random draws at the beginning of the play. If furthermore, the actions are observable, Kuhn’s theorem states that mixed and behavioural (and thus general) strategies have the same expressive power [Aum64].

These hypotheses have been inconspicuously challenged in recent papers. In this regard, the comparison between [BGG09] and [GS09] makes for an enlightening example. At first glance, these two papers look very similar: they both ponder the problem of the existence of almost-sure strategies in games where both players have (asymmetric) imperfect information. A closer examination reveals the differences: Bertrand, Genest, and Gimbert use general strategies, while Gripon and Serre use behavioural strategies; furthermore, in the latter paper, the players cannot observe their own actions. As a consequence, there are cases where the answer to the synthesis problem depends on which model is used. Consider for example the synchronous arena depicted in Figure 1, where Eve cannot distinguish vertices nor actions in the dashed area, is a losing sink state, and is her “target”, for either a reachability or a Büchi condition.

\gassetNh=4,Nw=4,polyangle=90 \node[Nmarks=i, iangle=270](i)(35,15) \node(delay)(35,27) \node(a)(10,21)\node(b)(59,21)\node[Nmarks=r](top)(35,39) \node(bottom)(35,3) \drawline[AHnb=0](36.45,4.45)(33.55,1.55) \drawline[AHnb=0](36.45,1.55)(33.55,4.45) \drawedge[curvedepth=2,ELpos=40](i,delay)\drawedge[curvedepth=2,ELpos=60](i,delay)\drawedge[curvedepth=3,ELside=r](i,a)\drawedge[curvedepth=-3](i,b)\drawedge[curvedepth=2,ELpos=40](delay,i)\drawedge[curvedepth=2,ELpos=60](delay,i)\drawedge[curvedepth=5](a,top)\drawedge[curvedepth=-5,ELside=r](a,bottom)\drawedge[curvedepth=-5,ELside=r](b,top)\drawedge[curvedepth=5](b,bottom)\drawedge(top,delay) \drawccurve[dash=3 1 1 10](5,25)(5,17)(65,17)(65,25)
Figure 1: Who wins?

With a behavioural strategy, Eve’s strategy can only depend on the length of the play. At any even move, if her strategy is to play with probability and with probability , Adam can answer by playing with probability and with probability , so the odds of the token going to or to are equal (they are worth each). In the next step, no matter what advocates, the odds of the token going to or to will again be equal. In the reachability game, this limits Eve’s prospects to half chances. In the Büchi game, the probability that she wins drops to .

On the flip side, she has an almost-sure mixed strategy for both objectives: the natural “uniform” measure over the strategies of the form guarantees that each sequence of two moves starting in the initial vertex has a probability of to send the token to , and a probability of to send the token back to the initial vertex, no matter what Adam does. It cannot go to , as Eve never plays or from the initial vertex.

The arena of Figure 1 is synchronous, so any behavioural strategy can be emulated by a mixed one. If we remove this hypothesis, it is not always the case, as in the one-player game of Figure 2, where Eve is unaware of any action or vertex.

\gassetNh=4,Nw=4,polyangle=90 \node[Nmarks=i, iangle=180](first)(7,9) \node(second)(21,9) \node[Nmarks=r](exit)(35,3) \node(fail)(35,15) \drawline[AHnb=0](36.45,16.45)(33.55,13.55) \drawline[AHnb=0](36.45,13.55)(33.55,16.45) \drawedge[ELside=r](first,second)\drawedge[curvedepth=1,ELside=r](second,fail)\drawedge[ELpos=42,curvedepth=3](first,fail)\drawedge[curvedepth=-1,ELside=r](second,exit)
Figure 2: The D.U.I. game

As Eve observes nothing, her strategy is completely determined by what she does on the empty word . She has only two pure strategies: and . Both lead to , and so does any mixed strategy of the form . The behavioural strategy , on the other hand, yields one chance out of four to to reach .

The case of games with perfect information and invisible actions is still open: there are mixed strategies which cannot be imitated by any behavioural one, so we cannot hope for a “generic” translation. But that does not rule out the possibility of specific, objective-dependent constructions which would yield a different strategy with the same value.

4 Memory issues

A refinement of the synthesis problem asks that the controller uses only finite memory, as a natural requirement for implementability. Pure strategies with memory are defined in the following way:

Definition 1

A pure strategy with memory is a triple where is the initial memory state; is the memory update function, which maps a signal and a memory state to a new memory state and is called at each new observation of Eve; and is the next-action function, which maps a signal and a memory state to an action and is called at each step.

Notice that any pure strategy can be represented as a strategy with memory , with , and . A strategy has finite memory if is a finite set, and is memoryless if is a singleton.

Randomized strategies with (countable) memory are defined with similar tuples, except that some of their elements use randomization.

Behavioural: In a behavioural strategy with memory , the next-action is randomized.

Mixed: In a mixed strategy with memory , the initial memory is randomised.

General: In a general strategy with memory , the next-action, initial memory, and memory-update are randomised.

The memory requirements can also depend of the type of strategy. In the game of Figure 1, for example, there is no almost-sure mixed strategy with finite memory (in the reachability game, there are -optimal strategies with finite (unbounded) memory; in the Büchi game, every mixed strategy with finite memory has value ). However, the strategy we described can be realized by a general strategy with four memory states , , and : in the memory state, she updates her memory at random to one of the states; in the states, she updates her memory to the corresponding state; in all states, she plays the action corresponding to her memory state.

4.1 Concurrent safety games

In [dAHK98], de Alfaro, Henzinger, and Kupferman study the problem of concurrent reachability/safety games and establish the qualitative determinacy of these games, as well as several results on the nature (memory and randomization) of the strategies needed to achieve various objectives. In particular, they show that positive strategies for safety objectives require, in general, an infinite amount of memory. The proof is based on the famous “snowball game” of [KS81], which is pictured in Figure 3.

\gassetNw=2,Nh=2,Nmr=6,loopdiam=2.8 \node(un)(2,3) \node[Nmarks=i,iangle=270](deux)(12,3) \node(trois)(22,3) \drawline[AHnb=0](22.7,3.7)(21.3,2.3) \drawline[AHnb=0](21.3,3.7)(22.7,2.3) \drawedge[ELside=r](deux,un)\drawloop(deux)\drawedge(deux,trois)\drawedge[ELside=r](deux,trois)
Figure 3: The snowball game

In this game, Adam loses if he never runs and Eve never throws, or if Eve happens to throw the snowball exactly at the moment he runs. It is clear that Adam has memoryless behavioural strategies with value arbitrarily close to one: if, at each step, he chooses to run with probability , he ensures a probability of winning of (Eve’s best chance is to throw the ball right away). It is also clear that he cannot win almost-surely: if he has a positive probability of never running, Eve can keep the snowball forever; and if he has a positive probability of running at any step, Eve can thwart him by throwing the ball with probability at each step.

By the qualitative determinacy of concurrent regular games, Eve has a positive strategy, i.e. a unique strategy which prevents Adam from winning almost-surely with any strategy. De Alfaro, Henzinger, and Kupferman use behavioural strategies, and argue that Eve needs infinite memory: the sequence must go to but never reach it. It is clear that there are no positive mixed strategies with finite memory, as pure strategies with memory states can only throw the snowball in the first steps.

On the other hand, there is a general strategy with only memory states: in the memory state , Eve keeps the snowball with probability , in the state , she throws it with probability ; the memory never changes, and the initial memory state is chosen at random. This strategy prevents Adam from winning almost-surely, since he can never be sure that Eve is not in the memory state . In fact, this is the case in every finite concurrent safety game:

Theorem 2

In every finite concurrent safety game, Eve has a positive general strategy with memory from her positive region.

Sketch of proof. It follows from the analysis of the fix-points in [dAHK98] that there is a total preorder  on the vertices such that:

  • the minimal vertices belong to the almost-sure region of Adam;

  • for each non-minimal vertex , there is an action of Eve such that, for any action of Adam:

    • either for any vertex , ,

    • or there is an action of Eve and a vertex such that

Notice that always playing the action is a pure and positional sure strategy for Eve in the maximal vertices (unless they are also minimal). For the vertices in between, we claim that the following strategy with two memory states is positive for Eve:

  • the memory states are called and ;

  • each time the token goes to a new -class, the memory state is updated to either or with equal probabilities; otherwise, the memory does not change;

  • in the state, Eve always plays the action of the current vertex;

  • in the state, Eve plays any action in with equal probabilities.

The situation is roughly the same as in the snowball game: if Adam’s actions have no chance to go to a lower vertex against the strategy, he will lose with probability ; if he takes a risk at any point, there is a positive probability that Eve was in the memory state all along, so he could end up in a greater vertex.

In addition to the finite memory, the strategy described in the proof of Theorem 2 is simple, generic, and uses only uniform probabilities. By comparison, the description of a positive behavioural strategy is in general very complex and uses probabilities of unbounded precision.

4.2 Memory bounds for Muller games

Even in the elementary case of simple Muller games, it is not clear that the memory needs are the same for behavioural and general strategies. Recall that a simple arena is an arena with turn-based moves and perfect information for both players, and a Muller condition is a condition depending only on the set of colours visited infinitely often:

Definition 3

A Muller condition on a set of colours is specified by a subset of . A play satisfies if and only if the set of colours occurring infinitely often in belongs to .

In such games, both players have pure optimal strategies with finite memory [BL69]. A follow-up problem is to determine, for a given Muller condition on a set of colours , the necessary and sufficient amount of memory needed to define optimal pure strategies in any arena coloured by . Gurevich and Harrington used the latest appearance record (LAR) structure of McNaughton to give a first upper bound of  [GH82]. Zielonka refined the LAR into a tree, whose leaves could be used as memory [Zie98]. Finally, Dziembowski, Jurdzinski, and Walukiewicz showed that each player needs only as much memory as the number of leaves in some particular sub-trees, establishing tight and asymmetrical bounds for pure strategies [DJW97].

It is clear from their proof that mixing strategies does not help, since the other player can efficiently adapt their strategy in the witness arenas. This is not the case for behavioural strategies: Chatterjee, de Alfaro, and Henzinger observed that upward-closed winning conditions admitted memoryless strategies [CdAH04], leading to smaller upper bounds for arbitrary Muller conditions [Cha07]. Horn established even smaller tight bounds for general strategies [Hor09] (see Figure 4 for a graphical representation of the three bounds on a Zielonka (sub-)tree). However, Horn’s upper-bound has not yet been proven (or refuted) for behavioural strategies.


\gassetNw=4,Nh=4,AHnb=0 \node[Nmr=4](abcd)(12,27) \node[Nmr=0](bcd)(6,20) \node[Nmr=0](acd)(12,20) \node[Nmr=0](abd)(18,20) \node[Nmr=3](ab)(18,13) \node[Nmr=0](a)(14,6) \node[Nmr=0](b)(22,6) \drawedge(abcd,bcd) \drawedge(abcd,acd) \drawedge(abcd,abd) \drawedge(abd,ab) \drawedge(ab,b) \drawedge(ab,a) \node[Nframe=n](1)(6,16)X \node[Nframe=n](2)(12,16)X \node[Nframe=n](3)(14,2)X \node[Nframe=n](4)(22,2)X
(a) Pure/mixed (tight)

\gassetNw=4,Nh=4,AHnb=0 \node[Nmr=4](abcd)(12,27) \node[Nmr=0](bcd)(6,20) \node[Nmr=0](acd)(12,20) \node[Nmr=0](abd)(18,20) \node[Nmr=3](ab)(18,13) \node[Nmr=0](a)(14,6) \node[Nmr=0](b)(22,6) \drawedge(abcd,bcd) \drawedge(abcd,acd) \drawedge(abcd,abd) \drawedge(abd,ab) \drawedge(ab,b) \drawedge(ab,a) \node[Nframe=n](1)(6,16)X \node[Nframe=n](2)(12,16)X \node[Nframe=n](3)(18,3)\node[Nframe=n](4)(18,1)X
(b) Behavioural (upper)

\gassetNw=4,Nh=4,AHnb=0 \node[Nmr=4](abcd)(12,27) \node[Nmr=0](bcd)(6,20) \node[Nmr=0](acd)(12,20) \node[Nmr=0](abd)(18,20) \node[Nmr=3](ab)(18,13) \node[Nmr=0](a)(14,6) \node[Nmr=0](b)(22,6) \drawedge(abcd,bcd) \drawedge(abcd,acd) \drawedge(abcd,abd) \drawedge(abd,ab) \drawedge(ab,b) \drawedge(ab,a) \node[Nframe=n](1)(9,17)\node[Nframe=n](2)(9,15)X \node[Nframe=n](3)(18,3)\node[Nframe=n](4)(18,1)X
(c) General (tight)
Figure 4: Memory bounds for simple Muller games

5 Discussion

We have compared three models of randomized strategies —behavioural, mixed, and general— in the context of graph games. Depending on the sub-case, we were able to expose variations in the amount of memory needed, the existence of finite-memory strategies, or even the values. In concurrent games with unobservable actions, the equivalence between the three models is still an open question.

In verification, the behavioural model has received most of the attention. Nevertheless, there is a priori nothing wrong with the other types of controllers. Furthermore, in several cases, general strategies can be much simpler than behavioural or mixed ones. On the other hand, general strategies are much less amenable to further analysis, as they introduce imperfect information. Even in simple safety games, one cannot compute the value of a general strategy —or even decide if it has positive value [GO09].

Each model has strengths and weaknesses, and we do not favour one over the others. Our point is rather to stress the importance of this initial choice, and to note that many memory-related problems which have been solved for behavioural strategies are still open in the mixed and general frameworks.

Footnotes

  1. except on Debian [BB08].
  2. As a matter of fact, these papers shows the quantitative determinacy, in behavioural strategies, of Borel games on concurrent arenas. An inspection of the proof yields the same result for pure strategies in the case of simple arenas.

References

  1. Robert J. Aumann. Mixed and Behavior Strategies in Infinite Extensive Games. In Advances in Game Theory, volume 52 of Annals of Mathematical Studies, pages 627–650. Princeton University Press, 1964.
  2. Luciano Bello and Maximiliano Bertacchini. Predictable RNG in the vulnerable Debian OpenSSL package, the What and the How. In the 2nd DEF CON Hacking Conference, 2008.
  3. Nathalie Bertrand, Blaise Genest, and Hugo Gimbert. Qualitative Determinacy and Decidability of Stochastic Games with Signals. In Proceedings of the 24rd Annual IEEE Symposium on Logic in Computer Science, LICS’09, pages 319–328. IEEE Computer Society, 2009.
  4. J. Richard Büchi and Lawrence H. Landweber. Solving Sequential Conditions by Finite-State Strategies. Transactions of the American Mathematical Society, 138:295–311, 1969.
  5. Krishnendu Chatterjee, Luca de Alfaro, and Thomas A. Henzinger. Trading Memory for Randomness. In Proceedings of the 1st International Conference on Quantitative Evaluation of Systems, QEST’04, pages 206–217. IEEE Computer Society, 2004.
  6. Krishnendu Chatterjee. Optimal Strategy Synthesis in Stochastic Muller Games. In Proceedings of the 10th International Conference on the Foundations of Software Science and Computational Structures, FoSSaCS’07, volume 4423 of Lecture Notes in Computer Science, pages 138–152. Springer-Verlag, 2007.
  7. Luca de Alfaro and Thomas A. Henzinger. Concurrent -regular Games. In Proceedings of the 16th Annual IEEE Symposium on Logic in Computer Science, LICS’00, pages 141–154. IEEE Computer Society, 2000.
  8. Luca de Alfaro, Thomas A. Henzinger, and Orna Kupferman. Concurrent Reachability Games. In Proceedings of the 39th Annual Symposium on Foundations of Computer Science, FoCS’98, pages 564–575. IEEE Computer Society, 1998.
  9. Laurent Doyen, Thomas A. Henzinger, and Jean-François Raskin. Equivalence of Labeled Markov Chains. International Journal of Foundations of Computer Science, 19(3):549–563, 2008.
  10. Philip K. Dick. Solar Lottery. Ace Books, 1955.
  11. Stefan Dziembowski, Marcin Jurdziński, and Igor Walukiewicz. How Much Memory is Needed to Win Infinite Games? In Proceedings of the 12th Annual IEEE Symposium on Logic in Computer Science, LICS’97, pages 99–110. IEEE Computer Society, 1997.
  12. Yuri Gurevich and Leo Harrington. Trees, Automata, and Games. In Proceedings of the 14th Annual ACM Symposium on Theory of Computing, STOC’82, pages 60–65. ACM Press, 1982.
  13. Hugo Gimbert and Youssouf Oualhadj. Automates Probabilistes : Problèmes Décidables et Indécidables. Technical Report hal-00422888, Laboratoire Bordelais de Recherche en Informatique, CNRS UMR 5800, 2009.
  14. Vincent Gripon and Olivier Serre. Qualitative Concurrent Games with Imperfect Information. In Proceedings of the 36th International Colloquium on Automata, Languages and Programming, ICALP’09, volume 5556 of Lecture Notes in Computer Science, pages 200–211. Springer-Verlag, 2009.
  15. Florian Horn. Random Fruits on the Zielonka Tree. In Proceedings of the 26th International Symposium on Theoretical Aspects of Computer Science, STACS’09, volume 3 of Leibniz International Proceedings in Informatics, pages 541–552. Schloß Dagstuhl, 2009.
  16. Panganamala R. Kumar and Tzong-Huei Shiau. Existence of Value and Randomized Strategies in Zero-sum Discrete-time Stochastic Dynamic Games. SIAM Journal on Control and Optimization, 19(5):617–634, 1981.
  17. Donald A. Martin. The Determinacy of Blackwell Games. Journal of Symbolic Logic, 63(4):1565–1581, 1998.
  18. Zohar Manna and Amir Pnueli. The Temporal Logic of Reactive and Concurrent Systems: Specification. Springer-Verlag, 1992.
  19. Ashok P. Maitra and William D. Sudderth. Finitely additive stochastic games with Borel measurable payoffs. International Journal of Game Theory, 27(2):257–267, 1998.
  20. Amir Pnueli and Roni Rosner. On the Synthesis of a Reactive Module. In Proceedings of the 16th Annual ACM Symposium on Principles of Programming Languages, POPL’89, pages 179–190, 1989.
  21. Susan Thomson, Thomas Narten, and Tatuya Jinmei. IPv6 Stateless Address Autoconfiguration. RFC 4862 (Draft Standard), 2007.
  22. David Williams. Probability with Martingales. Cambridge University Press, 1991.
  23. Tatu Ylonen and Chris Lonvick. The Secure Shell (SSH) Transport Layer Protocol. RFC 4253 (Proposed Standard), 2006.
  24. Wieslaw Zielonka. Infinite Games on Finitely Coloured Graphs with Applications to Automata on Infinite Trees. Theoretical Computer Science, 200(1–2):135–183, 1998.
100153
This is a comment super asjknd jkasnjk adsnkj
Upvote
Downvote
Edit
-  
Unpublish
""
The feedback must be of minumum 40 characters
The feedback must be of minumum 40 characters
Submit
Cancel
0
Comments 0
""
The feedback must be of minumum 40 characters
Add comment
Cancel

You are asking your first question!
How to quickly get a good answer:
  • Keep your question short and to the point
  • Check for grammar or spelling errors.
  • Phrase it like a question