The Agafonov and Schnorr-Stimm Theorems for Probabilistic Automata
Abstract
For a fixed alphabet , an infinite sequence is said to be normal if every word over appears in with the same frequency as any other word of the same length. A classical result of Agafonov (1966) relates normality to finite automata as follows: a sequence is normal if and only if any subsequence of selected by a finite automaton is itself normal. Another theorem of Schnorr and Stimm (1972) gives an alternative characterization: a sequence is normal if and only if no gambler can win large amounts of money by betting on the sequence using a strategy that can be described by a finite automaton. Both of these theorems are established in the setting of deterministic finite automata. This raises the question as to whether they can be extended to the setting of probabilistic finite automata. In the case of the Agafonov theorem, a partial positive answer was given by Léchine et al. (MFCS 2024) in a restricted case of probabilistic automata with rational transition probabilities.
In this paper, we settle the full conjecture by proving that both the Agafonov and the Schnorr-Stimm theorems hold true for arbitrary probabilistic automata. Specifically, we show that a sequence is normal if and only if any probabilistic automaton selects a normal subsequence of with probability and also show that a sequence is normal if and only if any probabilistic finite-state gambler fails to win on with probability .
Keywords and phrases:
Normality, Agafonov theorem, probabilistic automataFunding:
Laurent Bienvenu: supported in part by the ANR project FLITTLA ANR-21-CE48-0023.Copyright and License:
2012 ACM Subject Classification:
Theory of computation Formal languages and automata theoryEditors:
C. Aiswarya, Ruta Mehta, and Subhajit RoySeries and Publisher:
Leibniz International Proceedings in Informatics, Schloss Dagstuhl – Leibniz-Zentrum für Informatik
1 Introduction
Given a finite alphabet of letters, an infinite sequence of letters is said to be normal if every word of appear as subword of with the same frequency as any other word of the same length, namely, . The famous Champernowne sequence
can be shown to be normal over the alphabet . The number , for example, is conjectured to have a normal expansion in every base, though this very much remains an open question. Normal sequences are plentiful, and an easy way to generate a normal sequence is to draw each letter at random in the alphabet (all letters having the same probability ) independently of the other chosen letters . The law of large numbers tells us that we obtain a normal sequence with probability .
Of course, there are also plenty of examples of non-normal sequences:
-
Periodic, or ultimately periodic, sequences can never be normal: indeed if the period of is there are only possible subwords of of length hence most words of length will not appear and their frequency will be .
-
Sturmian sequences, which are sequences with only different subwords of length (such as the Fibonacci sequence 100101001001010010100100101001001… obtained by iterating the morphism and ) are not normal for the same reason.
-
The Thue-Morse sequence obtained by iterating the morphism and is not normal because and do not appear as subwords.
-
A sequence of ’s and ’s generated at random where each bit is chosen equal to the previous one with probability will have, with probability , all possible finite words as subwords but will not (still with probability ) be normal as for example the word will appear with frequency instead of .
It turns out that normality has a nice interpretation in terms of finite automata. Indeed, two classical results, one due to Agafonov [1] and the other due to Schnorr and Stimm [9], assert that an infinite sequence is normal if and only if it cannot be predicted by a finite automaton with better-than-average accuracy. Of course, one needs to specify what “predicted” means. We consider two prediction models.
-
(I)
In the Agafonov model, an automaton reads the infinite sequence one letter at a time and updates its state in the usual way. Some of its states have a “select” tag on them. When the current state has such a tag, the next letter will be selected and added to a subsequence . We consider the automaton successful at predicting if the subsequence built in this process is infinite and some letter of does not have asymptotic frequency in . This means that the automaton has exhibited a statistical anomaly in the sequence and isolated this anomaly in the subsequence .
-
(II)
In the Schnorr-Stimm model, the predictor is still an automaton but this time is viewed as a gambling strategy. The gambler starts with a capital of . Each state is labeled with a betting function . This function represents the amount by which the predictor would like her capital to be multiplied by depending on the value of the next bit. For example, suppose the player plays against a sequence . If her current capital is and the current state is labelled by a betting function such that , and , if the next letter is , her new capital will be , if it is her new capital will be and if it is her new capital will be . For the game to be fair, each betting function must have expectation equal to , i.e., must satisfy . We say that the predictor wins if the capital of the player takes arbitrarily large values throughout the (infinite) game. That is, the predictor has spotted some type of statistical anomaly and is exploiting it to get rich!
Both of these models lead to the same conclusion: an infinite sequence is normal if and only if it is unpredictable (by a finite automaton).
Theorem 1 (Agafonov [1]).
For , the following are equivalent.
-
(i)
is normal.
-
(ii)
For any automaton that selects a subsequence of as in model I, either is finite, or every letter of appears in with asymptotic frequency .
-
(iii)
For any automaton that selects a subsequence of as in model I, either is finite, or is normal.
Theorem 2 (Schnorr-Stimm [9]).
For , the following are equivalent.
-
(i)
is normal.
-
(ii)
Any automaton betting on according to model II does not win.
In both of these theorems, the finite automata used for prediction are assumed to be deterministic. Would the situation change if one allowed probabilistic automata? In principle, one would not expect an unpredictable sequence to become predictable in the presence of a random source. Indeed, given a sequence and a random source it seems, informally speaking, that almost surely will not “know” anything about and thus will not help in predicting . Surprisingly, this intuition is wrong in the setting where the predictors are not finite automata but are Turing machines, as shown by Bienvenu et al. [2] who built a sequence that is unpredictable by deterministic Turing machines (in either prediction model of selection or gambling) and becomes predictable (in either model) if one allows probabilistic Turing machines. Nonetheless, finite automata are much weaker than Turing machines and Bienvenu et al.’s construction does not work for such a memoryless model of computation. Indeed, recently, Léchine et al. showed that Agafonov’s theorems holds for probabilistic automata in the restricted case where the transition probabilities are rational.
Theorem 3 (Léchine et al. [6]).
For , the following are equivalent.
-
(i)
is normal.
-
(ii)
For any probabilistic automaton with rational probabilities, almost surely, selects a subsequence of that is either finite, or every letter of appears in with asymptotic frequency .
-
(iii)
For any probabilistic automaton with rational probabilities, almost surely, selects a subsequence of such that either is finite, or is normal.
This led them to conjecture that the probabilistic version of Agafonov’s theorem holds in the general case, where the probabilities are real-valued rather than being rational. In this paper, we prove this conjecture and also prove the probabilistic version of the Schnorr-Stimm theorem. Additionally, we establish a probabilistic version of the Schnorr-Stimm dichotomy regarding the winning rates of probabilistic gamblers.
Léchine et al.’s proof is a reduction of the rational probabilistic case to the deterministic case by some clever combinatorial reduction (which substantially increases the number of states). Our proof is also a reduction to the deterministic case but is of very different nature: Instead of encoding a probabilistic automaton into a deterministic one, we will use the fact that a probabilistic automaton can be seen as a deterministic one over a larger alphabet, without altering the number of states. For this, we will need an extension of the deterministic case to Bernoulli measures (an extension which was proved by Seiller and Simonsen [10]), which will be presented in the next section.
Notation and terminology
We finish this introduction by formalizing the concepts discussed so far and gathering the notation and terminology that will be used in the rest of the paper.
Given an alphabet , we denote by the set of finite words over , by the set of words of length , by the set of infinite sequences of letters and by the set . For , we denote by the -th letter of (by convention there is a -th letter) and by the word . Let denote the empty string.
Given a word of length and a word of length , we denote by the number of occurrences of the word in , i.e.,
and the frequency of occurrence of in is naturally defined by
When is an infinite sequence, we define
When and have the same value, we simply call this common value .
Given , we say that is balanced if all letters appear in with the expected frequency, i.e., for all . We say that is normal if all words appear in with the expected frequency, i.e., for all .
A deterministic finite automaton (DFA) is a tuple where is a finite set of states, a finite alphabet, the initial state and the transition function (in this paper, runs of automata are meant to be infinite hence there is no need for final states). We denote by the function from to defined inductively by and for and , , where is the concatenation of words.
An automatic selector (or selector for short) is a tuple where is a DFA and is a subset of , representing the selection states.
Given a selector , we define the selection function from to inductively by:
and for and :
If is an infinite sequence in , the sequence of words is non-decreasing with respect to the prefix order and thus converges to a sequence which we call the selected subsequence of selected by and denote by .
An automatic gambler (or gambler for short) is a tuple where is a DFA and is a function from to such that for all , . As said earlier, the value of should be interpreted as the multiplier by which the gambler, being currently in state , would like her capital to be multiplied by if the next read letter is . The condition ensures that the game is fair.
Given a gambler and , we define inductively by
and for and ,
and we say that a gambler wins against if
(otherwise we say that loses).
2 Deterministic prediction for Bernoulli measures
In classical normality, all letters of the alphabet occur with the same frequency. We can however consider the extension of normality to Bernoulli measures. A Bernoulli measure over is a probability measure where letters of an infinite sequence are drawn at random independently of one another but the distribution over the alphabet is non-uniform.
Definition 4.
Let be a distribution over the alphabet (hence satisfying ). The Bernoulli measure induced by , which we also denote by by abuse of notation, is the unique probability measure such that for all , for every word ,
We also denote by the quantity .
Normality generalizes very naturally to Bernoulli measures.
Definition 5.
Let be a Bernoulli measure. A sequence is -balanced if for all . It is -normal if for all words , .
We say that a Bernoulli measure is positive when for every letter . In the rest of the paper, all Bernoulli measures will be assumed to be positive, and we simply say “Bernoulli measure” to mean “positive Bernoulli measure”.
The Agafonov theorem can be extended to Bernoulli measures, as proven by Seiller and Simonsen [10]. It is this theorem that we will use in the next section to obtain a proof of the Agafonov theorem for probabilistic selectors.
Theorem 6 (Agafonov theorem for Bernoulli measures [10]).
For , the following are equivalent.
-
(i)
is -normal.
-
(ii)
For any selector that selects a subsequence of , either is finite or is -balanced.
-
(iii)
For any selector that selects a subsequence of , either is finite or is -normal.
We can also easily generalize the notion of gambler to the setting of Bernoulli measures: it suffices to define a -gambler as before but with the fairness condition on replaced by for every . The function and the notion of success are defined as before.
We will now prove that the Schnorr-Stimm theorem, just like the Agafonov theorem, can also be extended to Bernoulli measures.
Theorem 7 (Schnorr-Stimm theorem for Bernoulli measures).
For and a Bernoulli measure , the following are equivalent.
-
(i)
is -normal.
-
(ii)
No -gambler wins by betting on .
Proof.
.
Suppose that is -normal and consider a -gambler .
We can assume that the -gambler only has one state on which it places a non-trivial bet. Indeed, define for every the -gambler where for all and for . That is, is the gambler where all states but state are neutralized (no bet is placed while on them). By the multiplicative nature of the function , we have for all :
Thus if we can show that all quantities are bounded, we are done. Let us assume that there is a state that is the unique state on which bets. If instead of we consider the selector with only as the selecting state, we know by Agafonov’s theorem for Bernoulli measures (Theorem 6) that the subsequence of selected by is -normal, hence in particular -balanced. But this subsequence is precisely the values of on which bets!
We can further assume that is infinite, otherwise it means that the run of on passes by finitely often, hence certainly cannot win as other states are not betting states. Now, suppose that at the -th step of the run on the state has been visited times. We have
But since is -normal we have, for every , . Thus,
or equivalently
(here we assume that all values involved in the product are positive for if not then the capital falls to and we are done). Since we have ( being a distribution), we can use the strict concavity of the function on to apply Jensen’s inequality and get
with strict inequality when not all are equal (which is the case where makes non-trivial bets). But by the fairness condition, we have hence we see that is either or ultimately negative which either way means that does not win.
.
Assume that is not -normal. This means that there is some word such that does not converge to . Let us assume that is such a word of minimal length and write with and .
Consider the sequence of vectors defined by
or equivalently,
All of these vectors belong to the set . This is a compact set, hence the sequence must have cluster points. By definition of and , we know that does not converge to because does not converge to : Indeed, in the definition of , the denominator converges to which by minimality of is defined and equal to , while the numerator is equal to which by definition of does not converge to .
Therefore, the sequence must have at least one cluster point different from . Fix such a cluster point .
We now build our gambler . The idea is that the gambler will record the last bits it read and will only place bets when these exactly form the word . Let us thus take initial state and define by
Now, define whenever and for all . Observe that this is a valid -gambler as the fairness condition is satisfied: .
Suppose that after reading letters of the state has been visited times. First, observe that tends to . Indeed, the state is visited whenever is seen as a subword of . But we assumed that appears in with frequency , by minimality of .
Second, unfolding the definition of , we have,
This gives
Since is a cluster point of the sequence , for any fixed there are infinitely many (or ) such that
But the term is the relative entropy from to (also known as Kullback-Liebler divergence, see for example [5]) which we denote by . This quantity is non-negative in general and is positive when , which is the case here. We have thus established that for any fixed , there are arbitrarily large and such that
Taking any , this shows that
Let us note that this last proof actually gives us a finer analysis of normality, in terms of the rate of failure or success of the gambler. This was already observed by Schnorr and Stimm in their seminal paper, where they proved the following.
Theorem 8 (Schnorr-Stimm dichotomy theorem [9]).
Let be an infinite sequence in .
-
(i)
If is normal and is a gambler, then the capital of throughout the game either is ultimately constant or decreases at an exponential rate.
-
(ii)
If is not normal, then there exists a gambler which wins against at an “infinitely often” exponential rate (i.e., ).
As a byproduct of our proof of Theorem 7, we have the same dichotomy for positive Bernoulli measures (i.e., Bernoulli measures such that for every letter):
Theorem 9 (Schnorr-Stimm dichotomy theorem for Bernoulli measures).
Let be an infinite sequence in and a positive Bernoulli measure.
-
(i)
If is -normal and is a -gambler, then the capital of throughout the game either is ultimately constant or decreases at an exponential rate.
-
(ii)
If is not -normal, then there exists a -gambler which wins against at an “infinitely often” exponential rate.
Our proof of Theorem 7 almost establishes this, but we do need an additional technical lemma and the block-wise characterization of -normality. Let denote the block frequency of the word in , defined as the proportion of non-overlapping blocks in that are equal to . More precisely when and ,
For , let , and denote the lower, upper and limit block frequency of in defined similarly as , and . A sequence is -block normal if for all words , . The following was shown by Seiller and Simonsen.
Lemma 10 (Seiller-Simonsen [10]).
If a sequence is -normal then is -block normal.
We note that the converse implication also holds.
Lemma 11.
If a sequence is -block normal then is -normal.
The above lemmas completes the proof of the following equivalence theorem between -normality and -block normality.
Theorem 12.
A sequence is -normal if and only if is -block normal.
Proof of Lemma 11.
As in the proof of Theorem 3.1 from [8], for any finite length string and large enough ,
| (1) |
where are defined as follows:
In the above definition, is the set of strings of the form, where is some string of length and is some string of length .
Equation 1 shows that the frequency of in can be written as a sum of different block frequencies. The quantity counts the number of occurrences of inside disjoint -length blocks in , counts the number of occurrences crossing the boundaries of these -blocks, counts those crossing boundaries of -blocks, and so on. In general, counts the number of occurrences straddling boundaries of disjoint -length blocks. Each may miss at most occurrences of , which are accounted for by the error term.
Since is -block normal,
Now, when ,
Since is uniformly convergent, we have
Therefore from (1),
Hence, is -normal.
We require the following technical lemma in the proof of Theorem 9.
Lemma 13.
Let be a finite-state automaton, be a positive Bernoulli measure and let denote the number of times the state is visited upon running the automaton using the first bits of .
Then, for every , there exists a real number such that, for every -normal sequence :
-
either is ultimately constant (i.e., is visited only finitely often during the run on )
-
or, (i.e., the state is visited with asymptotic frequency ) and this second case can only happen when .
Proof.
On running an automaton upon a normal sequence, starting from any state, a strongly connected component must be reached in finitely many steps. Similar to [10], let us consider the Markov chain corresponding to the stochastic matrix where,
The proof follows using the Ergodic Theorem for Markov chains and the same steps in the proof of Lemma 4.5 from [9] by replacing the uniform measure with the positive Bernoulli measure induced by . We note that this line of proof uses -block normality, which is equivalent to -normality (Theorem 12).
Proof of Theorem 9.
In part of our proof of Theorem 7, we showed that on a given state , either is not a betting state ( is the constant on this state) or it is and then the gambler loses money exponentially fast in the number of times this state is visited, the exponent being which we proved to be negative. By Lemma 13, betting states are either visited finitely often or with positive asymptotic density. If they are all visited finitely often, the capital stabilizes after the last bet is made. Otherwise, if betting states are visited with frequency and at least one is positive, then the gambler loses at an exponential rate, where the exponent is .
In part , under the assumption of non--normality of , we built a gambler satisfying the following: for any fixed , there are arbitrarily large and such that . Here, is the number of visits to a state . In the proof of Theorem 7 part , this state is visited with frequency for large enough . Hence the gambler has an “infinitely often” exponential rate of success with exponent .
3 The Agafonov theorem for PFAs
We now want to prove the extension of Theorem 6 to probabilistic automata/selectors. A probabilistic finite automaton (PFA) is a tuple where is a finite set of states, is a finite alphabet, is the initial state and is a probabilistic transition function, that is, is the set of probability distributions over . In this setting, we define inductively the random variables by and for and , the event is defined as the union
where is a family of independent random variables such that for all , the distribution of is .
Modulo this change of type of transition, probabilistic selectors as well as are defined as before. This makes a random variable for every given .
In order to lift the deterministic Agafonov theorem for Bernoulli measures to the probabilistic case, we will need some preliminary lemmas about normality.
Given two alphabets and , and given two sequences and , we denote by the sequence over the alphabet where . For two words and of the same length over and respectively, the product is defined in the same way. Likewise, if and are Bernoulli measures over and respectively, is the Bernoulli measure over where , for all . Finally, if is a sequence over a product alphabet , we denote by and its first and second projection respectively (in other words, and for all ).
Lemma 14.
Let and be two alphabets and , two Bernoulli measures over and respectively. If a sequence is -normal and a sequence is drawn at random according to , then -almost surely111For a probability 1 set of according to the measure ., is -normal.
Proof.
While it is a bit easier to prove this lemma for block-normality, we prefer a more direct proof which does not rely on the equivalence between normality and block-normality.
Let be a non-empty word over . Let be a large integer. We split into blocks of length :
with , for all . Introduce the random variables which count the number of occurrences of in each block:
For every integer , within there are complete blocks. Some occurrences of in do occur inside a block while some other do not because they overlap two contiguous blocks. There can be at most such overlapping occurrences between two given blocks. That observation leads to two ways to count the number of occurrences of : the exact way and the -block way , two random variables satisfying:
| (2) |
Let us focus on first. The variables can be grouped into different buckets, with respect to the corresponding value of . For every , set , which is non-empty for large enough, since is normal. The random variables are mutually independent, and so are the random variables . Moreover, for a fixed , the random variables are distributed identically, denote by the corresponding probability distribution on . Then measures the expected number of occurrences of in a sequence where the left part is fixed, equal to , and the right part is independently generated according to . Thus
According to the Law of Large Numbers, . Since is normal, every word occurs with frequency in thus, almost-surely,
| (3) |
Since is normal, the right part in (3) converges to when grows large. Using (2), we get the desired result:
Lemma 15.
If a sequence is -normal over , then and are -normal and -normal respectively.
Proof.
Suppose is -normal. We only need to show that is -normal, the proof of the -normality of is the same by symmetry. Let . We have
which, by -normality of , implies
The same holds for , hence we have proven that , which is what we wanted.
We are now ready to prove Agafonov’s theorem for PFAs.
Theorem 16 (Agafonov’s theorem for PFA).
Let , and a Bernoulli measure over . The following are equivalent.
-
(i)
is -normal.
-
(ii)
For any deterministic selector that selects a subsequence of , either is finite or is -normal.
-
(iii)
For any probabilistic selector that selects a subsequence of , almost surely, either is finite, or is -normal.
Proof.
Since deterministic selectors (for which we already have Agafonov’s theorem) are a subset of the probabilistic ones, all that is left to prove is .
Let be a probabilistic selector. Recall that each transition (where and ) is some probability distribution over . Consider the set of all functions . We can put a distribution over such that if is chosen according to , for every , the marginal distribution of is . An easy way to do this is to take .
This construction means that, for a fixed sequence , an equivalent way to simulate the probabilistic run of on is, every time we are in a state and read a letter , to pick at random according to and move to state . But is a finite set and a Bernoulli measure over it. This Bernoulli measure might not be positive. If it is not, we simply remove from all functions whose -probability is . Reformulating slightly, yet another equivalent way to simulate the run of over is to do the following:
-
1.
First choose at random with respect to the Bernoulli measure .
-
2.
Then run on the sequence the deterministic selector whose set of states is and transition is defined by .
Now, the two following random variables have the same distribution:
-
The subsequence of selected by .
-
The sequence , where is the subsequence of selected by when is chosen randomly according to .
Since is -normal, by Lemma 14, is -normal -almost surely. Thus, by Agafonov’s theorem for deterministic selectors and Bernoulli measures (Theorem 6), almost surely, the subsequence selected by is -normal. Finally, by Lemma 15, this implies that almost surely, the subsequence of is -normal. This concludes the proof.
4 The Schnorr-Stimm theorem for probabilistic gamblers
Finally, we establish the Schnorr-Stimm theorem for probabilistic finite-state automata. The definition of probabilistic (-)gambler is the same as the definition of deterministic (-)gambler except that it is based on PFAs instead of DFAs. In this setting, the quantities become random variables.
Theorem 17 (Schnorr-Stimm theorem for PFAs).
For and a Bernoulli measure, the following are equivalent.
-
(i)
is -normal.
-
(ii)
Any probabilistic -gambler betting on loses almost surely.
Proof.
We already have by Theorem 7 (deterministic gamblers being a subset of probabilistic ones).
For , the idea is similar to the proof of Theorem 16. To simulate the run of a probabilistic -gambler on a sequence , one can equivalently run on the sequence (where is -random sequence)222 and are defined in the same way as in Theorem 16. the deterministic gambler where again for all and (indeed the bet placed by a gambler at a given stage does not take into account which state will be reached next). Note that the fairness condition is respected since for every ,
(the second-to-last equality holds by fairness condition on and the last one because is a distribution).
In this way, the two following random variables will have the same distribution:
-
.
-
where is chosen at random according to .
Since is -normal, by Lemma 14 again, is -normal -almost surely. Thus, by Theorem 7, for -almost all , is bounded by a constant independent on . By the equivalence of the two random variables above, this means that almost surely, there exists a constant such that for all . In other words, almost surely, loses on the sequence .
We remark that the dichotomy theorem for Bernoulli measures yields the following dichotomy for probabilistic -gamblers.
Theorem 18 (Schnorr-Stimm dichotomy theorem for PFAs).
Let be an infinite sequence in and a Bernoulli measure.
-
(i)
If is -normal and is a -gambler, then almost surely the capital of throughout the game either is ultimately constant or decreases at an exponential rate.
-
(ii)
If is not -normal, then there exists a -gambler which wins against at an “infinitely often” exponential rate almost surely.
5 Conclusion
The main contributions of this paper are a generalization of the Agafonov theorem for PFA with arbitrary transition probabilities which settles the open question posed by Léchine et al. [6], and an extension of Schnorr-Stimm theorem to probabilistic gamblers.
While we proved the probabilistic Agafonov theorem (Theorem 16) by reduction to the deterministic setting, it is also possible to follow with a more direct approach (i.e., without appealing to the Seiller-Simonsen result), similar to the one followed by Carton to generalize Agafonov’s theorem for DFA [3]. This however makes the argument somewhat more complicated.
An interesting direction for future research is to explore whether the “uselessness of randomness” also holds for pushdown automata. These are a more powerful model of computation and indeed some normal sequences can be predicted by pushdown automata (some in a rather dramatic way, as proven by Carton and Perifel [4]333The result proven by Carton and Perifel considers a slightly different paradigm, namely compression (which we did not discuss in this paper) instead of prediction.). We can for example ask: If some probabilistic pushdown selector selects a biased subsequence from a sequence , does there necessarily exist a deterministic pushdown selector which also selects a biased subsequence? Similarly, if some probabilistic pushdown gambler wins against a sequence , does there necessarily exist a deterministic pushdown gambler which wins against that same sequence ?
A related question concerns the speed of success in the gambling model. In the case of finite-state automata, Schnorr and Stimm proved that either a sequence cannot be predicted or some gambler wins on it at an exponential rate. This dichotomy no longer holds for pushdown automata, but one may ask the following question: If some probabilistic pushdown gambler wins at an exponential rate against a sequence , does there necessarily exist a pushdown gambler which wins against that sequence at an exponential rate?
References
- [1] V. N. Agafonov. Normal sequences and finite automata. Soviet Mathematics Doklady, 9:324–325, 1968.
- [2] Laurent Bienvenu, Valentino Delle Rose, and Tomasz Steifer. Probabilistic vs deterministic gamblers. In 39th International Symposium on Theoretical Aspects of Computer Science, STACS 2022, volume 219 of LIPIcs, pages 11:1–11:13. Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2022. doi:10.4230/LIPICS.STACS.2022.11.
- [3] Olivier Carton. A direct proof of Agafonov’s theorem and an extension to shift of finite type. CoRR, abs/2005.00255, 2020. arXiv:2005.00255.
- [4] Olivier Carton and Sylvain Perifel. Deterministic pushdown automata can compress some normal sequences. Logical Methods in Computer Science, Volume 20, Issue 3, 2024.
- [5] T.M. Cover and J.A. Thomas. Elements of Information Theory. Wiley, 2nd edition edition, 2006.
- [6] Ulysse Léchine, Thomas Seiller, and Jakob Grue Simonsen. Agafonov’s theorem for probabilistic selectors. In 49th International Symposium on Mathematical Foundations of Computer Science, MFCS 2024, volume 306 of LIPIcs, pages 67:1–67:15. Schloss Dagstuhl – Leibniz-Zentrum für Informatik, 2024. doi:10.4230/LIPICS.MFCS.2024.67.
- [7] John E. Maxfield. A short proof of Pillai’s theorem on normal numbers. Pacific Journal of Mathematics, 2(1):23–24, 1952.
- [8] Satyadev Nandakumar, Subin Pulari, Prateek Vishnoi, and Gopal Viswanathan. An analogue of Pillai’s theorem for continued fraction normality and an application to subsequences. Bulletin of the London Mathematical Society, 53(5):1414–1428, 2021.
- [9] Claus Schnorr and Hermann Stimm. Endliche Automaten und Zufallsfolgen. Acta Informatica, 1(4):345–359, 1972. doi:10.1007/BF00289514.
- [10] Thomas Seiller and Jakob Grue Simonsen. Agafonov’s theorem for finite and infinite alphabets and probability distributions different from equidistribution. Ergodic Theory and Dynamical Systems, 45(11):3506–3539, 2025.
