SciELO - Scientific Electronic Library Online

 
vol.5 número1A mathematically assisted reconstruction of the initial focus of the yellow fever outbreak in Buenos Aires (1871) índice de autoresíndice de materiabúsqueda de artículos
Home Pagelista alfabética de revistas  

Servicios Personalizados

Revista

Articulo

Indicadores

  • No hay articulos citadosCitado por SciELO

Links relacionados

  • No hay articulos similaresSimilares en SciELO

Compartir


Papers in physics

versión On-line ISSN 1852-4249

Pap. Phys. vol.5 no.1 La Plata jun. 2013

http://dx.doi.org/10.4279/PIP.050001 

 

LT2C2: A language of thought with Turing-computable Kolmogorov complexity

 

Sergio Romano,1* Mariano Sigman,2'3 Santiago Figueira1'3*

*E-mail: sgromano@dc.uba.ar
E-mail: sigman@df.uba.ar
E-mail: santiago@dc.uba.ar

1 Department of Computer Science, FCEN, University of Buenos Aires, Pabell´on I, Ciudad Universitaria (C1428EGA) Buenos Aires, Argentina.
2 Laboratory of Integrative Neuroscience, Physics Department, FCEN, University of Buenos Aires, Pabellon I, Ciudad Universitaria (C1428EGA) Buenos Aires, Argentina.
3 CONICET, Argentina.

 

In this paper, we present a theoretical effort to connect the theory of program size to psychology by implementing a concrete language of thought with Turing-computable Kolmogorov complexity (LT2C2) satisfying the following requirements: 1) to be simple enough so that the complexity of any given finite binary sequence can be computed, 2) to be based on tangible operations of human reasoning (prínting, repeating,...), 3) to be sufficiently powerful to genérate all possible sequences but not too powerful as to identify regularices which would be invisible to humans. We first formalize LT2C2, giving its syntax and semantics and defining an adequate notion of program size. Our setting leads to a Kolmogorov complexity function relative to LT2C2 which is computable in polynomial time, and it also induces a prediction algorithm in the spirit of Solomonoff’s inductive infer-ence theory. We then prove the efficacy of this language by investigating regularities in strings produced by participants attempting to genérate random strings. Participants had a profound understanding of randomness and henee avoided typical misconceptions such as exaggerating the number of alternations. We reasoned that remaining regularities would express the algorithmic nature of human thoughts, revealed in the form of specific patterns. Kolmogorov complexity relative to LT2C2 passed three expected tests examined here: 1) human sequences were less complex than control PRNG sequences, 2) human sequences were not stationary, showing decreasing valúes of complexity resulting from fatigue, 3) each individual showed traces of algorithmic stability since fitting of partial sequences was more effective to predict subsequent sequences than average fits. This work extends on previous efforts to combine notions of Kolmogorov complexity theory and algorithmic information theory to psychology, by explicitly proposing a language which may describe the patterns of human thoughts.

 

I. Introduction

Although people feel they understand the concept of randomness 1, humans are unable to produce random sequences, even when instructed to do so 2-6, and to perceive randomness in a way that is inconsistent with probability theory 7-10. For instance, random sequences are not perceived by participants as such because runs appear too long to be random 11,12 and, similarly, sequences produced by participants aiming to be random have too many alternations 13, 14. This bias, known as the gambler’s fallacy, is thought to result from an expectation of local representativeness (LR) of randomness 10 which ascribes chance to a self-correcting mechanism, promptly restoring the balance whenever disrupted. In words of Tversky and Kahneman 5, people apply the law of large num-bers too hastily, as if it were the law of small num-bers. The gambler’s fallacy leads to classic psycho-logical illusions in real-world situations such as the hot hand perception by which people assume spe-cific states of high performance, while analysis of records show that sequences of hits and misses are largely compatible with Bernoulli (random) process 15, 16.

Despite massive evidence showing that perception and productions of randomness shows system-atic distortions, a mathematical and psychological theory of randomness remains partly elusive. From a mathematical point of view -as discussed below- a notion of randomness for finite sequences presents a major challenge.

From a psychological point of view, it remains difficult to ascribe whether the inability to produce and perceive randomness adequately results from a genuine misunderstanding of randomness or, in-stead, as a consequence of the algorithmic nature of human thoughts which is revealed in the forms of patterns and, hence, in the impossibility of producing genuine chance.

In this work, we address both issues by developing a framework based on a specific language of thought by instantiating a simple device which induces a computable (and efficient) definition of algorithmic complexity 17-19.

The notion of algorithmic complexity is de-scribed in greater detail below but, in short, it as-signs a measure of complexity to a given sequence as the length of the shortest program capable of producing it. If a sequence is algorithmically com-pressible, it implies that there may be a certain pattern embedded (described succinctly by the pro-gram) and hence it is not random. For instance, the binary version of Champernowne’s sequence 20

01101110010111011110001001101010111100 ...

consisting of the concatenation of the binary rep-resentation of all the natural numbers, one after another, is known to be normal in the scale of 2,

which means that every finite word of length n oc-curs with a limit frequency of 2-n —e.g., the string 1 occurs with probability 2-1, the string 10 with probability 2-2, and so on. Although this sequence may seem random based on its probability distri-bution, every prefix of length n is produced by a program much shorter than n.

The theory of program size, developed si-multaneously in the ’60s by Kolmogorov 17, Solomonoff 21 and Chaitin 22, had a major influ-ence in theoretical computer science. Its practical relevance was rather obscure because most notions, tools and problems were undecidable and, overall, because it did not apply to finite sequences. A problem at the heart of this theory is that the com-plexity of any given sequence depends on the chosen language. For instance, the sequence

x1 = 1100101001111000101000110101100110011100

which seems highly complex, may be trivially ac-counted by a single character if there is a symbol (or instruction of a programming language) which accounts for this sequence. This has its psycholog-ical analog in the kind of regularities people often extract:

x2 = 1010101010101010101010101010101010101010

is obviously a non-random sequence, as it can suc-cinctly be expressed asrepeat20times: print ‘10’.                (1)

Instead, the sequence

x3 = 0010010000111111011010101000100010000101

appears more random and yet it is highly compress-ible as it consists of the first 40 binary digits of π after the decimal point. This regularity is sim-ply not extracted by the human-compressor and demonstrates how the exceptions to randomness re-veal natural patterns of thoughts 23.

The genesis of a practical (computable) algorith-mic information theory 24 has had an influence (although not yet a major impact) in psychology. Variants of Kolmogorov complexity have been ap-plied to human concept learning 25, to general theories of cognition 26 and to subjective random-ness 23, 27. In this last work, Falk and Konold showed that a simple measure, inspired in algorithmic notions, was a good correlate of perceived ran-domness 27. Griffiths & Tenenbaum developed statistical models that incorporate the detection of certain regularities, which are classified in terms of the Chomsky hierarchy 23. They showed the exis-tence of motifs (repetition, symmetry) and related their probability distributions to Kolmogorov complexity via Levin’s coding theorem (cf. section VII. for more details).

The main novelty of our work is to develop a class of specific programming languages (or Turing machines) which allows us to stick to the theory of pro-gram size developed by Kolomogorov, Solomonoff and Chaitin. We use the patterns of sequences of humans aiming to produce random strings to fit, for each individual, the language which captures these regularities.

II. Mathematical theory of randomness

The idea behind Kolmogorov complexity theory is to study the length of the descriptions that a formal language can produce to identify a given string. All descriptions are finite words over a finite alphabet, and henee each description has a finite length or, more generally a suitable notion of size. One string may have many descriptions, but any description should describe one and only one string. Roughly, the Kolmogorov complexity 17 of a string x is the length of the shortest description of x. So a string is ‘simple’ if it has at least one short description, and it is ‘complex’ if all its descriptions are long. Random strings are those with high complexity.

As we have mentioned, Kolmogorov complexity uses programming languages to describe strings. Some programming languages are Turing complete, which means that any partial computable function can be represented in it. The commonly used programming languages, like C++ or Java, are all Turing complete. However, there are also Turing in-complete programming languages, which are less powerful but more convenient for specific tasks.

In any reasonable imperative language, one can describe x2 above with a program like (1), of length 26, which is considerably smaller than 40, the size of the described string. It is clear that x2 is ‘simple’. The case of x3 is a bit tricky. Although at first sight it seems to have a complete lack of struc-ture, it contains a hidden pattern: it consists of the first forty binary digits of tt after the decimal point. This pattern could hardly be recognized by the reader, but once it is revealed to us, we agree that x3 must also be tagged as ‘simple’. Observe that the underlying programming language is central: x3 is ‘simple’ with the proviso that the language is strong enough to represent (in a reasonable way) an algorithm for computing the bits of tt —a language to which humans are not likely to have access when they try to find patterns in a string. Finally, for x1 the best way to describe it seems to be something like

print ‘1100101001111000101000110101100110011100’,

which includes the string in question verbatim, length 48. Henee x1 only has long descriptions and henee it is ‘complex’.

In general, both the string of length n which al-ternates Os and ls and the string which consists of the first n binary digits of tt after the decimal point can be computed by a program of length « log n — and this applies to any computable sequence. The idea of the algorithmic randomness theory is that a truly random string of length n necessarily needs a program of length « n (cf. section ii. for details).

i. Languages, Turing machines and Kolmogorov complexity

Any programming language C can be formalized with a Turing machine M¿, so that programs of C are represented as inputs of Mc via an adequate binary codification. If C is Turing complete then the corresponding machine Mc is called universal, which is equivalent to say that Mc can simúlate any other Turing machine.

Let {0,1}* denote the set of finite words over the binary alphabet. Given a Turing machine M, a program p and a string x (p, x G {0,1}*), we say that p is an M-description of x if M{p) = x — Le., the program p, when executed in the machine M, computes x. Here we do not care about the time that the computation needs, or the memory it consumes. The Kolmogorov complexity of x G {0,1}* relative to M is defined by the length of the shorter M-description of x. More formally,

where \p\ denotes the length of p. Here M is any given Turing machine, possibly one with a very specific behavior, so it may be the case that a given string x does not have any M-description at all. In this case, M(x) = oo. In practical terms, a machine M is a useful candidate to measure complexity if it computes a surjective function. In this case, every string x has at least one M-description and there-fore KM(x) < oo.

ii. Randomness for finite words

The strength of Kolmogorov complexity appears when M is set to any universal Turing machine U. The invariance theorem states that Kv is minimal, in the sense that for every Turing machine M there is a constant cM such that for all x G {0,1}* we have Ku(x) < KM{c) + cM- Here, cM can be seen as the specification of the language M in U (i.e., the information contained in cM tells U that the machine to be simulated is M). If U and U' are two universal Turing then Kv and Kw differ at most by a constant. In a few words, Kv{x) represents the length of the ultimate compressed versión of x, performed by means of algorithmic processes.

For analysis of arbitrarily long sequences, cM be-comes negligible and henee for nonpractical aspeets of the theory the choice of the machine is not rel-evant. However, for short sequences, as we study here, this becomes a fundamental problem, as no-tions of complexity are highly dependent on the choice of the underlying machine through the constant cM- The most trivial example, as referred in the introduction, is that for any given sequence, say xi, there is a machine M for which xx has minimal complexity.

iii. Solomonoff induction

Here we have presented compression as a frame-work to understand randomness. Another very in-fluential paradigm proposed by Schnorr is to use the notion of marüngale (roughly, a betting strat-egy), by which a sequence is random if there is no computable martingale capable of predicting fortheoming symbols (say, of a binary alphabet {0,1}) better than chance 28,29. In the 1960s, Solomonoff 21 proposed a universal prediction method which successfully approximates any dis-tribution /x, with the only requirement of ¡j, being computable.

This theory brings together concepts of algorithmic information, Kolmogorov complexity and prob-

ability theory. Roughly, the idea is that amongst all ‘explanations’ of x, those which are ‘simple’ are more relevant, henee following Occam’s razor principie: amongst all hypothesis that are consistent with the data, choose the simplest. Here the ‘explanations’ are formalized as programs computing x, and ‘simple’ means low Kolmogorov complexity.

Solomonoff’s theory, builds on the notion of monotone (and prefix) Turing machines. Mono-tone machines are ordinary Turing machines with a one-way read-only input tape, some work tapes, and a one-way write-only output tape. The out-put is written one symbol at a time, and no eras-ing is possible in it. The output can be finite if the machine halts, or infinite in case the machine computes forever. The output head of monotone machines can only “print and move to the right” so they are well suited for the problem of inference of fortheoming symbols based on partial (and finite) states of the output sequence. Any monotone machine N has the monotonicüy property (henee its ñame) with respect to extensión: if p, q G {0,1}* then N(p) is a prefix of N(p~q), where pq denotes the concatenation of p and q.

One of Solomonoff’s fundamental results is that given a finite observed sequence x G {0,1}*, the most likely finite continuation is the one in which the concatenation of x and y is less complex in a Kolmogorov sense. This is formalized in the following result (see theorem 5.2.3 of 24): for almost all infinite binary sequences X (in the sense of /x) we have

Here, X\n represents the first n symbols of X, and Kmu is the monotone Kolmogorov complexity rel-ative to a monotone universal machine U. That is, Kmu(x) is defined as the length of the shortest programp such that the output of U{p) starts with x —and possibly has a (finite or infinite) continuation.

In other words, Solomonoff inductive inference leads to a method of prediction based on data compression, whose idea is that whenever the source has output the string x, it is a good heuristic to choose the extrapolation y of x that minimizes Kmu{xry). For instance, if one has observed x2, it is more likely for the continuation to be 1010 rather than 0101, as the former can be succinctly described by a program like

repeat 22 times: print ‘10’.             (2)

and the latter looks more difficult to describe; in-deed the shorter program describing it seems to be something like

repeat 20 times: print ‘10’;            (3)

print ‘0101’.

Intuitively, as program (2) is shorter than (3), x2~1010 is more probable than x2~0101. Henee, ií we have seen x2, it seems to be a better strategy to predict 1.

III. A framework for human thoughts

The notion oí thought is not well grounded. We lack an operative working definition and, as also happens with other terms in neuroscience (con-sciousness, sel!, ...), the word thought is highly pol-ysemic in common language. It may reíer, for ex-ample, to a belieí, to an idea or to the contení oí the conscious mind. Due to this difficulty, the mere notion oí thought has not been a principal or directed object oí study in neuroscience, although oí course it is always present implicitly, vaguely, without a formal definition.

Here we do not intend to elabórate an extensive review on the philosophical and biological concep-tions of thoughts (see 30 for a good review on thoughts). Ñor are we in a theoretical position to provide a full formal definition of a thought. Instead, we point to the key assumptions of our framework about the nature of thoughts. This accounts to defining constraints in the class of thoughts which we aim to describe. In other words, we do not claim to provide a general theory of human thoughts (which is not amenable at this stage lacking a full definition of the class) but rather of a subset of thoughts which satisfy certain constraints defined below.

For instance, E.B. Titchener and W. Wundt, the founders of structuralist school in psychology (seek-ing structure in the mind without evoking meta-physical conceptions, a tradition which we inherit and to which we adhere), believed that thoughts

were images (there are not imageless thoughts) and hence can be broken down to elementary sensations 30. While we do not necessarily agree with this propositions (see Carey 31 for more contemporary versions denying the sensory foundations of conceptual knowledge), here we do not intend to explain all possible thoughts but rather a subset, a sim-pler class which —in agreement with the Wundt and Titchener— can be expressed in images. More precisely, we develop a theory which may account for Boole’s 32 notion of thoughts as propositions and statements about the world which can be rep-resented symbolically. Hence, a first and crucial assumption of our framework is that thoughts are discrete. Elsewhere we have extensively discussed 33-39 how the human brain, whose architecture is quite different from Turing machines, can emerge in a form of computation which is discrete, symbolic and resembles Turing devices.

Second, here we focus on the notion of “prop-less” mental activity, i.e., whatever (symbolic) com-putations can be carried out by humans without resorting to external aids such as paper, marbles, computers or books. This is done by actually asking participants to perform the task “in their heads”. Again, this is not intended to set a proposi-tion about the universality of human thoughts but, instead, a narrower set of thoughts which we con-ceive is theoretically addressable in this mathemat-ical framework.

Summarizing:

1.  We think we do not have a good mathematical (even philosophical) conception of thoughts, as mental structures, yet.

2.  Intuitively (and philosophically), we adhere to a materialistic and computable approach to thoughts. Broadly, one can think (to pic-ture, not to provide a formal framework) that thoughts are formations of the mind with certain stability which defines distinguishable clusters or objects 40-42.

3.  While the set of such objects and the rules of their transitions may be of many different forms (analogous, parallel, unconscious, un-linked to sensory experience, non-linguistic, non-symbolic), here we work on a subset of thoughts, a class defined by Boole’s attempt to formalize thought as symbolic propositions about the world.

4.  This states which may correspond to human “conscious rational thoughts”, the seed of Boole and Turing foundations 34,34 are discrete and defined by symbols and poten-tially represented by a Turing device.

5.  We focus on an even narrower space of thoughts. Binary formations (right or left, zero or one) to focus on what kind of language bet-ter describes these transitions. This work can be naturally extended to understand discrete transitions in conceptual formations 43-45.

6.  We concéntrate on prop-less mental activity to understand limitations of the human mind when it does not have evident external support (paper, computer...)

IV. Implementing a language of thought with Turing-computable complexity

As explained in section II.L, Kolmogorov complexity considers all possible computable compressors and assigns to a string x the length of the short-est of the corresponding compressions. This seems to be a perfect theory of compression but it has a drawback: the function Kv is not computable, that is, there is no effective procedure to calcúlate Ku(x) given x.

On the other hand, the definition of randomness introduced in section II.L, having very deep and intricate connections with algorithmic information and computability theories, is simply too strong to explain our own perception of randomness. To de-tect that x3 consists of the first twenty bits of tt is incompatible with human patterns of thought.

Henee, the intrinsic algorithms (or observed patterns) which make human sequences not random are too restricted to be accounted by a universal machine and may be better described by a specific machine. Furthermore, our hypothesis is that each person uses his own particular specific machine or algorithm to genérate a random string.

As a first step in this complicated enterprise, we propose to work with a specific language LT2C2 which meets the following requirements:

•  LT2C2 must reflect some plausible features of our mental activity when finding succinct de-scriptions of words. For instance, finding rep-etitions in a sequence such as x2 seems to be something easy for our brain, but detecting nu-merical dependencies between its digits as in x3 seems to be very unlikely.

•  LT2C2 must be able to describe any string in {0,1}*. This means that the map given by the induced machine N = NLT2C2 must be surjective.

•  N must be simple enough so that KN —the Kolmogorov complexity relative to be-comes computable. This requirement clearly makes LT2C2 Turing incomplete, but as we have seen before, this is consistent with human deviations from randomness.

•  The rate of compression given by KN must be sensible for very short strings, since our exper-iments will produce such strings. For instance, the approach, followed in 46, of using the size of the compressed file via general-purpose compressors like Lempel-Ziv based dictionary (gzip) or block based (bzip2) to approximate the Kolmogorov complexity does not work in our setting. This method works best for long files.

•  LT2C2 should have certain degrees of freedom, which can be adjusted in order to approximate the specific machine that each individual fol-lows during the process of randomness gener-ation.

We will not go into the details on how to codify the instructions of LT2C2 into binary strings of N: for the sake of simplicity we take ¡Vasa surjective total mapping LT2C2 ->• {0,1}*. We restrict ourselves to describe the grammar and semantics of our proposed programming language LT2C2. It is basically an imperative language with only two classes of instructions: a sort of print i, which prints the bit i in the output; and a sort of repeat n times P, which for a fixed n G N it repeats n times the program P. The former is simply represented as i and the latter as (P)n.

Formally, we set the alphabet {0,1, (,),° ,... ,9 } and define LT2C2 over such alphabet with the following grammar:

i. Kolmogorov complexity for LT2C2

The Kolmogorov complexity relative to N (and henee to the language LT2C2) is defined as

KN(x) = mm{\\p\\: p G LT2C2, N(p) = x},

where ||p||, the size of a program p, is inductively defined as:

operation and the repeat n times operation. In the sequel, we drop the subindex of KN and simply write K = KN. Table 1 shows some examples of the size of JV-programs wnen b = r = 1. Observe that for all x we have K(x) < \\x\\.

It is not difiicult to see that K(x) depends only on the valúes of K(y), where y is any nonempty and proper substring of x. Since || • || is computable in polynomial time, using dynamic programming one can calcúlate K(x) in polynomial time. This, of course, is a major difference with respect to the Kolmogorov complexity relative to a universal machine, which is not computable.

ii. Prom compression to prediction

As one can imagine, the perfect universal prediction method described in section Il.iii. is, again, non-computable. We define a computable prediction algorithm based on Solomonoff’s theory of induc-tive inference but using K, the Kolmogorov complexity relative to LT2C2, instead of Kmv (which depends on a universal machine). To predict the next symbol of x G {0,1}*, we follow the idea described in section Il.iii.: amongst all extrapolations y of x we choose the one that minimizes K{x~y). If such y starts with 1, we predict 1, else we predict 0. Since we cannot examine the infinitely many extrapolations, we restrict to those up to a fixed given length lF. Also, we do not take into account the whole x but only a suffix of length tP. Both £F and tP are parameters which control, respectively, how many extrapolation bits are examined {tF many Future bits) and how many bits of the tail of x (£P many Past bits) are considered.

Let {0,1}™ (resp. {0,1}^™) be the set of words over the binary alphabet {0,1} of length n (resp. at most n). Formally, the prediction method is as follows. Suppose x = xx---xn (x¿ G {0,1}) is a string. The next symbol is determined as follows:

and g : {O, í}£p -> {O,1} is defined as g{z) = i if the number of occurrences of i in z is greater than the number of occurrences of 1 - i in z; in case the number of occurrences of ls and Os in z coincide then g(z) is defined as the last bit of z.

V.     Methods

Thirty eight volunteers (mean age = 24) partici-pated in an experiment to examine the capacity of LT2C2 to identify regularities in production of bi-nary sequences. Participants were asked to produce random sequences, without further instruction.

All the participants were college students or graduates with programming experience and knowledge of the theoretical foundations of ran-domness and computability. This was intended to test these ideas in a hard sample where we did not expect typical errors which results from a misun-derstanding of chance.

The experiment was divided in four blocks. In each block the participant pressed freely the left or right arrow 120 times.

After each key press, the participant received a notification with a green square which progressively filled a line to indicate the participant the number of choices made. At the end of the block, participants were provided feedback of how many times the predictor method has correctly predicted their input. After this point, a new trial would start.

38 participants performed 4 sequences, yielding a total of 152 sequences. 14 sequences were excluded from analysis because they had an extremely high level of predictability. Including these sequences would have actually improved all the scores re-ported here.

The experiment was programmed in Action-Script and can be seen at http://gamesdata. lafhis-server.exp.dc.uba.ar/azarexp.

VI.     Results

i. Law of large numbers

Any reasonable notion of randomness for strings on base 2 should imply Borel’s normality, or the law of large numbers in the sense that if x G {0,1}™ is random then the number of occurrences of any given string y in x divided by n should tend to 2-1^1, as n goes to infinity.

A well-known result obtained in some investiga-tions on generation or perception of randomness in binary sequences is that people tend to increase the number of alternations of symbols with respect to the expected valué 27. Given a string x of length n with r runs, there are n - 1 transitions between successive symbols and the number of alternations between symbol types is r - 1. The probaUUty of alternation of the string x is defined as

In our experiment, the average P(x) of participants was 0.51, very cióse to the expected probaUUty of alternation of a random sequence which should be 0.5. A t-test on the P(x) of the strings produced by participants, where the nuil hypothesis is that they are a random sample from a normal distribution with mean 0.5, shows that the hypothesis cannot be rejected as the p-valué is 0.31 and the confidence in-terval on the mean is 0.49, 0.53. This means that the probaUUty of alternation is not a good mea-sure to distinguish participant’s strings from random ones, or at least, that the participants in this very experiment can bypass this validation.

Although the probaUUty of alternation was cióse to the expected valué in a random string, participants tend to produce n-grams of length > 2 with probability distributions which are not equiprob-able (see Fig. 1). Strings containing more alternations (like 1010, 0101, 010, 101) and 3- and 4- runs have a higher frequency than expected by chance. This might be seen as an effort from participants to keep the probaUUty of alternation cióse to 0.5 by compensating the excess of alternations with blocks of repetitions of the same symbol.

ii. Comparing human randomness with other random sources

We asked whether K, the Kolmogorov complexity relative to LT2C2 defined in section IV.L, is able to detect and compress more patterns in strings generated by participants than in strings produced by other sources, which are considered random for many practical issues. In particular, we stud-ied strings originated by two sources: Pseudo-Random Number Generator (PRNG) and Atmo-spheric Noise (AN).

 

For the PRNG source, we chose the Mersenne Twister algorithm 47 (specifically, the second revisión from 2002 that is currently implemented in GNU Scientific Library). The atmospheric noise was taken from random.org site (property of Ran-domness and Integrity Services Limited) which also runs real-time statistic tests recommended by the US National Institute of Standards and Technol-ogy to ensure the random quality of the numbers produced over time.

The mean and median of K increases when comparing participant’s string with PRNG or AN strings. This difference was significant, as con-firmed by a t-test (p-value of 4.9 x lO-11 when comparing participant’s sample with PRNG one, a p-value of 1.2 x lO-15 when comparing participant’s with AN and a p-valué of 1.4 x 10-2 when comparing PRNG with AN sample).

Therefore, despite the simplicity of LT2C2, based merely on prints and repeats, it is rich enough to identify regularities of human sequences. The K function relative to LT2C2 is an effective and significant measure to distinguish strings produced by participants with profound understanding in the mathematics of randomness, from PRNG and AN strings. As expected, humans produce less complex (Le., less random) strings than those produced by PRNG or atmospheric noise sources.

iii. Mental fatigue

On cognitively demanding tasks, fatigue affects performance by deteriorating the capacity to or-ganize behavior 48-52. Specifically, Weiss claim that boredom may be a factor that increases non-randomness 48. Henee, as another test to the abil-ity of K relative to LT2C2 to identify idiosyncratic elements of human regularities, we asked whether the random quality of the participant’s string dete-riorated with time.

For each of the 138 strings x = X1 . . . x120 (x¿ G {0,1}) produced by the participants, we measured the K complexity of all the sub-strings of length 30.

Specifically, we calculated the average K(xí---xí+30) from the 138 strings for each i G 0,90 (see Fig. 2), using the same parameters as in section Vl.ii. (6 = r = 1), and compared to the same sliding average procedure for PRNG (Fig. 3) and AN sources (Fig. 4).

The solé source which showed a significant linear regression was human generated data (see Table 3) which, as expected, showed a negative correlation indicating that participants produced less complex or random strings over time (slope -0.007, p < 0.02).

The finding of a fatigue-related effect shows that the unpropped, i.e., resource-limited, human Tur-ing machine is not only limited in terms of the lan-guage it can parse, but also in terms of the amount of time it can dedicate to a particular task.

 

iv. Predictability

In section IV.ii., we introduced a prediction method with two parameters: £Fand£P. A predictor based on LT2C2 achieved levéis of predictability cióse to 56% which were highly significant (see Table 4). The predictor, as expected, performed at chance for the control PRNG and AN data. This fit was relatively insensitive to the valúes of £P and £F, contrary to our intuition that there may be a mem-ory scale which would correspond in this framework to a given length.

A very important aspect of this investigation, in line with the prior work of 23, is to inquire whether specific parameters are stable for a given individual. To this aim, we optimized, for each participant, the parameters using the first 80 symbols of the sequence and then tested these parameters in the second half of each segment (last 80 symbols of the sequence)

After this optimization procedure, mean predictability increased significantly to 58.14% (p < 0.002, see Table 5). As expected, the optimization based on partial data of PRNG and AN resulted in no improvement in the classifier, which remained at chance with no significant difference (p < 0.3, p<0.2, respectively).

Henee, while the specific parameters for compres-sion vary widely across each individual, they show stability in the time-scale of this experiment.

VII. Discussion

Here we analyzed strings produced by participants attempting to generate random strings. Partici-pants had a profound understanding of randomness and hence avoided typical misconceptions such as exaggerating the number of alternations. We reasoned that remaining regularities would express the algorithmic nature of human thoughts, revealed in the form of specific patterns.

Our effort here was to bridge the gap between Kolmogorov theory and psychology, developing a concrete language, LT2C2, satisfying the follow-ing requirements: 1) to be simple enough so that the complexity of any given sequence can be com-puted, 2) to be based on tangible operations of human reasoning ( printing, repeating, . . . ), 3) to be sufficiently powerful to generate all possible se-quences but not too powerful as to identify regu-larities which would be invisible to humans.

More specifically, our aim is to develop a class of languages with certain degrees of freedom which can then be fit to an individual (or an individual in a specific context and time). Here, we opted for a comparably easier strategy by only allowing the relative cost of each operation to vary. However, a natural extension of this framework is to generate classes of languages where structural and qualita-tive aspects of the language are free to vary. For instance, one can devise a program structure for repeating portions of (not necessarily neighboring) code, or considering the more general framework of for-programs where the repetitions are more general than in our setting: for i=1 to n do P(i), where P is a program that uses the successive values of i = 1, 2, . . ., n in each iteration. For instance, the following program

for i=1 to 6 do print ‘0’ repeat i times: print ‘1’ would describe the string

010110111011110111110111111.

The challenge from the computational theoretical point of view is to define an extension which induces a computable (even more, feasible, whenever possi-ble) Kolmogorov complexity. For instance, adding simple control structures like conditional jumps or

allowing the use of imperative program variables may turn the language into Turing-complete, with the theoretical consequences that we already men-tioned. The aim is to keep the language simple and yet include structures to compact some patterns which are compatible with the human language of thought.

We emphasize that our aim here was not to genérate an optimal predictor of human sequences. Clearly, restricting LT2C2 to a very rudimentary language is not the way to go to identify vast classes of patterns. Our goal, instead, was to use human sequences to calibrate a language which expresses and captures specific patterns of human thought in a tangible and concrete way.

Our model is based on ideas from Kolmogorov complexity and Solomonoff’s induction. It is im-portant to compare it to what we think is the clos-est and more similar approach in previous stud-ies: the work 23 of Griffiths and Tenenbaum’s. Griffiths and Tenenbaum devise a series of statisti-cal models that account for different kind of reg-ularities. Each model Z is fixed and assigns to every binary string x a probability Pz(x). This probabilistic approach is connected to Kolmogorov complexity theory via Levin’s famous Coding The-orem, which points out a remarkably numerical re-lation between the algorithmic probability Pv{x) (the probability that the universal prefix Turing machine U outputs x when the input is filled-up with the results of coin tosses) and the (prefix) Kolmogorov complexity Kv described in section Il.i. Formally, the theorem states that there is a con-stant c such that for any string x G {0,1}* such that

(the reader is referred to section 4.3.4 of 24 for more details). Griffiths & Tenenbaum’s bridge to Kolmogorov complexity is only established through this last theoretical result: replacing Pv by Pz in Eq. (4) should automatically give us some Kolmogorov complexity Kz with respect to some un-derlying Turing machine Z.

While there is henee a formal relation to Kolmogorov complexity, there is no explicit definition of the underlying machine, and henee no notion of program.

On the contrary, we propose a specific language of thought, formalized as the programming language LT2C2 or, alternatively, as a Turing machine N, which assigns formal semantics to each program. Semantics are given, precisely, through the behavior of N. The fundamental introduc-tion of program semantics and the clear distinc-tion between inputs (programs of N) and outputs (binary strings) allows us to give a straightfor-ward definition of Kolmogorov complexity relative to N, denoted KN, which because of the choice of LT2C2 becomes computable in polynomial time. Once we count with a complexity function, we ap-ply Solomonoff’s ideas of inductive inference to ob-tain a predictor which tries to guess the continu-ation of a given string under the assumption that the most probable one is the most compressible in terms of LT2C2-Kolmogorov complexity. As in 23, we also make use of the Coding Theorem (4), but in the opposite direction: given the complexity KN, we derive an algorithmic probability PN.

This work is mainly a theoretical development, to develop a framework to adapt Kolmogorov ideas in a constructive procedure (i.e., defining an explicit language) to identify regularities in human se-quences. The theory was validated experimentally, as three tests were satisfied: 1) human sequences were less complex than control PRNG sequences, 2) human sequences were non-stationary, showing decreasing values of complexity, 3) each individual showed traces of algorithmic stability since fitting of partial data was more effective to predict sub-sequent data than average fits. Our hope is that this theory may constitute, in the future, a useful framework to ground and describe the patterns of human thoughts.

 

Acknowledgements - The authors are thank-ful to Daniel Gorín and Guillermo Cecchi for useful discussions. S. Figueira is partially supported by grants PICT-2011-0365 and UBACyT 20020110100025.

1 M Kac, What is random?, Am. Sci. 71, 405 (1983).

2 H Reichenbach, The Theory of Probabil-ity, University of California Press, Berkeley (1949).

3 G S Tune, Response preferences: A review of some relevant literature, Psychol. Bull. 61, 286 (1964).

4 A D Baddeley, The capacity for generating in-formation by randomization, Q. J. Exp. Psy-chol. 18, 119 (1966).

5 A Tversky, D Kahneman, Belief in the law of small numbers, Psychol. Bull. 76, 105 (1971).

6 W A Wagenaar, Randomness and randomiz-ers: Maybe the problem is not so big, J. Behav. Decis. Making 4, 220 (1991).

7 R Falk, Perception of randomness, Unpub-lished doctoral dissertation, Hebrew Univer-sity of Jerusalem (1975).

8 R Falk, The perception of randomness, In: Proceedings of the fifth international confer-ence for the psychology of mathematics ed-ucation, Vol. 1, Pag. 222, Grenoble, France (1981).

9 D Kahneman, A Tversky, Subjective probabil-ity: A judgment of representativeness, Cogni-tive Psychol. 3, 430 (1972).

10 A Tversky, D Kahneman, Subjective probabil-ity: A judgment of representativeness, Cogni-tive Psychol. 3, 430 (1972).

11 T Gilovich, R Vallone, A Tversky, The hot hand in basketball: On the misperception of random sequences, Cognitive Psychol. 17, 295 (1985).

12 W A Wgenaar, G B Kerenm, Cance and luck are not the same, J. Behav. Decis. Making 1, 65 (1988).

13 D Budescu, A Rapoport, Subjective random-ization in one-and two-person games, J. Be-hav. Decis. Making 7, 261 (1994).

14 A Rapoport, D V Budescu, Generation of ran-dom series in two-person strictly competitive games, J. Exp. Psychol. Gen. 121, 352 (1992).

15 P D Larkey, R A Smith, J B Kadane, It’s okay to believe in the hot hand, Chance: New Directions for Statistics and Computing 2, 22-30 (1989).

16 A Tversky, T Gilovich, The ”hot hand”: Sta-tistical reality or cognitive illusion?, Chance: New Directions for Statistics and Computing 2, 31 (1989).

17 A N Kolmogorov, Three approaches to the quantitative definition of information, Probl. Inf. Transm. 1, 1 (1965).

18 G J Chaitin, A theory of program size formally identical to information theory, J. AMC 22, 329 (1975).

19 L A Levin, A K Zvonkin, The complexity of finite objects and the development of the con-cepts of information and randomness by means of the theory of algorithms, Russ. Math. Surv. 25, 83 (1970).

20 D G Champernowne, The construction of dec-imals in the scale of ten, J. London Math. Soc. 8, 254 (1933).

21 R J Solomonoff, A formal theory of inductive inference: Part I, Inform. Control 7, 1 (1964); ibid. Part II 7, 224 (1964).

22 G Chaitin, On the length of programs for com-puting finite binary sequences: statistical con-siderations, J. ACM 13, 547 (1969).

23 T L Griffiths, J B Tenenbaum, Probability, al-gorithmic complexity, and subjective random-ness, In: Proceedings of the Twenty-Fifth An-nual Conference of the Cognitive Science So-ciety, Eds. R Alterman, D Hirsh, Pag. 480, Cognitive Science Society, Boston (MA, USA), (2003).

24 M Li, P M Vitányi, An introduction to Kolmogorov complexity and its applications, Springer, Berlin, 3rd edition (2008).

25 J Feldman, Minimization of boolean copmlex-ity in human concept learning, Nature London 407, 630 (2000).

26 N Chater, The search for simplicity: A fundamental cognitive principle?, Q. J. Exp. Psy-chol. 52A, 273 (1999).

27 R Falk, C Konold, Making sense of random-ness: Implicit encoding as a bias for judgment, Psychol. Rev. 104, 301 (1997).

28 C P Schnorr, Zuf¨alligkeit und Wahrschein-lichkeit, Lecture Notes in Mathematics vol. 218. Springer-Verlag, Berlin, New York (1971).

29 C P Schnorr, A unified approach to the defini-tion of a random sequence, Math. Syst. Theory 5, 246 (1971).

30 D Dellarosa, A history of thinking, In: The psychology of human thought, Eds. R J Stern-berg, E E Smith, Cambridge University Press, Cambridge (USA) (1988).

31 S Carey, The origin of concepts, Oxford Uni-versity Press, Oxford (USA) (2009).

32 G Boole, An investigation of the laws of thought: on which are founded the mathemat-ical theories of logic and probabilities, Vol. 2, Walton and Maberly, London (1854).

33 A Zylberberg, S Dehaene, G Mindlin, M Sig-man, Neurophysiological bases of exponential sensory decay and top-down memory retrieval: a model, Front. Comput. Neurosci. 3, 4 (2009).

34 A Zylberberg, S Dehaene, P Roelfsema, M Sig-man, The human Turing machine: a neural framework for mental programs, Trends. Cogn. Sci. 15, 293 (2011).

35 M Graziano, P Polosecki, D Shalom, M Sig-man, Parsing a perceptual decision into a se-quence of moments of thought, Front. Integ. Neurosci. 5, 45 (2011).

36 A Zylberberg, P Barttfeld, M Sigman, The construction of confidence in a perceptual de-cision, Front. Integ. Neurosci. 6, 79 (2012).

37 D Shalom, B Dagnino, M Sigman, Looking at breakout: Urgency and predictability direct eye events, Vision Res. 51, 1262 (2011).

38 S Dehaene, M Sigman, From a single decision to a multi-step algorithm, Curr. Opin. Neuro-biol. 22, 937 (2012).

39 J Kamienkowski, H Pashler, S Dehaene, M Sigman, Effects of practice on task archi-tecture: Combined evidence from interference experiments and random-walk models of deci-sion making, Cognition 119, 81 (2011).

40 A Zylberberg, D Slezak, P Roelfsema, S De-haene, M Sigman, The brain’s router: a cortical network model of serial processing in the primate brain, PLoS Comput. Biol. 6, e1000765 (2010).

41 L Gallos, H Makse, M Sigman, A small world of weak ties provides optimal global integration of self-similar modules in functional brain net-works, P. Natl. Acad. Sci. USA 109, 2825 (2012).

42 L Gallos, M Sigman, H Makse, The co-nundrum of functional brain networks: small-world efficiency or fractal modularity, Front. Physiol. 3 123 (2012).

43 M Costa, F Bonomo, M Sigman, Scale-invariant transition probabilities in free word association trajectories, Front. Integ. Neu-rosci. 3 19 (2009).

44 N Mota, N Vasconcelos, N Lemos, A Pieretti, O Kinouchi, G Cecchi, M Copelli, S Ribeiro, Speech graphs provide a quantitative measure of thought disorder in psychosis, PloS one 7, e34928 (2012).

45 M Sigman, G Cecchi, Global organization of the wordnet lexicon, P. Natl. Acad. Sci. USA 99, 1742 (2002).

46 R Cilibrasi, P M Vitányi, Clustering by com-pression, IEEE T. Inform. Theory 51, 1523 (2005).

47 M Matsumoto, T Nishimura, Mersenne twister: a 623-dimensionally equidistributed uniform pseudo-random number generator, ACM Trans. Model. Comput. Simul. 8, 3 (1998).

48 R L Weiss, On producing random responses, Psychon. Rep. 14, 931 (1964).

49 F Bartlett, Fatigue following highly skilled work, Nature (London) 147, 717 (1941).

50 D E Broadbent, Is a fatigue test now possible?, Ergonomics 22, 1277 (1979).

51 W Floyd, A Welford, Symposium on fatigue and symposium on human factors in equip-ment design, Eds. W F Floyd, A T Welford, Arno Press, New York (1953).

52 R Hockey, Stress and fatigue in human performance, Wiley, Chichester (1983).

Creative Commons License Todo el contenido de esta revista, excepto dónde está identificado, está bajo una Licencia Creative Commons