Algorithmic Specified Complexity DRAFT. Winston Ewert, William A. Dembski, and Robert J. Marks II. October 16, 2012

Size: px
Start display at page:

Download "Algorithmic Specified Complexity DRAFT. Winston Ewert, William A. Dembski, and Robert J. Marks II. October 16, 2012"

Transcription

1 Algorithmic Specified Complexity Winston Ewert, William A. Dembski, and Robert J. Marks II October 16, 2012 Abstract As engineers we would like to think that we produce something different from that of a chaotic system. The Eiffel tower is fundamentally different from the same components lying in a heap on the ground. Mt. Rushmore is fundamentally different from a random mountainside. But we lack a good method for quantifying this idea. This has led some to reject the idea that we can detect engineered or designed systems. Various methods have been proposed each of which has various faults. Some have trouble distinguishing noise from data, some are subjective, etc. We propose to use conditional Kolmogorov complexity to measure the degree of specification of an object. The Kolmogorov complexity of an object, is the length of the shortest computer program required to describe that object. Conditional Kolmogorov complexity is Kolmogorov complexity, with access to a context. The program can extract information from the context in a variety of ways allowing more compression. The more compressible an object is the more we may deem the object specified. Random noise is incompressible, and so compression indicates that the object is not simply random noise. We hope this model launches further dialog on use of conditional Kolmogorov complexity in the measurement of specified complexity. 1 Introduction Intuitively, humans identify objects such as the carved faces at Mount Rushmore as qualitatively different from that of a random mountainside. However, quantifying this concept is an objective manner has proved difficult. Both mountainsides are made up of the same material components. They are both subject to the same physical forces, and will react the same to almost all physical tests. Yet, there does appear to be something quite different about Mount Rushmore. There is a special something about carved faces that separates it from the rock it is carved in. This special something is information. Information is what distinguishes an empty hard disk from a full one. Information is the difference between random scribbling and carefully printed prose. Information is the difference between car parts strewn over a lawn, and a working truck. While humans operate using an intuitive concept of information, attempts to develop a theory of information has thus far fallen short of the intuitive concept. Claude Shannon developed what its today known as Shannon information theory [15]. Shannon s concern was studying the problem of communication, that of sending information from one point to another. However, Shannon explicitly avoided the question of the meaningfulness of the information being transmitted, thus not quite capturing the concept of information as we are defining it. In fact, under Shannon s model a random signal has the highest amount of information, the precise opposite of the intuitive concept. Another model of information is that of algorithmic information theory [4, 16, 12]. Techniques such as Kolmogorov complexity measure the complexity of an object as the minimum length computer program required to recreate the object; Chaitin refers to such minimum length programs as elegant [5]. As with Shannon information, random noise is the most complex because it requires a long computer program to describe. In contrast, simple patterns are not complex because a short computer program can describe the pattern. But neither simple patterns or random noise are what we think of as information. As with Shannon information, there is a disconnect between Kolmogorov complexity and conceptual information. 1

2 Other models are based on algorithmic information theory, but also take in account the computational resources required for the programs being run. Levin complexity adds the log of the execution time to the complexity of the problem [13]. Logical depth, on the other hand, is concerned with the execution time of the shortest program [3]. There is a class of objects which are easy to describe but expensive to actually produce. It is argued [? ] that objects in this class must have been produced over a long history. Such objects are interesting, but do not seem to capture all of what we consider to be the intuitive concept of information. English text or Mount Rushmore correspond to what we think of as information, but its not clear that they can be most efficiently described as long running programs. One approach to information is the specified complexity as expressed by Dembski [7]. Dembski s concern is that of detecting design, the separation of that which can be explained by chance or necessity from that which is the product of intelligence. In order to infer design, and object must be both complex and specified. Complexity refers essentially to improbability. The probability of any given object depends on the chance hypothesis proposed to explain it. Improbability is a necessary but not sufficient condition for rejecting a chance hypothesis. Events which have a high probability under a given chance hypothesis do not give us reason to reject that hypothesis. Specification is defined as conforming to an independently given pattern. The requirement for the pattern to be independent of the object being investigated is fundamental. Given absolute freedom of pattern selection, any object can be made specified by selecting that object as the pattern. It is not impressive to hit a bullseye if the bullseye is painted on after the arrow has hit the wall. It is impressive to hit the bullseye if the bullseye was painted before the arrow was fired. Investigators are often not in the position of being able to choose the target prior to investigating the object. Consider the example of life. Life is a self-replicating process, and it would seem that an appropriate specification would be self-replication. Self-replication is what makes life such a fascinating area of investigation as compared to rocks. We know about self-replication because of our knowledge of life, not as an independent fact. Therefore it does not qualify as an independent specification. If we did not already have examples of self-replicating entities, we would not have picked as the specification. The same is true of almost any specification in biology. It is tempting to consider flight a specification, but we would only be defining the pattern of flight because we have seen flying animals. As with life in general, specific features in biology cannot be specified independently of the objects themselves. The concept of specification has been criticized for being imprecisely defined and unquantifiable. It is also charged that the maintaining the independence of the patterns is difficult. But specification has been defined in a mathematically rigorous manner in several different ways [7, 8, 9]. Kolmogorov complexity, or a similar concept, is a persistent method used in this definitions. Our goal is to present and defend a simple measure of specification that clearly alleviates these concerns. Towards this end, we propose to use conditional Kolmogorov complexity to quantify the degree of specification in an object. Combining this with the complexity, we can quantify the specified complexity as algorithmic specified complexity. As noted, Kolmogorov complexity has been suggested as a method for measuring specification. The novelty in method presented in this paper is the use of conditional Kolmogorov complexity. However, this paper also elucidates a number of examples of algorithmic compressibility demonstrating wider applicability then is often realized. 2 Method 2.1 Kolmogorov Kolmogorov complexity is a method of measuring information. It is defined as the minimum length computer program, in bits, required to produce a binary string. where K(X) = min p (1) U(p,)=X p P 2

3 K(X) is the Kolmogorov complexity of X P is the set of all possible computer programs U(p, ) is the output of program p run without input The definition is given for producing binary strings. Kolmogorov complexity measures the degree to which a given bitstring follows a pattern. The more a bitstring follows a pattern, the shorter the program required to reproduce it. In contrast, if a bitstring exhibits no patterns, it is simply random, and a much longer program will be required to produce it. Consider the example of a random binary string, It can be produced by the following Python program: print In contrast, we have the string, , which can be produced by print 0 * 24 Both strings are of the same length, but the string following a pattern requires a shorter program to produce. Thus we have a technique for measuring the degree to which a binary string follows a pattern. Specification is defined as following an independently given pattern. Kolmogorov complexity gives us the ability to precisely define and quantify the degree to which a binary string follows a pattern. Therefore, it seems plausible that we can measure specification using Kolmogorov complexity. The more compressible a bitstring, the more specified it is. However, Kolmogorov complexity seems unable to capture the entirety of what is intended by specification. Natural language text is not reducible to a simple pattern; however, it is an example of what we d consider specification. The design of an electronic circuit should also be specified, but it is not reducible to a simple pattern. In fact, the cases of specification that Kolmogorov complexity seems able to capture are limited to objects which exhibit some very simple pattern. But these are not the objects of most interest in terms of specification. We use an extension of Kolmogorov complexity known as conditional Kolmogorov complexity [11]. The program now has access to additional data as its input. K(X Y ) = min U(p,Y )=X p P p (2) where U(p, Y ) is the output of running program p with input Y. The input provides additional data to the program. As a result, the program is no longer restricted to exploiting pattern in the desired output, but can take advantage of the information provided by the input. Henceforth, we will refer to this input as the context. The use of context allows the measure to capture a broader range of specifications. It is possible to describe many bitstrings by combining a short program along with the contextual information. A useful range of specifications can be captured using this technique. 2.2 Algorithmic Specified Complexity To combine the measurement of specification and complexity, we use the following formula for algorithmic specified complexity (ASC). A(X, C, p) = log p(x) K(X C) (3) where X is the bitstring being investigated C is the context as a bitstring 3

4 p is the probability distribution which we suppose X to have been selected from p(x) is the probability of X occurring according to the chance hypothesis under consideration Since high compressibility corresponds to specification, we subtract the compressed length of the string. Thus high improbability counts for specified complexity, but incompressible strings count against it. For this number to become large requires X to be both complex, (i.e. improbable), and specified, (i.e. compressible). Failing on either of these counts will produce a low or negative value. Since Kolmogorov complexity can, at best, be upper bounded, the ASC can, at best, be lower bounded. At best this measure can reject a given probability distribution. It makes no attempt to rule out chance based hypothesis in general. However, it can conclude that a given probability distribution does a poor job in explaining a particular item. The value of ASC gives a measure of the confidence we can have in rejecting a chance hypothesis. 2.3 Functionality Perhaps the most interesting form of specification is that of functionality. It is clear that machines, biological structures, and buildings all have functionality. But quantifying that in an objective manner has proven difficult. However, ASC gives us the ability to do this. Any machine can be described, in part, by tests that it will pass. You can test the functionality of a car by seeing whether it accelerates when the gas or brake pedals are pushed. You can test the functionality of a cell by seeing whether it self-replicates. A test, or a number of tests, can be defined to identify the functionality of an object The existence of a test gives us the ability to compress the object. Consider the following pseuocode program: counter = 0 for each possible building design if building won t fall over counter += 1 if counter == X return building design where X is some number. This program will output the design for a specific building. Different values of X will produce different buildings. But any building that will not fall over can be expressed by this program. It may take a considerable amount of space to encode this number. However, if few designs are stable, the number will take much less space than required to actually specify the building plans. Thus the stability of the building plan enables compression, which in turn indicates specification. Kolmogorov complexity is not limited to exploiting what humans perceive as simple patterns. It can also capture other aspect such as functionality. Functionality can be described as passing a test. As a result functional objects are compressible. 3 Examples 3.1 Natural Language As the first example, consider the sentence: the quick brown fox jumps over the lazy dog. We can encode this sentence as UTF-32, a system for encoding that allows the encoding of symbols from almost any alphabet. Since each character takes 32 bits, the message will be encoded as a total of 1376 bits. In this case, we will take the context to be the English alphabet along with a space. This is a minimal level of information about the English language. 4

5 To specify one of the 27 characters, will require log 2 27 bits. To specify the 43 character in the sentence will thus take 43 log 2 27 bits. We also need to record the number of characters at 2 log bits. 1 Altogether, the bits required to specify the message requires 43 log log bits. However, in order to actually give a bound for Kolmogorov complexity, we must also include the length of the computer program which interprets the bits. Here is an example computer program in Python which could interpret the message print.join(alphabet[index] for index in encoded_message) This assumes that the alphabet and encoded message are readily available and in form amenable to processing within the language. It may be that the input has to be preprocessed, which would make the program longer. Additionally, the length of the program will vary heavily depending on which programming language is used. However, the distances between different computers and languages only differs by a constant [6]. As a result, it is common practice in algorithmic information theory, to discount any actual program length and merely include that length as a constant, c. Consequently, we can express the conditional Kolmogorov complexity as K(X C) bits + c. (4) The expression is less than rather than equal, because it is possible that an even more efficient way of expressing the sentence exists. But we know that at least this efficiency is possible. The encoded version of the sentence requires 32 bits for each character, giving a total of 1376 bits. We adopt a simplistic probability model, supposing that each bit is generated by the equivalent of a coin flip. This means that the complexity, log P (X) is 1376 bits. Using equation 3, A(X, C, p) = log(p) K(X C) 1376 bits bits c = bits c. (5) This means that we have 1166 bits of algorithmic specified complexity by equation 3. Those 1166 bits are a measure of the confidence in rejecting the hypothesis that the sentence was generated by random coin flips. The large number of bits gives a good indication that is highly unlikely that this sentence was generated by randomly choosing bits. However, we can also analyze the hypothesis that the sentence was generated by choosing random English letters. In this case we can calculate the probability of this sentence as The complexity is then ( ) 43 1 P (X) =. (6) 27 ( ) 43 1 log P (X) = log = 43 log bits. (7) 27 In which case the algorithmic specified complexity becomes A(X, C, p) = log p(x) K(X C) bits bits c = bits c. (8) The negative bound indicates that we have no reason to suppose that this sentence could not have been generated by a random choice of English letters. The bound is negative as a result of two factors. In the specification, bits were required to encode the length. On the other hand, the probability model assumes a length. Hence the negative bits indicate information which the probability model had, but was not provided in the context. Since the only provided context is that of English letters, this is not a surprising result. We did not identify any sort of pattern beyond that explained by the probability model. We can also expand the context. Instead of providing the English alphabet as our context, we provide the word list of the Oxford English Dictionary [14]. In the second edition of that dictionary there were 1 A more compact representation for numbers is available. See the log method in Cover and Thomas [6]. 5

6 615,100 word forms defined or illustrated. For the purpose of the alphabet context, we encode each letter as a number corresponding to that character. In this case, we can choose a number corresponding to words in the dictionary. We can thus encode this message: K(X C) 9 log 2 615, log c c. (9) Access to the context of the English dictionary allows much better compression than simply the English alphabet as comparing equations 4 and 9 shows. Using equation 3, we determine A(X, C, p) = log p(x) K(X C) bits bits c = bits c. (10) This gives us confidence to say this sentence was not generated by randomly choosing letters from the English alphabet. It would be possible to adopt a probability model that selected random words from the English language. Such a probability model would explain all of the specification in the sentence. It would also be possible to include more information about the English language such that the specification would increase. This technique depends on the fact that the numbers of words in the English language is much smaller then the number of possible combinations of letters. If the dictionary contained every possible combination of letters up to some finite length, it would not allow compression, and thus we would not be able to indicate specification. A language where all possible combinations of letters were valid words could still show specification, but another technique would have to be used to allow compression. But one could also use a much smaller dictionary. A dictionary of 10 words would be sufficient to include all the words in this sentence. The ASC formula would give a much smaller compressed bound: K(X C) 9 log log bits. (11) This is a reduction of over 100 bits from equation 9. This is because it takes about 16 bits less to encode each word when the dictionary is this small. This is because the sentence is much more closely related to the context. It requires much less additional information to use the context. But it is possible to include words not included in the dictionary. The program would have to fall back on spelling the word one letter at a time. Only bounds of the ASC can be computed. It is always possible a better compression exists; i.e. the object could be more specified than we realize. 3.2 Random Noise While natural language is an example of something that should be specified, random noise is an example of something which should not. We will calculate the ASC of a random bitstring. This bitstring will contain 1000 bits, where each bit is assigned with equal probability 1 or 0. Since randomness is incompressible, calculating the Kolmogorov complexity is easy. The only way of reproducing a random bitstring is to describe the whole bitstring. K(X) 2 log c 1020 bits + c (12) The probability of each bitstring is and thus the complexity will be 1000 bits. Calculating the ASC: A(X, C, p) = log p(x) K(X C) 1000 bits 1020 bits c = 20 bits c. (13) As expected, the ASC is negative, there is no evidence of pattern in the string which are not explained by the probability model. However, we can also consider the case of a biased distribution. That is, 1 and 0 are not equally likely. Instead, a given bit will be one two thirds of the time, while zero only one third of the time. The entropy of each bit can be expressed as H(X i ) = 1 3 log log bits. (14) 3 6

7 for any i The entropy of a bit is the number of bits required in an optimal encoding to encode each bit. This means we can describe the whole sequence as K(X) 2 log H(X i ) + c bits + c. (15) If we adopt the uniform probability model, the complexity is still 1000 bits and A(X, C, p) = log p(x) K(X C) 1000 bits bits c = bits c. (16) This random sequence has a high bound of algorithmic specified complexity. It is important to remember that the ASC bound only serves to measure the plausibility of the random model. It does not exclude the existence of another more accurate model that explains the data. In this case, if we use the actual probability model used to generate the message log 2 (p) = H(X i ) bits. (17) and the resulting ASC: A(X, C, p) = log p(x) K(X C) bits bits c = 20 bits c. (18) The bound of ASC gives us reason to reject a uniform noise explanation for this data, but not the biased coin distribution. Dembski [7] has considered the example of ballot rigging where a political party is almost always given the top billing on the ballot listing candidates. Since the selection is supposed to be chosen on the basis of a fair coin toss, this is suspicious. ASC can quantify this situation. We can describe the outcome by giving the numbers of heads and tails, follow by the same representation as for the biased coin distribution. ( ) Xt + X h K(X) 2 log X h + 2 log X t + log + c (19) where X h is the number of heads, X t is the number of tails We assume a probability model of a fair coin X h log 2 (p) = X h + X t bits. (20) This gives us: ( ) ( ( )) Xt + X h A(X, C, p) = X h +X t 2 log X h 2 log X t log c = X h +X t log XhX 2 t 2 Xt + X h c. (21) X h Figure 1 shows the result of plotting this equation for varying numbers of head and tails given 20 coin tosses. As expected, for either high numbers of tails or high number of heads, the bound of ASC is high. However, for an instance which looks like a random sequence, the ASC is minimized. 3.3 Playing Cards Another interesting case is that of playing cards for poker. In playing cards if the distribution is not uniform; somebody is likely cheating. For the purpose of investigating card hands, we can simply assume a uniform random distribution over all five-card poker hands. We will consider the hands for the game of poker. A poker hand is made up of 5 cards. Some categories of hands are rarer then others. Table 1 shows the frequency of the different hands. Given a uniform distribution, every poker hand has the same probability, and thus the same complexity. There are 2,598,960 possible poker hands. This gives us 1 log 2 p(x) = log 2 ( ) 21.3 bits. (22) 2, 598, X h

8 ASC (bits) number of heads Figure 1: ASC for varyingly biased coin sequences and 20 coin tosses Name Frequency Royal Flush 4 Straight Flush 36 Four of Kind 624 Full House 3744 Flush 5108 Straight Three of a Kind Two Pair One Pair None Table 1: Poker hand frequency 8

9 Name Frequency Complexity Compressed Length ASC Royal Flush Straight Flush Four of a Kind Full House Flush Straight Three of Kind Two pair One pair None Table 2: The ASC of the various poker card hands While the probability of every poker hand is the same, the Kolmogorov complexity is not. To describe a royal flush requires specifying that is a royal flush, and which suit it is in. However, describing a pair requires specifying which the paired value as well as both suit in addition to the three cards not involved in the pair. In general, describing hand requires specifying the type of hand, which of all the hands with that type. This gives us K(H i C) log log 2 H + c. (23) where 10 is the number of types of hands. H is the set of all hands of a particular type, and H i is a particular hand in that set. There are 1,098,240 possible pairs. Putting this in Equation 23 gives: K(H i C) log log 2 H + c bits + c. (24) On the other hand, describing a pair without using the context gives K(H i C) log 2 2, 598, c 21.3 bits + c. (25) Single pairs are so common that the space required to record that it was a pair is more then the space required to record the duplicate card straightforwardly. Accordingly, we must take the minimum of the two methods K(H i C) min(log log 2 H, log 2 2, 598, 960) + c. (26) Table 2 shows the ASC for the various poker hands. Rare hands have large ASC, but command hands have low ASC. This parallels what we would expect, because a rare hand might cause us to expect cheating, but a common hand will not. In other card games, a card is turned over after hands have been dealt to determine trump. The suit of the card is taken to trump for that round of the game. If the same suit is repeatedly chosen as trump, someone may ask what the odds are. This question can be difficult to answer because every possible sequence of trump suits is equally likely. Yet, it is deemed unusual that the same suit is trump repeatedly. Algorithmic specified complexity allows us to capture this. We represent the suits as a bit sequence, using two bits for each suit. K(X) = log log 2 H + c = 2 + log 2 H + c (27) where 4 is the number of suits, and H is the number of hands played. The complexity of the sequence is log P (X) = 4 X 2 = 2H. (28) The ASC is then ASC(X, p) = 2H 2 log 2 H c. (29) 9

10 ASC (bits) hands Figure 2: A plot of ASC for getting the same suit repeatedly 10

11 Note that this equation becomes c when H = 1. A pattern repeating once is no pattern at all, and doesn t provide specification. Figure 2 shows the ASC for increasing numbers of hands. The more times the same suit is chosen as trump, the larger the number of bits of ASC. Given the same trump for many rounds becomes less and less probable. 3.4 Folding Proteins In biology, an important prerequisite to a protein being functional is that it folds. The number of all possible proteins folding has been estimated. the overall prevalence of sequences performing a specific function by any domain-sized fold may be as low as 1 in [2] We can create a program which outputs a particular protein, given the laws of physics. for all proteins of length L run protein in a physics simulator if protein folds add to list of folding proteins output the Xth protein from the list This program given different choices of L and N will output any folding protein that we choose. This means that we can describe the protein by providing those two numbers. K(X C) = 2 log 2 L + log 2 F L + c (30) where C is the context, in this case the law of physics. F L is the number of folding proteins of length L. Taking Axe s estimate [2], an assuming simplistically, that it applies for all lengths of proteins: F L = L (31) so log F L = 77 log 10 + L log 4 (32) K(X C) = 2 log 2 L + log 2 F L + c = 2 log 2 L + 77 log 10 + L log 4. (33) For our probability model, we will suppose that each base along the DNA chain is uniformly chosen. It should be emphasized that according to the Darwinian model of evolution, the bases are not uniformly chosen. This supposition only serves to test a simplistic chance model of protein origin. We can calculate the probability as log 2 Pr(X) = log 2 4 L = L log 2 4. (34) Caution should be used in apply this formula. It assumes that proportion of functional proteins is applicable for all lengths, and implies that a fractional number of proteins fold. Finally calculating the ASC, ASC(X, p) = L log 4 2 log 2 L + 77 log 2 10 L log 2 4 c = 2 log 2 L + 77 log 2 10 c. (35) The final bound for ASC depends little on the length of the protein sequence which only comes to play in the logarithmic term. The significant term is the 77 log bits. This means that we have good reason to believe that folding sequences were not generated randomly from a uniform distribution. 11

12 3.5 Functional Sequence Complexity Kirk Durston et. al. have defined the idea of functional sequence complexity [10]. Functional sequence complexity is related to a special case of algorithmic specified complexity. A protein is made from a sequence of amino acids. Some sequences have functionality and some do not. The case considered in section 3.4 above of folding is one particular case. But perhaps more interesting is considering the case of various proteins which perform useful biological functions. Let Ω be the set of all proteins. Let F be the set of all proteins which pass a functionality test. Let f(x) be a probability distribution over F. Both F and f(x) can be produced by a simple algorithm using a functionality test on each element of Ω. Consequently, F and f(x) can be described using a constant program length. Consider the average for ASC over all elements in F. f(x)a(x, C, p) = f(x)( log p(x) K(x C)) = f(x) log p(x) f(x)k(x C)) (36) xinf We can describe any element x given the probability distribution and log f(x) bits. Given that we can calculate f(x) and F with a constant program gives: K(x C) log f(x) + c. (37) Place this into equation 36. f(x)a(x, C, p) f(x) log p(x) f(x) log f(x) c (38) The middle term is recognized as the Shannon entropy. f(x)a(x, C, p) f(x) log p(x) H(f) c f(x) (39) If we assume that the distribution p is uniform, p(x) = 1 Ω, f(x)a(x, C, p) log 2 Ω f(x) H(f) c f(x). (40) The two summations over F, are summations over a probability distribution and therefore 1. f(x)a(x, C, p) log 2 Ω H(f) c (41) Equation 5 in Durston s work, adjusting for notation is log Ω H(f). (42) This equation derives from making the same uniformity assumption that we have made here. Thus, for the uniform probability distribution case, (f(x)a(x, C, p)) + c log Ω H(f). (43) This establishes the relationship between ASC and FSC. The difference is that the ASC is a lower bound, and includes a constant. This is the same constant as elsewhere: the length of the program required to describe the specification. 12

13 4 Objections 4.1 Natural Law We have argued that compressibility in the presence of context is a necessary condition for information. This is in contrast to other who have argued that lack of compressibility is a necessary condition for information [1]. But compressible objects lack complexity. Because a compressible object is describable as some simple pattern, it is amenable to being produced by a simple process. Many objects in the real world follow simple patterns. Water tends to collect at lower elevations. Beaches follow a slopping pattern. Sparks fly upwards. But these patterns are the result of the operation of simple law-like processes. Even if the explanations for these patterns were unknown, we would suppose due to the simplicity of the pattern shown that some simple explanation existed. The premise behind this use of compressibility is that it identifies what human would see as simple patterns. Abel writes: A sequence is compressible because it contains redundant order and patterns. [1] The problem is that algorithms are very versatile and allow the description of many patterns beyond that which humans would see as patterns. As has been shown by the various examples in this paper, many objects which do not exhibit what humans typically identify as redundant order and patterns are in fact compressible. Significantly, we have argued that functionality actually allows compressibility. Contrary to what Abel states, functional sequences are compressible by virtue of the functionality they exhibit. All of the sequences that Abel holds to be mostly incompressible are actually compressible. But are compressible objects amenable to explanation by simple processes? Do all compressible objects lack complexity? If this were true, it would be problematic for algorithmic specified complexity because all specified objects would also be not complex, and no object would ever be both specified and complex. But many compressible objects do not appear to amenable to explanation by a simple process. As discussed, English text is compressible given a knowledge of the English language. This does not somehow make it probable that English text will appear on a beach carved out by waves. Ninety degree angles are very compressible; yet, they are not typically found in nature. The existence of an explanation from the laws of nature does not appear to follow from compressibility. Kolmogorov complexity deliberately ignores how long a program takes to run. It is only concerned with the length of the program s description. A program may be short but take an astronomical amount of time to run. Many of the specifications considered in this paper fall into that category. These objects are compressible, but that compression does not give an practical way to reproduce the object. But if there is no practical way to reproduce the object, we have no reason to suggest law-like processes as a plausible explanation. 4.2 Context is Subjective The ASC of any object will depend on the context chosen. Any object can be made to have high ASC by using a specifically chosen context. But this appears to be the way that information works. If the authors, who do not understand Arabic, look at Arabic text, it appears to be no better then scribbling. The problem is not that Arabic lacks information content, but that we are unable to identify it without the necessary context. As a result, this subjectivity appears to capture something about the way information works in the human experience. As with specification, it is important the context be chosen independent of object under investigation. While a specification will rarely be independent of the object under investigation, we believe it is much easier to maintain this independence in the case of a context. 4.3 Incalculability It is not possible to calculate the Kolmogorov complexity of an object. However, it is possible to upper bound the Kolmogorov complexity and thus lower-bound the algorithmic specified complexity. This means 13

14 that we can say that something is at least this specified, although we cannot rule out the possibility that it is even more specified. This means that we cannot mechanically detect that something has specification, although we can objectively identify it when we see it. 5 Conclusions Dembski argued that information can be detected by looking for specified complexity. We propose that all or most forms of specification can be represented as algorithms, using Kolmogorov complexity. The shorter the algorithm, the more specified the object is. In order to measure a broader range of specification, we include the context and thus make use of conditional Kolmogorov complexity. We have defined the concept of Algorithmic Specified Complexity which takes into account the probabilistic complexity as well as the Kolmogorov complexity. We have presented a number of examples showing how this can represent the specification in a variety of cases. We hope that this paper introduces discussion on the use of conditional Kolmogorov complexity as a method for measuring specification as well as the use of Algorithmic Specified Complexity. Acknowledgements The approach of using compressibility as a measurement of specification was suggested to the authors by Eric Holloway. We have attempted to extend the approach to apply to many more types of specifications. We are grateful for his initial suggestion and answer to our initial objections to the idea. References [1] David L Abel and Jack T Trevors. Three subsets of sequence complexity and their relevance to biopolymeric information. Theoretical biology & medical modelling, 2:29, January [2] Douglas D Axe. Estimating the prevalence of protein sequences adopting functional enzyme folds. Journal of molecular biology, 341(5): , August [3] CH Bennett. Logical depth and physical complexity. The Universal Turing Machine A Half-Century Survey, pages , [4] Gregory J. Chaitin. On the length of programs for computing finite binary sequences. Journal of the ACM (JACM), 13, [5] Gregory J. Chaitin. Conversations with a mathematician : math, art, science, and the limits of reason : a collection of his most wide-ranging and non-technical lectures and interviews. Springer, New York, New York, USA, [6] Thomas M Cover and Joy A Thomas. Elements Of Information Theory. Wiley-Interscience, Hoboken, NJ, second edi edition, [7] William A. Dembski. The Design Inference: Eliminating Chance through Small Probabilities, volume 112. Cambridge University Press, [8] William A. Dembski. No free lunch : why specified complexity cannot be purchased without intelligence. Rowman & Littlefield, Lanham MD, [9] William A. Dembski. Specification: the pattern that signifies intelligence. Philosophia Christi, 7(2): ,

15 [10] Kirk K Durston, David K Y Chiu, David L Abel, and Jack T Trevors. Measuring the functional sequence complexity of proteins. Theoretical biology & medical modelling, 4:47, January [11] AN Kolmogorov. Logical basis for information theory and probability theory. Information Theory, IEEE Transactions on, 14(5): , September [12] AN Kolmogorov. Three approaches to the quantitative definition of information*. International Journal of Computer Mathematics, [13] L. A. Levin. Various Measures of Complexity for Finite Objects. (Axiomatic Description). Soviet Math, 17(522), [14] Oxford English Dictionary Online. Oxford English Dictionary Online, [15] Claude E Shannon, Warren Weaver, and Norbert Wiener. The Mathematical Theory of Communication. Physics Today, 3(9):31, [16] RJ Solomonoff. A preliminary report on a general theory of inductive inference

Algorithmic Probability

Algorithmic Probability Algorithmic Probability From Scholarpedia From Scholarpedia, the free peer-reviewed encyclopedia p.19046 Curator: Marcus Hutter, Australian National University Curator: Shane Legg, Dalle Molle Institute

More information

CISC 876: Kolmogorov Complexity

CISC 876: Kolmogorov Complexity March 27, 2007 Outline 1 Introduction 2 Definition Incompressibility and Randomness 3 Prefix Complexity Resource-Bounded K-Complexity 4 Incompressibility Method Gödel s Incompleteness Theorem 5 Outline

More information

Entropy. Probability and Computing. Presentation 22. Probability and Computing Presentation 22 Entropy 1/39

Entropy. Probability and Computing. Presentation 22. Probability and Computing Presentation 22 Entropy 1/39 Entropy Probability and Computing Presentation 22 Probability and Computing Presentation 22 Entropy 1/39 Introduction Why randomness and information are related? An event that is almost certain to occur

More information

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14 CS 70 Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14 Introduction One of the key properties of coin flips is independence: if you flip a fair coin ten times and get ten

More information

Lecture 11: Information theory THURSDAY, FEBRUARY 21, 2019

Lecture 11: Information theory THURSDAY, FEBRUARY 21, 2019 Lecture 11: Information theory DANIEL WELLER THURSDAY, FEBRUARY 21, 2019 Agenda Information and probability Entropy and coding Mutual information and capacity Both images contain the same fraction of black

More information

Entropy and Ergodic Theory Lecture 3: The meaning of entropy in information theory

Entropy and Ergodic Theory Lecture 3: The meaning of entropy in information theory Entropy and Ergodic Theory Lecture 3: The meaning of entropy in information theory 1 The intuitive meaning of entropy Modern information theory was born in Shannon s 1948 paper A Mathematical Theory of

More information

Shannon's Theory of Communication

Shannon's Theory of Communication Shannon's Theory of Communication An operational introduction 5 September 2014, Introduction to Information Systems Giovanni Sileno g.sileno@uva.nl Leibniz Center for Law University of Amsterdam Fundamental

More information

1. When applied to an affected person, the test comes up positive in 90% of cases, and negative in 10% (these are called false negatives ).

1. When applied to an affected person, the test comes up positive in 90% of cases, and negative in 10% (these are called false negatives ). CS 70 Discrete Mathematics for CS Spring 2006 Vazirani Lecture 8 Conditional Probability A pharmaceutical company is marketing a new test for a certain medical condition. According to clinical trials,

More information

MATH MW Elementary Probability Course Notes Part I: Models and Counting

MATH MW Elementary Probability Course Notes Part I: Models and Counting MATH 2030 3.00MW Elementary Probability Course Notes Part I: Models and Counting Tom Salisbury salt@yorku.ca York University Winter 2010 Introduction [Jan 5] Probability: the mathematics used for Statistics

More information

Discrete Mathematics and Probability Theory Spring 2014 Anant Sahai Note 10

Discrete Mathematics and Probability Theory Spring 2014 Anant Sahai Note 10 EECS 70 Discrete Mathematics and Probability Theory Spring 2014 Anant Sahai Note 10 Introduction to Basic Discrete Probability In the last note we considered the probabilistic experiment where we flipped

More information

Information Theory and Coding Techniques: Chapter 1.1. What is Information Theory? Why you should take this course?

Information Theory and Coding Techniques: Chapter 1.1. What is Information Theory? Why you should take this course? Information Theory and Coding Techniques: Chapter 1.1 What is Information Theory? Why you should take this course? 1 What is Information Theory? Information Theory answers two fundamental questions in

More information

3F1 Information Theory, Lecture 3

3F1 Information Theory, Lecture 3 3F1 Information Theory, Lecture 3 Jossy Sayir Department of Engineering Michaelmas 2013, 29 November 2013 Memoryless Sources Arithmetic Coding Sources with Memory Markov Example 2 / 21 Encoding the output

More information

Lecture 1: Probability Fundamentals

Lecture 1: Probability Fundamentals Lecture 1: Probability Fundamentals IB Paper 7: Probability and Statistics Carl Edward Rasmussen Department of Engineering, University of Cambridge January 22nd, 2008 Rasmussen (CUED) Lecture 1: Probability

More information

Multimedia Communications. Mathematical Preliminaries for Lossless Compression

Multimedia Communications. Mathematical Preliminaries for Lossless Compression Multimedia Communications Mathematical Preliminaries for Lossless Compression What we will see in this chapter Definition of information and entropy Modeling a data source Definition of coding and when

More information

Compression Complexity

Compression Complexity Compression Complexity Stephen Fenner University of South Carolina Lance Fortnow Georgia Institute of Technology February 15, 2017 Abstract The Kolmogorov complexity of x, denoted C(x), is the length of

More information

3F1 Information Theory, Lecture 3

3F1 Information Theory, Lecture 3 3F1 Information Theory, Lecture 3 Jossy Sayir Department of Engineering Michaelmas 2011, 28 November 2011 Memoryless Sources Arithmetic Coding Sources with Memory 2 / 19 Summary of last lecture Prefix-free

More information

Lecture 11: Continuous-valued signals and differential entropy

Lecture 11: Continuous-valued signals and differential entropy Lecture 11: Continuous-valued signals and differential entropy Biology 429 Carl Bergstrom September 20, 2008 Sources: Parts of today s lecture follow Chapter 8 from Cover and Thomas (2007). Some components

More information

Life and Information Dr. Heinz Lycklama

Life and Information Dr. Heinz Lycklama Life and Information Dr. Heinz Lycklama heinz@osta.com www.heinzlycklama.com/messages @ Dr. Heinz Lycklama 1 Biblical View of Life Gen. 1:11-12, 21, 24-25, herb that yields seed according to its kind,,

More information

Is there an Elegant Universal Theory of Prediction?

Is there an Elegant Universal Theory of Prediction? Is there an Elegant Universal Theory of Prediction? Shane Legg Dalle Molle Institute for Artificial Intelligence Manno-Lugano Switzerland 17th International Conference on Algorithmic Learning Theory Is

More information

Kolmogorov complexity and its applications

Kolmogorov complexity and its applications CS860, Winter, 2010 Kolmogorov complexity and its applications Ming Li School of Computer Science University of Waterloo http://www.cs.uwaterloo.ca/~mli/cs860.html We live in an information society. Information

More information

A Mathematical Theory of Communication

A Mathematical Theory of Communication A Mathematical Theory of Communication Ben Eggers Abstract This paper defines information-theoretic entropy and proves some elementary results about it. Notably, we prove that given a few basic assumptions

More information

STAT2201. Analysis of Engineering & Scientific Data. Unit 3

STAT2201. Analysis of Engineering & Scientific Data. Unit 3 STAT2201 Analysis of Engineering & Scientific Data Unit 3 Slava Vaisman The University of Queensland School of Mathematics and Physics What we learned in Unit 2 (1) We defined a sample space of a random

More information

N-grams. Motivation. Simple n-grams. Smoothing. Backoff. N-grams L545. Dept. of Linguistics, Indiana University Spring / 24

N-grams. Motivation. Simple n-grams. Smoothing. Backoff. N-grams L545. Dept. of Linguistics, Indiana University Spring / 24 L545 Dept. of Linguistics, Indiana University Spring 2013 1 / 24 Morphosyntax We just finished talking about morphology (cf. words) And pretty soon we re going to discuss syntax (cf. sentences) In between,

More information

PERFECT SECRECY AND ADVERSARIAL INDISTINGUISHABILITY

PERFECT SECRECY AND ADVERSARIAL INDISTINGUISHABILITY PERFECT SECRECY AND ADVERSARIAL INDISTINGUISHABILITY BURTON ROSENBERG UNIVERSITY OF MIAMI Contents 1. Perfect Secrecy 1 1.1. A Perfectly Secret Cipher 2 1.2. Odds Ratio and Bias 3 1.3. Conditions for Perfect

More information

CHAPTER 4 ENTROPY AND INFORMATION

CHAPTER 4 ENTROPY AND INFORMATION 4-1 CHAPTER 4 ENTROPY AND INFORMATION In the literature one finds the terms information theory and communication theory used interchangeably. As there seems to be no wellestablished convention for their

More information

HW2 Solutions, for MATH441, STAT461, STAT561, due September 9th

HW2 Solutions, for MATH441, STAT461, STAT561, due September 9th HW2 Solutions, for MATH44, STAT46, STAT56, due September 9th. You flip a coin until you get tails. Describe the sample space. How many points are in the sample space? The sample space consists of sequences

More information

An introduction to basic information theory. Hampus Wessman

An introduction to basic information theory. Hampus Wessman An introduction to basic information theory Hampus Wessman Abstract We give a short and simple introduction to basic information theory, by stripping away all the non-essentials. Theoretical bounds on

More information

INTRODUCTION TO INFORMATION THEORY

INTRODUCTION TO INFORMATION THEORY INTRODUCTION TO INFORMATION THEORY KRISTOFFER P. NIMARK These notes introduce the machinery of information theory which is a eld within applied mathematics. The material can be found in most textbooks

More information

Final Exam Comments. UVa - cs302: Theory of Computation Spring < Total

Final Exam Comments. UVa - cs302: Theory of Computation Spring < Total UVa - cs302: Theory of Computation Spring 2008 Final Exam Comments < 50 50 59 60 69 70 79 80 89 90 94 95-102 Total 2 6 8 22 16 16 12 Problem 1: Short Answers. (20) For each question, provide a correct,

More information

Remarks on Random Sequences

Remarks on Random Sequences Remarks on Random Sequences Branden Fitelson & Daniel Osherson 1 Setup We consider evidence relevant to whether a (possibly idealized) physical process is producing its output randomly. For definiteness,

More information

God doesn t play dice. - Albert Einstein

God doesn t play dice. - Albert Einstein ECE 450 Lecture 1 God doesn t play dice. - Albert Einstein As far as the laws of mathematics refer to reality, they are not certain; as far as they are certain, they do not refer to reality. Lecture Overview

More information

Information in Biology

Information in Biology Lecture 3: Information in Biology Tsvi Tlusty, tsvi@unist.ac.kr Living information is carried by molecular channels Living systems I. Self-replicating information processors Environment II. III. Evolve

More information

P (E) = P (A 1 )P (A 2 )... P (A n ).

P (E) = P (A 1 )P (A 2 )... P (A n ). Lecture 9: Conditional probability II: breaking complex events into smaller events, methods to solve probability problems, Bayes rule, law of total probability, Bayes theorem Discrete Structures II (Summer

More information

COS597D: Information Theory in Computer Science October 19, Lecture 10

COS597D: Information Theory in Computer Science October 19, Lecture 10 COS597D: Information Theory in Computer Science October 9, 20 Lecture 0 Lecturer: Mark Braverman Scribe: Andrej Risteski Kolmogorov Complexity In the previous lectures, we became acquainted with the concept

More information

Computer Sciences Department

Computer Sciences Department Computer Sciences Department 1 Reference Book: INTRODUCTION TO THE THEORY OF COMPUTATION, SECOND EDITION, by: MICHAEL SIPSER Computer Sciences Department 3 ADVANCED TOPICS IN C O M P U T A B I L I T Y

More information

Lecture 22: Quantum computational complexity

Lecture 22: Quantum computational complexity CPSC 519/619: Quantum Computation John Watrous, University of Calgary Lecture 22: Quantum computational complexity April 11, 2006 This will be the last lecture of the course I hope you have enjoyed the

More information

2. Polynomials. 19 points. 3/3/3/3/3/4 Clearly indicate your correctly formatted answer: this is what is to be graded. No need to justify!

2. Polynomials. 19 points. 3/3/3/3/3/4 Clearly indicate your correctly formatted answer: this is what is to be graded. No need to justify! 1. Short Modular Arithmetic/RSA. 16 points: 3/3/3/3/4 For each question, please answer in the correct format. When an expression is asked for, it may simply be a number, or an expression involving variables

More information

Kolmogorov Complexity

Kolmogorov Complexity Kolmogorov Complexity 1 Krzysztof Zawada University of Illinois at Chicago E-mail: kzawada@uic.edu Abstract Information theory is a branch of mathematics that attempts to quantify information. To quantify

More information

MACHINE LEARNING INTRODUCTION: STRING CLASSIFICATION

MACHINE LEARNING INTRODUCTION: STRING CLASSIFICATION MACHINE LEARNING INTRODUCTION: STRING CLASSIFICATION THOMAS MAILUND Machine learning means different things to different people, and there is no general agreed upon core set of algorithms that must be

More information

DEMBSKI S SPECIFIED COMPLEXITY: A SIMPLE ERROR IN ARITHMETIC

DEMBSKI S SPECIFIED COMPLEXITY: A SIMPLE ERROR IN ARITHMETIC DEMBSKI S SPECIFIED COMPLEXITY: A SIMPLE ERROR IN ARITHMETIC HOWARD A. LANDMAN Abstract. We show that the derivation of Dembski s formula for specified complexity contains a simple but enormous error,

More information

Lecture 8: Conditional probability I: definition, independence, the tree method, sampling, chain rule for independent events

Lecture 8: Conditional probability I: definition, independence, the tree method, sampling, chain rule for independent events Lecture 8: Conditional probability I: definition, independence, the tree method, sampling, chain rule for independent events Discrete Structures II (Summer 2018) Rutgers University Instructor: Abhishek

More information

Kotebe Metropolitan University Department of Computer Science and Technology Multimedia (CoSc 4151)

Kotebe Metropolitan University Department of Computer Science and Technology Multimedia (CoSc 4151) Kotebe Metropolitan University Department of Computer Science and Technology Multimedia (CoSc 4151) Chapter Three Multimedia Data Compression Part I: Entropy in ordinary words Claude Shannon developed

More information

Mutual Information & Genotype-Phenotype Association. Norman MacDonald January 31, 2011 CSCI 4181/6802

Mutual Information & Genotype-Phenotype Association. Norman MacDonald January 31, 2011 CSCI 4181/6802 Mutual Information & Genotype-Phenotype Association Norman MacDonald January 31, 2011 CSCI 4181/6802 2 Overview What is information (specifically Shannon Information)? What are information entropy and

More information

Remarks on Random Sequences

Remarks on Random Sequences Australasian Journal of Logic Remarks on Random Sequences Branden Fitelson * and Daniel Osherson ** * Department of Philosophy, Rutgers University ** Department of Psychology, Princeton University Abstract

More information

Turing Machines, diagonalization, the halting problem, reducibility

Turing Machines, diagonalization, the halting problem, reducibility Notes on Computer Theory Last updated: September, 015 Turing Machines, diagonalization, the halting problem, reducibility 1 Turing Machines A Turing machine is a state machine, similar to the ones we have

More information

Information in Biology

Information in Biology Information in Biology CRI - Centre de Recherches Interdisciplinaires, Paris May 2012 Information processing is an essential part of Life. Thinking about it in quantitative terms may is useful. 1 Living

More information

Preface These notes were prepared on the occasion of giving a guest lecture in David Harel's class on Advanced Topics in Computability. David's reques

Preface These notes were prepared on the occasion of giving a guest lecture in David Harel's class on Advanced Topics in Computability. David's reques Two Lectures on Advanced Topics in Computability Oded Goldreich Department of Computer Science Weizmann Institute of Science Rehovot, Israel. oded@wisdom.weizmann.ac.il Spring 2002 Abstract This text consists

More information

Introduction to Information Theory. Uncertainty. Entropy. Surprisal. Joint entropy. Conditional entropy. Mutual information.

Introduction to Information Theory. Uncertainty. Entropy. Surprisal. Joint entropy. Conditional entropy. Mutual information. L65 Dept. of Linguistics, Indiana University Fall 205 Information theory answers two fundamental questions in communication theory: What is the ultimate data compression? What is the transmission rate

More information

Dept. of Linguistics, Indiana University Fall 2015

Dept. of Linguistics, Indiana University Fall 2015 L645 Dept. of Linguistics, Indiana University Fall 2015 1 / 28 Information theory answers two fundamental questions in communication theory: What is the ultimate data compression? What is the transmission

More information

Classification & Information Theory Lecture #8

Classification & Information Theory Lecture #8 Classification & Information Theory Lecture #8 Introduction to Natural Language Processing CMPSCI 585, Fall 2007 University of Massachusetts Amherst Andrew McCallum Today s Main Points Automatically categorizing

More information

Lecture 6 - Random Variables and Parameterized Sample Spaces

Lecture 6 - Random Variables and Parameterized Sample Spaces Lecture 6 - Random Variables and Parameterized Sample Spaces 6.042 - February 25, 2003 We ve used probablity to model a variety of experiments, games, and tests. Throughout, we have tried to compute probabilities

More information

Chaos, Complexity, and Inference (36-462)

Chaos, Complexity, and Inference (36-462) Chaos, Complexity, and Inference (36-462) Lecture 7: Information Theory Cosma Shalizi 3 February 2009 Entropy and Information Measuring randomness and dependence in bits The connection to statistics Long-run

More information

AN ALGORITHMIC INFORMATION THEORY CHALLENGE TO INTELLIGENT DESIGN

AN ALGORITHMIC INFORMATION THEORY CHALLENGE TO INTELLIGENT DESIGN AN ALGORITHMIC INFORMATION THEORY CHALLENGE TO INTELLIGENT DESIGN by Sean Devine Abstract. William Dembski claims to have established a decision process to determine when highly unlikely events observed

More information

Complex Systems Methods 2. Conditional mutual information, entropy rate and algorithmic complexity

Complex Systems Methods 2. Conditional mutual information, entropy rate and algorithmic complexity Complex Systems Methods 2. Conditional mutual information, entropy rate and algorithmic complexity Eckehard Olbrich MPI MiS Leipzig Potsdam WS 2007/08 Olbrich (Leipzig) 26.10.2007 1 / 18 Overview 1 Summary

More information

Lecture 3: Randomness in Computation

Lecture 3: Randomness in Computation Great Ideas in Theoretical Computer Science Summer 2013 Lecture 3: Randomness in Computation Lecturer: Kurt Mehlhorn & He Sun Randomness is one of basic resources and appears everywhere. In computer science,

More information

Lecture Notes on Inductive Definitions

Lecture Notes on Inductive Definitions Lecture Notes on Inductive Definitions 15-312: Foundations of Programming Languages Frank Pfenning Lecture 2 September 2, 2004 These supplementary notes review the notion of an inductive definition and

More information

MATH STUDENT BOOK. 12th Grade Unit 9

MATH STUDENT BOOK. 12th Grade Unit 9 MATH STUDENT BOOK 12th Grade Unit 9 Unit 9 COUNTING PRINCIPLES MATH 1209 COUNTING PRINCIPLES INTRODUCTION 1. PROBABILITY DEFINITIONS, SAMPLE SPACES, AND PROBABILITY ADDITION OF PROBABILITIES 11 MULTIPLICATION

More information

6.895 Randomness and Computation March 19, Lecture Last Lecture: Boosting Weak Learners Into Strong Learners

6.895 Randomness and Computation March 19, Lecture Last Lecture: Boosting Weak Learners Into Strong Learners 6.895 Randomness and Computation March 9, 2008 Lecture 3 Lecturer: Ronitt Rubinfeld Scribe: Edwin Chen Overview. Last Lecture: Boosting Weak Learners Into Strong Learners In the last two lectures, we showed

More information

Introduction to Languages and Computation

Introduction to Languages and Computation Introduction to Languages and Computation George Voutsadakis 1 1 Mathematics and Computer Science Lake Superior State University LSSU Math 400 George Voutsadakis (LSSU) Languages and Computation July 2014

More information

Module 1. Probability

Module 1. Probability Module 1 Probability 1. Introduction In our daily life we come across many processes whose nature cannot be predicted in advance. Such processes are referred to as random processes. The only way to derive

More information

Production-rule complexity of recursive structures

Production-rule complexity of recursive structures Production-rule complexity of recursive structures Konstantin L Kouptsov New York University klk206@panix.com Complex recursive structures, such as fractals, are often described by sets of production rules,

More information

Basic Probability. Introduction

Basic Probability. Introduction Basic Probability Introduction The world is an uncertain place. Making predictions about something as seemingly mundane as tomorrow s weather, for example, is actually quite a difficult task. Even with

More information

6.02 Fall 2012 Lecture #1

6.02 Fall 2012 Lecture #1 6.02 Fall 2012 Lecture #1 Digital vs. analog communication The birth of modern digital communication Information and entropy Codes, Huffman coding 6.02 Fall 2012 Lecture 1, Slide #1 6.02 Fall 2012 Lecture

More information

MATH 521, WEEK 2: Rational and Real Numbers, Ordered Sets, Countable Sets

MATH 521, WEEK 2: Rational and Real Numbers, Ordered Sets, Countable Sets MATH 521, WEEK 2: Rational and Real Numbers, Ordered Sets, Countable Sets 1 Rational and Real Numbers Recall that a number is rational if it can be written in the form a/b where a, b Z and b 0, and a number

More information

Lecture 7: DecisionTrees

Lecture 7: DecisionTrees Lecture 7: DecisionTrees What are decision trees? Brief interlude on information theory Decision tree construction Overfitting avoidance Regression trees COMP-652, Lecture 7 - September 28, 2009 1 Recall:

More information

Intro to Information Theory

Intro to Information Theory Intro to Information Theory Math Circle February 11, 2018 1. Random variables Let us review discrete random variables and some notation. A random variable X takes value a A with probability P (a) 0. Here

More information

CSE468 Information Conflict

CSE468 Information Conflict CSE468 Information Conflict Lecturer: Dr Carlo Kopp, MIEEE, MAIAA, PEng Lecture 02 Introduction to Information Theory Concepts Reference Sources and Bibliography There is an abundance of websites and publications

More information

Bits. Chapter 1. Information can be learned through observation, experiment, or measurement.

Bits. Chapter 1. Information can be learned through observation, experiment, or measurement. Chapter 1 Bits Information is measured in bits, just as length is measured in meters and time is measured in seconds. Of course knowing the amount of information is not the same as knowing the information

More information

CS1800: Strong Induction. Professor Kevin Gold

CS1800: Strong Induction. Professor Kevin Gold CS1800: Strong Induction Professor Kevin Gold Mini-Primer/Refresher on Unrelated Topic: Limits This is meant to be a problem about reasoning about quantifiers, with a little practice of other skills, too

More information

CS 124 Math Review Section January 29, 2018

CS 124 Math Review Section January 29, 2018 CS 124 Math Review Section CS 124 is more math intensive than most of the introductory courses in the department. You re going to need to be able to do two things: 1. Perform some clever calculations to

More information

Fundamentals of Probability CE 311S

Fundamentals of Probability CE 311S Fundamentals of Probability CE 311S OUTLINE Review Elementary set theory Probability fundamentals: outcomes, sample spaces, events Outline ELEMENTARY SET THEORY Basic probability concepts can be cast in

More information

ORF 245 Fundamentals of Statistics Chapter 5 Probability

ORF 245 Fundamentals of Statistics Chapter 5 Probability ORF 245 Fundamentals of Statistics Chapter 5 Probability Robert Vanderbei Oct 2015 Slides last edited on October 14, 2015 http://www.princeton.edu/ rvdb Sample Spaces (aka Populations) and Events When

More information

Discrete distribution. Fitting probability models to frequency data. Hypotheses for! 2 test. ! 2 Goodness-of-fit test

Discrete distribution. Fitting probability models to frequency data. Hypotheses for! 2 test. ! 2 Goodness-of-fit test Discrete distribution Fitting probability models to frequency data A probability distribution describing a discrete numerical random variable For example,! Number of heads from 10 flips of a coin! Number

More information

Chapter 4. Basic Set Theory. 4.1 The Language of Set Theory

Chapter 4. Basic Set Theory. 4.1 The Language of Set Theory Chapter 4 Basic Set Theory There are two good reasons for studying set theory. First, it s a indispensable tool for both logic and mathematics, and even for other fields including computer science, linguistics,

More information

P (A B) P ((B C) A) P (B A) = P (B A) + P (C A) P (A) = P (B A) + P (C A) = Q(A) + Q(B).

P (A B) P ((B C) A) P (B A) = P (B A) + P (C A) P (A) = P (B A) + P (C A) = Q(A) + Q(B). Lectures 7-8 jacques@ucsdedu 41 Conditional Probability Let (Ω, F, P ) be a probability space Suppose that we have prior information which leads us to conclude that an event A F occurs Based on this information,

More information

Computability Theory for Neuroscience

Computability Theory for Neuroscience Computability Theory for Neuroscience by Doug Rubino Abstract Neuronal circuits are ubiquitously held to be the substrate of computation in the brain, information processing in single neurons is though

More information

arxiv:math/ v1 [math.nt] 17 Feb 2003

arxiv:math/ v1 [math.nt] 17 Feb 2003 arxiv:math/0302183v1 [math.nt] 17 Feb 2003 REPRESENTATIONS OF Ω IN NUMBER THEORY: FINITUDE VERSUS PARITY TOBY ORD AND TIEN D. KIEU Abstract. We present a new method for expressing Chaitin s random real,

More information

Massachusetts Institute of Technology Handout J/18.062J: Mathematics for Computer Science May 3, 2000 Professors David Karger and Nancy Lynch

Massachusetts Institute of Technology Handout J/18.062J: Mathematics for Computer Science May 3, 2000 Professors David Karger and Nancy Lynch Massachusetts Institute of Technology Handout 48 6.042J/18.062J: Mathematics for Computer Science May 3, 2000 Professors David Karger and Nancy Lynch Quiz 2 Solutions Problem 1 [10 points] Consider n >

More information

Opleiding Informatica

Opleiding Informatica Opleiding Informatica Tape-quantifying Turing machines in the arithmetical hierarchy Simon Heijungs Supervisors: H.J. Hoogeboom & R. van Vliet BACHELOR THESIS Leiden Institute of Advanced Computer Science

More information

Turing Machines Part Three

Turing Machines Part Three Turing Machines Part Three What problems can we solve with a computer? What kind of computer? Very Important Terminology Let M be a Turing machine. M accepts a string w if it enters an accept state when

More information

About the relationship between formal logic and complexity classes

About the relationship between formal logic and complexity classes About the relationship between formal logic and complexity classes Working paper Comments welcome; my email: armandobcm@yahoo.com Armando B. Matos October 20, 2013 1 Introduction We analyze a particular

More information

Shannon s Noisy-Channel Coding Theorem

Shannon s Noisy-Channel Coding Theorem Shannon s Noisy-Channel Coding Theorem Lucas Slot Sebastian Zur February 2015 Abstract In information theory, Shannon s Noisy-Channel Coding Theorem states that it is possible to communicate over a noisy

More information

Introduction to Information Theory

Introduction to Information Theory Introduction to Information Theory Impressive slide presentations Radu Trîmbiţaş UBB October 2012 Radu Trîmbiţaş (UBB) Introduction to Information Theory October 2012 1 / 19 Transmission of information

More information

Complexity 6: AIT. Outline. Dusko Pavlovic. Kolmogorov. Solomonoff. Chaitin: The number of wisdom RHUL Spring Complexity 6: AIT.

Complexity 6: AIT. Outline. Dusko Pavlovic. Kolmogorov. Solomonoff. Chaitin: The number of wisdom RHUL Spring Complexity 6: AIT. Outline Complexity Theory Part 6: did we achieve? Algorithmic information and logical depth : Algorithmic information : Algorithmic probability : The number of wisdom RHUL Spring 2012 : Logical depth Outline

More information

Lecture Notes on From Rules to Propositions

Lecture Notes on From Rules to Propositions Lecture Notes on From Rules to Propositions 15-816: Substructural Logics Frank Pfenning Lecture 2 September 1, 2016 We review the ideas of ephemeral truth and linear inference with another example from

More information

Deterministic Finite Automata

Deterministic Finite Automata Deterministic Finite Automata COMP2600 Formal Methods for Software Engineering Ranald Clouston Australian National University Semester 2, 2013 COMP 2600 Deterministic Finite Automata 1 Pop quiz What is

More information

Lecture 1: September 25, A quick reminder about random variables and convexity

Lecture 1: September 25, A quick reminder about random variables and convexity Information and Coding Theory Autumn 207 Lecturer: Madhur Tulsiani Lecture : September 25, 207 Administrivia This course will cover some basic concepts in information and coding theory, and their applications

More information

Distribution of Environments in Formal Measures of Intelligence: Extended Version

Distribution of Environments in Formal Measures of Intelligence: Extended Version Distribution of Environments in Formal Measures of Intelligence: Extended Version Bill Hibbard December 2008 Abstract This paper shows that a constraint on universal Turing machines is necessary for Legg's

More information

LECSS Physics 11 Introduction to Physics and Math Methods 1 Revised 8 September 2013 Don Bloomfield

LECSS Physics 11 Introduction to Physics and Math Methods 1 Revised 8 September 2013 Don Bloomfield LECSS Physics 11 Introduction to Physics and Math Methods 1 Physics 11 Introduction to Physics and Math Methods In this introduction, you will get a more in-depth overview of what Physics is, as well as

More information

Lecture 8: Probability

Lecture 8: Probability Lecture 8: Probability The idea of probability is well-known The flipping of a balanced coin can produce one of two outcomes: T (tail) and H (head) and the symmetry between the two outcomes means, of course,

More information

BLAST: Target frequencies and information content Dannie Durand

BLAST: Target frequencies and information content Dannie Durand Computational Genomics and Molecular Biology, Fall 2016 1 BLAST: Target frequencies and information content Dannie Durand BLAST has two components: a fast heuristic for searching for similar sequences

More information

3 PROBABILITY TOPICS

3 PROBABILITY TOPICS Chapter 3 Probability Topics 135 3 PROBABILITY TOPICS Figure 3.1 Meteor showers are rare, but the probability of them occurring can be calculated. (credit: Navicore/flickr) Introduction It is often necessary

More information

17.1 Binary Codes Normal numbers we use are in base 10, which are called decimal numbers. Each digit can be 10 possible numbers: 0, 1, 2, 9.

17.1 Binary Codes Normal numbers we use are in base 10, which are called decimal numbers. Each digit can be 10 possible numbers: 0, 1, 2, 9. ( c ) E p s t e i n, C a r t e r, B o l l i n g e r, A u r i s p a C h a p t e r 17: I n f o r m a t i o n S c i e n c e P a g e 1 CHAPTER 17: Information Science 17.1 Binary Codes Normal numbers we use

More information

Sophistication Revisited

Sophistication Revisited Sophistication Revisited Luís Antunes Lance Fortnow August 30, 007 Abstract Kolmogorov complexity measures the ammount of information in a string as the size of the shortest program that computes the string.

More information

Lecture 13: Foundations of Math and Kolmogorov Complexity

Lecture 13: Foundations of Math and Kolmogorov Complexity 6.045 Lecture 13: Foundations of Math and Kolmogorov Complexity 1 Self-Reference and the Recursion Theorem 2 Lemma: There is a computable function q : Σ* Σ* such that for every string w, q(w) is the description

More information

Pseudorandom Generators

Pseudorandom Generators 8 Pseudorandom Generators Great Ideas in Theoretical Computer Science Saarland University, Summer 2014 andomness is one of the fundamental computational resources and appears everywhere. In computer science,

More information

Some Concepts in Probability and Information Theory

Some Concepts in Probability and Information Theory PHYS 476Q: An Introduction to Entanglement Theory (Spring 2018) Eric Chitambar Some Concepts in Probability and Information Theory We begin this course with a condensed survey of basic concepts in probability

More information

MATH 3C: MIDTERM 1 REVIEW. 1. Counting

MATH 3C: MIDTERM 1 REVIEW. 1. Counting MATH 3C: MIDTERM REVIEW JOE HUGHES. Counting. Imagine that a sports betting pool is run in the following way: there are 20 teams, 2 weeks, and each week you pick a team to win. However, you can t pick

More information

Inferring the Causal Decomposition under the Presence of Deterministic Relations.

Inferring the Causal Decomposition under the Presence of Deterministic Relations. Inferring the Causal Decomposition under the Presence of Deterministic Relations. Jan Lemeire 1,2, Stijn Meganck 1,2, Francesco Cartella 1, Tingting Liu 1 and Alexander Statnikov 3 1-ETRO Department, Vrije

More information

Topics in Computer Mathematics

Topics in Computer Mathematics Random Number Generation (Uniform random numbers) Introduction We frequently need some way to generate numbers that are random (by some criteria), especially in computer science. Simulations of natural

More information