fbpx
Wikipedia

Solomonoff's theory of inductive inference

Solomonoff's theory of inductive inference is a mathematical theory of induction introduced by Ray Solomonoff, based on probability theory and theoretical computer science.[1][2][3] In essence, Solomonoff's induction derives the posterior probability of any computable theory, given a sequence of observed data. This posterior probability is derived from Bayes' rule and some universal prior, that is, a prior that assigns a positive probability to any computable theory.

Solomonoff's induction naturally formalizes Occam's razor[4][5][6][7][8] by assigning larger prior credences to theories that require a shorter algorithmic description.

Origin Edit

Philosophical Edit

The theory is based in philosophical foundations, and was founded by Ray Solomonoff around 1960.[9] It is a mathematically formalized combination of Occam's razor[4][5][6][7][8] and the Principle of Multiple Explanations.[10] All computable theories which perfectly describe previous observations are used to calculate the probability of the next observation, with more weight put on the shorter computable theories. Marcus Hutter's universal artificial intelligence builds upon this to calculate the expected value of an action.

Principle Edit

Solomonoff's induction has been argued to be the computational formalization of pure Bayesianism.[3] To understand, recall that Bayesianism derives the posterior probability   of a theory   given data   by applying Bayes rule, which yields  , where theories   are alternatives to theory  . For this equation to make sense, the quantities   and   must be well-defined for all theories   and  . In other words, any theory must define a probability distribution over observable data  . Solomonoff's induction essentially boils down to demanding in addition that all such probability distributions be computable.

Interestingly, the set of computable probability distributions is a subset of the set of all programs, which is countable. Similarly, the sets of observable data considered by Solomonoff were finite. Without loss of generality, we can thus consider that any observable data is a finite bit string. As a result, Solomonoff's induction can be defined by only invoking discrete probability distributions.

Solomonoff's induction then allows to make probabilistic predictions of future data  , by simply obeying the laws of probability. Namely, we have  . This quantity can be interpreted as the average predictions   of all theories   given past data  , weighted by their posterior credences  .

Mathematical Edit

The proof of the "razor" is based on the known mathematical properties of a probability distribution over a countable set. These properties are relevant because the infinite set of all programs is a denumerable set. The sum S of the probabilities of all programs must be exactly equal to one (as per the definition of probability) thus the probabilities must roughly decrease as we enumerate the infinite set of all programs, otherwise S will be strictly greater than one. To be more precise, for every   > 0, there is some length l such that the probability of all programs longer than l is at most  . This does not, however, preclude very long programs from having very high probability.

Fundamental ingredients of the theory are the concepts of algorithmic probability and Kolmogorov complexity. The universal prior probability of any prefix p of a computable sequence x is the sum of the probabilities of all programs (for a universal computer) that compute something starting with p. Given some p and any computable but unknown probability distribution from which x is sampled, the universal prior and Bayes' theorem can be used to predict the yet unseen parts of x in optimal fashion.

Mathematical guarantees Edit

Solomonoff's completeness Edit

The remarkable property of Solomonoff's induction is its completeness. In essence, the completeness theorem guarantees that the expected cumulative errors made by the predictions based on Solomonoff's induction are upper-bounded by the Kolmogorov complexity of the (stochastic) data generating process. The errors can be measured using the Kullback–Leibler divergence or the square of the difference between the induction's prediction and the probability assigned by the (stochastic) data generating process.

Solomonoff's uncomputability Edit

Unfortunately, Solomonoff also proved that Solomonoff's induction is uncomputable. In fact, he showed that computability and completeness are mutually exclusive: any complete theory must be uncomputable. The proof of this is derived from a game between the induction and the environment. Essentially, any computable induction can be tricked by a computable environment, by choosing the computable environment that negates the computable induction's prediction. This fact can be regarded as an instance of the no free lunch theorem.

Modern applications Edit

Artificial intelligence Edit

Though Solomonoff's inductive inference is not computable, several AIXI-derived algorithms approximate it in order to make it run on a modern computer. The more computing power they are given, the closer their predictions are to the predictions of inductive inference (their mathematical limit is Solomonoff's inductive inference).[11][12][13]

Another direction of inductive inference is based on E. Mark Gold's model of learning in the limit from 1967 and has developed since then more and more models of learning.[14] The general scenario is the following: Given a class S of computable functions, is there a learner (that is, recursive functional) which for any input of the form (f(0),f(1),...,f(n)) outputs a hypothesis (an index e with respect to a previously agreed on acceptable numbering of all computable functions; the indexed function may be required consistent with the given values of f). A learner M learns a function f if almost all its hypotheses are the same index e, which generates the function f; M learns S if M learns every f in S. Basic results are that all recursively enumerable classes of functions are learnable while the class REC of all computable functions is not learnable.[citation needed] Many related models have been considered and also the learning of classes of recursively enumerable sets from positive data is a topic studied from Gold's pioneering paper in 1967 onwards. A far reaching extension of the Gold’s approach is developed by Schmidhuber's theory of generalized Kolmogorov complexities,[15] which are kinds of super-recursive algorithms.

Turing machines Edit

The third mathematically based direction of inductive inference makes use of the theory of automata and computation. In this context, the process of inductive inference is performed by an abstract automaton called an inductive Turing machine (Burgin, 2005). Inductive Turing machines represent the next step in the development of computer science providing better models for contemporary computers and computer networks (Burgin, 2001) and forming an important class of super-recursive algorithms as they satisfy all conditions in the definition of algorithm. Namely, each inductive Turing machine is a type of effective method in which a definite list of well-defined instructions for completing a task, when given an initial state, will proceed through a well-defined series of successive states, eventually terminating in an end-state. The difference between an inductive Turing machine and a Turing machine is that to produce the result a Turing machine has to stop, while in some cases an inductive Turing machine can do this without stopping. Stephen Kleene called procedures that could run forever without stopping by the name calculation procedure or algorithm (Kleene 1952:137). Kleene also demanded that such an algorithm must eventually exhibit "some object" (Kleene 1952:137). This condition is satisfied by inductive Turing machines, as their results are exhibited after a finite number of steps, but inductive Turing machines do not always tell at which step the result has been obtained.

Simple inductive Turing machines are equivalent to other models of computation. More advanced inductive Turing machines are much more powerful. It is proved (Burgin, 2005) that limiting partial recursive functions, trial and error predicates, general Turing machines, and simple inductive Turing machines are equivalent models of computation. However, simple inductive Turing machines and general Turing machines give direct constructions of computing automata, which are thoroughly grounded in physical machines. In contrast, trial and error predicates, limiting recursive functions and limiting partial recursive functions present syntactic systems of symbols with formal rules for their manipulation. Simple inductive Turing machines and general Turing machines are related to limiting partial recursive functions and trial and error predicates as Turing machines are related to partial recursive functions and lambda-calculus.

Note that only simple inductive Turing machines have the same structure (but different functioning semantics of the output mode) as Turing machines. Other types of inductive Turing machines have an essentially more advanced structure due to the structured memory and more powerful instructions. Their utilization for inference and learning allows achieving higher efficiency and better reflects learning of people (Burgin and Klinger, 2004).

Some researchers confuse computations of inductive Turing machines with non-stopping computations or with infinite time computations. First, some of computations of inductive Turing machines halt. As in the case of conventional Turing machines, some halting computations give the result, while others do not give. Second, some non-stopping computations of inductive Turing machines give results, while others do not give. Rules of inductive Turing machines determine when a computation (stopping or non-stopping) gives a result. Namely, an inductive Turing machine produces output from time to time and once this output stops changing, it is considered the result of the computation. It is necessary to know that descriptions of this rule in some papers are incorrect. For instance, Davis (2006: 128) formulates the rule when result is obtained without stopping as "once the correct output has been produced any subsequent output will simply repeat this correct result." Third, in contrast to the widespread misconception, inductive Turing machines give results (when it happens) always after a finite number of steps (in finite time) in contrast to infinite and infinite-time computations. There are two main distinctions between conventional Turing machines and simple inductive Turing machines. The first distinction is that even simple inductive Turing machines can do much more than conventional Turing machines. The second distinction is that a conventional Turing machine always informs (by halting or by coming to a final state) when the result is obtained, while a simple inductive Turing machine in some cases does inform about reaching the result, while in other cases (where the conventional Turing machine is helpless), it does not inform. People have an illusion that a computer always itself informs (by halting or by other means) when the result is obtained. In contrast to this, users themselves have to decide in many cases whether the computed result is what they need or it is necessary to continue computations. Indeed, everyday desktop computer applications like word processors and spreadsheets spend most of their time waiting in event loops, and do not terminate until directed to do so by users.

Evolutionary inductive Turing machines Edit

Evolutionary approach to inductive inference is accomplished by another class of automata called evolutionary inductive Turing machines (Burgin and Eberbach, 2009; 2012). An evolutionary inductive Turing machine is a (possibly infinite) sequence E = {A[t]; t = 1, 2, 3, ... } of inductive Turing machines A[t] each working on generations X[t] which are coded as words in the alphabet of the machines A[t]. The goal is to build a “population” Z satisfying the inference condition. The automaton A[t] called a component, or a level automaton, of E represents (encodes) a one-level evolutionary algorithm that works with input generations X[i] of the population by applying the variation operators v and selection operator s. The first generation X[0] is given as input to E and is processed by the automaton A[1], which generates/produces the first generation X[1] as its transfer output, which goes to the automaton A[2]. For all t = 1, 2, 3, ..., the automaton A[t] receives the generation X[t − 1] as its input from A[t − 1] and then applies the variation operator v and selection operator s, producing the generation X[i + 1] and sending it to A[t + 1] to continue evolution.

See also Edit

References Edit

  1. ^ Zenil, Hector (2011-02-11). Randomness Through Computation: Some Answers, More Questions. World Scientific. ISBN 978-981-4462-63-1.
  2. ^ Solomonoff, Ray J. (2009), Emmert-Streib, Frank; Dehmer, Matthias (eds.), "Algorithmic Probability: Theory and Applications", Information Theory and Statistical Learning, Boston, MA: Springer US, pp. 1–23, doi:10.1007/978-0-387-84816-7_1, ISBN 978-0-387-84816-7, retrieved 2020-07-21
  3. ^ a b Hoang, Lê Nguyên. The equation of knowledge : from Bayes' rule to a unified philosophy of science (First ed.). Boca Raton, FL. ISBN 978-0-367-85530-7. OCLC 1162366056.
  4. ^ a b JJ McCall. Induction: From Kolmogorov and Solomonoff to De Finetti and Back to Kolmogorov – Metroeconomica, 2004 – Wiley Online Library.
  5. ^ a b D Stork. Foundations of Occam's razor and parsimony in learning from ricoh.com – NIPS 2001 Workshop, 2001
  6. ^ a b A.N. Soklakov. Occam's razor as a formal basis for a physical theory from arxiv.org – Foundations of Physics Letters, 2002 – Springer
  7. ^ a b Jose Hernandez-Orallo (1999). "Beyond the Turing Test" (PDF). Journal of Logic, Language and Information. 9.
  8. ^ a b M Hutter. On the existence and convergence of computable universal priors arxiv.org – Algorithmic Learning Theory, 2003 – Springer
  9. ^ Samuel Rathmanner and Marcus Hutter. A philosophical treatise of universal induction. Entropy, 13(6):1076–1136, 2011
  10. ^ Ming Li and Paul Vitanyi, An Introduction to Kolmogorov Complexity and Its Applications. Springer-Verlag, N.Y., 2008p 339 ff.
  11. ^ J. Veness, K.S. Ng, M. Hutter, W. Uther, D. Silver. "A Monte Carlo AIXI Approximation" – Arxiv preprint, 2009 arxiv.org
  12. ^ J. Veness, K.S. Ng, M. Hutter, D. Silver. "Reinforcement Learning via AIXI Approximation" Arxiv preprint, 2010 – aaai.org
  13. ^ S. Pankov. A computational approximation to the AIXI model from agiri.org – Artificial general intelligence, 2008: proceedings of …, 2008 – books.google.com
  14. ^ Gold, E. Mark (1967). "Language identification in the limit" (PDF). Information and Control. 10 (5): 447–474. doi:10.1016/S0019-9958(67)91165-5.
  15. ^ J. Schmidhuber (2002). "Hierarchies of generalized Kolmogorov complexities and nonenumerable universal measures computable in the limit" (PDF). International Journal of Foundations of Computer Science. 13 (4): 587–612. doi:10.1142/S0129054102001291.

Sources Edit

  • Angluin, Dana; Smith, Carl H. (Sep 1983). "Inductive Inference: Theory and Methods". Computing Surveys. 15 (3): 237–269. doi:10.1145/356914.356918. S2CID 3209224.
  • Burgin, M. (2005), Super-recursive Algorithms, Monographs in computer science, Springer. ISBN 0-387-95569-0
  • Burgin, M., "How We Know What Technology Can Do", Communications of the ACM, v. 44, No. 11, 2001, pp. 82–88.
  • Burgin, M.; Eberbach, E., "Universality for Turing Machines, Inductive Turing Machines and Evolutionary Algorithms", Fundamenta Informaticae, v. 91, No. 1, 2009, 53–77.
  • Burgin, M.; Eberbach, E., "On Foundations of Evolutionary Computation: An Evolutionary Automata Approach", in Handbook of Research on Artificial Immune Systems and Natural Computing: Applying Complex Adaptive Technologies (Hongwei Mo, Ed.), IGI Global, Hershey, Pennsylvania, 2009, 342–360.
  • Burgin, M.; Eberbach, E., "Evolutionary Automata: Expressiveness and Convergence of Evolutionary Computation", Computer Journal, v. 55, No. 9, 2012, pp. 1023–1029.
  • Burgin, M.; Klinger, A. Experience, Generations, and Limits in Machine Learning, Theoretical Computer Science, v. 317, No. 1/3, 2004, pp. 71–91
  • Davis, Martin (2006) "The Church–Turing Thesis: Consensus and opposition]". Proceedings, Computability in Europe 2006. Lecture Notes in Computer Science, 3988 pp. 125–132.
  • Gasarch, W.; Smith, C. H. (1997) "A survey of inductive inference with an emphasis on queries". Complexity, logic, and recursion theory, Lecture Notes in Pure and Appl. Math., 187, Dekker, New York, pp. 225–260.
  • Hay, Nick. "Universal Semimeasures: An Introduction," CDMTCS Research Report Series, University of Auckland, Feb. 2007.
  • Jain, Sanjay ; Osherson, Daniel ; Royer, James ; Sharma, Arun, Systems that Learn: An Introduction to Learning Theory (second edition), MIT Press, 1999.
  • Kleene, Stephen C. (1952), Introduction to Metamathematics (First ed.), Amsterdam: North-Holland.
  • Li Ming; Vitanyi, Paul, An Introduction to Kolmogorov Complexity and Its Applications, 2nd Edition, Springer Verlag, 1997.
  • Osherson, Daniel ; Stob, Michael ; Weinstein, Scott, Systems That Learn, An Introduction to Learning Theory for Cognitive and Computer Scientists, MIT Press, 1986.
  • Solomonoff, Ray J. (1999). "Two Kinds of Probabilistic Induction" (PDF). The Computer Journal. 42 (4): 256. CiteSeerX 10.1.1.68.8941. doi:10.1093/comjnl/42.4.256.
  • Solomonoff, Ray (March 1964). "A Formal Theory of Inductive Inference Part I" (PDF). Information and Control. 7 (1): 1–22. doi:10.1016/S0019-9958(64)90223-2.
  • Solomonoff, Ray (June 1964). "A Formal Theory of Inductive Inference Part II" (PDF). Information and Control. 7 (2): 224–254. doi:10.1016/S0019-9958(64)90131-7.

External links Edit

  • Algorithmic probability – Scholarpedia

solomonoff, theory, inductive, inference, this, article, multiple, issues, please, help, improve, discuss, these, issues, talk, page, learn, when, remove, these, template, messages, this, article, confusing, unclear, readers, particular, large, swathes, this, . This article has multiple issues Please help improve it or discuss these issues on the talk page Learn how and when to remove these template messages This article may be confusing or unclear to readers In particular large swathes of this article are filled with jargon and unsourced vague generalizations Please help clarify the article There might be a discussion about this on the talk page June 2017 Learn how and when to remove this template message This article needs additional citations for verification Please help improve this article by adding citations to reliable sources Unsourced material may be challenged and removed Find sources Solomonoff s theory of inductive inference news newspapers books scholar JSTOR June 2017 Learn how and when to remove this template message Learn how and when to remove this template message Solomonoff s theory of inductive inference is a mathematical theory of induction introduced by Ray Solomonoff based on probability theory and theoretical computer science 1 2 3 In essence Solomonoff s induction derives the posterior probability of any computable theory given a sequence of observed data This posterior probability is derived from Bayes rule and some universal prior that is a prior that assigns a positive probability to any computable theory Solomonoff s induction naturally formalizes Occam s razor 4 5 6 7 8 by assigning larger prior credences to theories that require a shorter algorithmic description Contents 1 Origin 1 1 Philosophical 1 2 Principle 1 3 Mathematical 2 Mathematical guarantees 2 1 Solomonoff s completeness 2 2 Solomonoff s uncomputability 3 Modern applications 3 1 Artificial intelligence 3 2 Turing machines 3 2 1 Evolutionary inductive Turing machines 4 See also 5 References 6 Sources 7 External linksOrigin EditPhilosophical Edit The theory is based in philosophical foundations and was founded by Ray Solomonoff around 1960 9 It is a mathematically formalized combination of Occam s razor 4 5 6 7 8 and the Principle of Multiple Explanations 10 All computable theories which perfectly describe previous observations are used to calculate the probability of the next observation with more weight put on the shorter computable theories Marcus Hutter s universal artificial intelligence builds upon this to calculate the expected value of an action Principle Edit Solomonoff s induction has been argued to be the computational formalization of pure Bayesianism 3 To understand recall that Bayesianism derives the posterior probability P T D displaystyle mathbb P T D of a theory T displaystyle T given data D displaystyle D by applying Bayes rule which yields P T D P D T P T P D T P T A T P D A P A displaystyle mathbb P T D mathbb P D T mathbb P T mathbb P D T mathbb P T sum A neq T mathbb P D A mathbb P A where theories A displaystyle A are alternatives to theory T displaystyle T For this equation to make sense the quantities P D T displaystyle mathbb P D T and P D A displaystyle mathbb P D A must be well defined for all theories T displaystyle T and A displaystyle A In other words any theory must define a probability distribution over observable data D displaystyle D Solomonoff s induction essentially boils down to demanding in addition that all such probability distributions be computable Interestingly the set of computable probability distributions is a subset of the set of all programs which is countable Similarly the sets of observable data considered by Solomonoff were finite Without loss of generality we can thus consider that any observable data is a finite bit string As a result Solomonoff s induction can be defined by only invoking discrete probability distributions Solomonoff s induction then allows to make probabilistic predictions of future data F displaystyle F by simply obeying the laws of probability Namely we have P F D E T P F T D T P F T D P T D displaystyle mathbb P F D mathbb E T mathbb P F T D sum T mathbb P F T D mathbb P T D This quantity can be interpreted as the average predictions P F T D displaystyle mathbb P F T D of all theories T displaystyle T given past data D displaystyle D weighted by their posterior credences P T D displaystyle mathbb P T D Mathematical Edit The proof of the razor is based on the known mathematical properties of a probability distribution over a countable set These properties are relevant because the infinite set of all programs is a denumerable set The sum S of the probabilities of all programs must be exactly equal to one as per the definition of probability thus the probabilities must roughly decrease as we enumerate the infinite set of all programs otherwise S will be strictly greater than one To be more precise for every ϵ displaystyle epsilon gt 0 there is some length l such that the probability of all programs longer than l is at most ϵ displaystyle epsilon This does not however preclude very long programs from having very high probability Fundamental ingredients of the theory are the concepts of algorithmic probability and Kolmogorov complexity The universal prior probability of any prefix p of a computable sequence x is the sum of the probabilities of all programs for a universal computer that compute something starting with p Given some p and any computable but unknown probability distribution from which x is sampled the universal prior and Bayes theorem can be used to predict the yet unseen parts of x in optimal fashion Mathematical guarantees EditSolomonoff s completeness Edit The remarkable property of Solomonoff s induction is its completeness In essence the completeness theorem guarantees that the expected cumulative errors made by the predictions based on Solomonoff s induction are upper bounded by the Kolmogorov complexity of the stochastic data generating process The errors can be measured using the Kullback Leibler divergence or the square of the difference between the induction s prediction and the probability assigned by the stochastic data generating process Solomonoff s uncomputability Edit Unfortunately Solomonoff also proved that Solomonoff s induction is uncomputable In fact he showed that computability and completeness are mutually exclusive any complete theory must be uncomputable The proof of this is derived from a game between the induction and the environment Essentially any computable induction can be tricked by a computable environment by choosing the computable environment that negates the computable induction s prediction This fact can be regarded as an instance of the no free lunch theorem Modern applications EditArtificial intelligence Edit Though Solomonoff s inductive inference is not computable several AIXI derived algorithms approximate it in order to make it run on a modern computer The more computing power they are given the closer their predictions are to the predictions of inductive inference their mathematical limit is Solomonoff s inductive inference 11 12 13 Another direction of inductive inference is based on E Mark Gold s model of learning in the limit from 1967 and has developed since then more and more models of learning 14 The general scenario is the following Given a class S of computable functions is there a learner that is recursive functional which for any input of the form f 0 f 1 f n outputs a hypothesis an index e with respect to a previously agreed on acceptable numbering of all computable functions the indexed function may be required consistent with the given values of f A learner M learns a function f if almost all its hypotheses are the same index e which generates the function f M learns S if M learns every f in S Basic results are that all recursively enumerable classes of functions are learnable while the class REC of all computable functions is not learnable citation needed Many related models have been considered and also the learning of classes of recursively enumerable sets from positive data is a topic studied from Gold s pioneering paper in 1967 onwards A far reaching extension of the Gold s approach is developed by Schmidhuber s theory of generalized Kolmogorov complexities 15 which are kinds of super recursive algorithms Turing machines Edit This section does not cite any sources Please help improve this section by adding citations to reliable sources Unsourced material may be challenged and removed June 2017 Learn how and when to remove this template message The third mathematically based direction of inductive inference makes use of the theory of automata and computation In this context the process of inductive inference is performed by an abstract automaton called an inductive Turing machine Burgin 2005 Inductive Turing machines represent the next step in the development of computer science providing better models for contemporary computers and computer networks Burgin 2001 and forming an important class of super recursive algorithms as they satisfy all conditions in the definition of algorithm Namely each inductive Turing machine is a type of effective method in which a definite list of well defined instructions for completing a task when given an initial state will proceed through a well defined series of successive states eventually terminating in an end state The difference between an inductive Turing machine and a Turing machine is that to produce the result a Turing machine has to stop while in some cases an inductive Turing machine can do this without stopping Stephen Kleene called procedures that could run forever without stopping by the name calculation procedure or algorithm Kleene 1952 137 Kleene also demanded that such an algorithm must eventually exhibit some object Kleene 1952 137 This condition is satisfied by inductive Turing machines as their results are exhibited after a finite number of steps but inductive Turing machines do not always tell at which step the result has been obtained Simple inductive Turing machines are equivalent to other models of computation More advanced inductive Turing machines are much more powerful It is proved Burgin 2005 that limiting partial recursive functions trial and error predicates general Turing machines and simple inductive Turing machines are equivalent models of computation However simple inductive Turing machines and general Turing machines give direct constructions of computing automata which are thoroughly grounded in physical machines In contrast trial and error predicates limiting recursive functions and limiting partial recursive functions present syntactic systems of symbols with formal rules for their manipulation Simple inductive Turing machines and general Turing machines are related to limiting partial recursive functions and trial and error predicates as Turing machines are related to partial recursive functions and lambda calculus Note that only simple inductive Turing machines have the same structure but different functioning semantics of the output mode as Turing machines Other types of inductive Turing machines have an essentially more advanced structure due to the structured memory and more powerful instructions Their utilization for inference and learning allows achieving higher efficiency and better reflects learning of people Burgin and Klinger 2004 Some researchers confuse computations of inductive Turing machines with non stopping computations or with infinite time computations First some of computations of inductive Turing machines halt As in the case of conventional Turing machines some halting computations give the result while others do not give Second some non stopping computations of inductive Turing machines give results while others do not give Rules of inductive Turing machines determine when a computation stopping or non stopping gives a result Namely an inductive Turing machine produces output from time to time and once this output stops changing it is considered the result of the computation It is necessary to know that descriptions of this rule in some papers are incorrect For instance Davis 2006 128 formulates the rule when result is obtained without stopping as once the correct output has been produced any subsequent output will simply repeat this correct result Third in contrast to the widespread misconception inductive Turing machines give results when it happens always after a finite number of steps in finite time in contrast to infinite and infinite time computations There are two main distinctions between conventional Turing machines and simple inductive Turing machines The first distinction is that even simple inductive Turing machines can do much more than conventional Turing machines The second distinction is that a conventional Turing machine always informs by halting or by coming to a final state when the result is obtained while a simple inductive Turing machine in some cases does inform about reaching the result while in other cases where the conventional Turing machine is helpless it does not inform People have an illusion that a computer always itself informs by halting or by other means when the result is obtained In contrast to this users themselves have to decide in many cases whether the computed result is what they need or it is necessary to continue computations Indeed everyday desktop computer applications like word processors and spreadsheets spend most of their time waiting in event loops and do not terminate until directed to do so by users Evolutionary inductive Turing machines Edit Evolutionary approach to inductive inference is accomplished by another class of automata called evolutionary inductive Turing machines Burgin and Eberbach 2009 2012 An evolutionary inductive Turing machine is a possibly infinite sequence E A t t 1 2 3 of inductive Turing machines A t each working on generations X t which are coded as words in the alphabet of the machines A t The goal is to build a population Z satisfying the inference condition The automaton A t called a component or a level automaton of E represents encodes a one level evolutionary algorithm that works with input generations X i of the population by applying the variation operators v and selection operator s The first generation X 0 is given as input to E and is processed by the automaton A 1 which generates produces the first generation X 1 as its transfer output which goes to the automaton A 2 For all t 1 2 3 the automaton A t receives the generation X t 1 as its input from A t 1 and then applies the variation operator v and selection operator s producing the generation X i 1 and sending it to A t 1 to continue evolution See also EditAlgorithmic information theory Bayesian inference Language identification in the limit Inductive inference Inductive probability Mill s methods Minimum description length Minimum message length For a philosophical viewpoint see Problem of induction and New riddle of inductionReferences Edit Zenil Hector 2011 02 11 Randomness Through Computation Some Answers More Questions World Scientific ISBN 978 981 4462 63 1 Solomonoff Ray J 2009 Emmert Streib Frank Dehmer Matthias eds Algorithmic Probability Theory and Applications Information Theory and Statistical Learning Boston MA Springer US pp 1 23 doi 10 1007 978 0 387 84816 7 1 ISBN 978 0 387 84816 7 retrieved 2020 07 21 a b Hoang Le Nguyen The equation of knowledge from Bayes rule to a unified philosophy of science First ed Boca Raton FL ISBN 978 0 367 85530 7 OCLC 1162366056 a b JJ McCall Induction From Kolmogorov and Solomonoff to De Finetti and Back to Kolmogorov Metroeconomica 2004 Wiley Online Library a b D Stork Foundations of Occam s razor and parsimony in learning from ricoh com NIPS 2001 Workshop 2001 a b A N Soklakov Occam s razor as a formal basis for a physical theory from arxiv org Foundations of Physics Letters 2002 Springer a b Jose Hernandez Orallo 1999 Beyond the Turing Test PDF Journal of Logic Language and Information 9 a b M Hutter On the existence and convergence of computable universal priors arxiv org Algorithmic Learning Theory 2003 Springer Samuel Rathmanner and Marcus Hutter A philosophical treatise of universal induction Entropy 13 6 1076 1136 2011 Ming Li and Paul Vitanyi An Introduction to Kolmogorov Complexity and Its Applications Springer Verlag N Y 2008p 339 ff J Veness K S Ng M Hutter W Uther D Silver A Monte Carlo AIXI Approximation Arxiv preprint 2009 arxiv org J Veness K S Ng M Hutter D Silver Reinforcement Learning via AIXI Approximation Arxiv preprint 2010 aaai org S Pankov A computational approximation to the AIXI model from agiri org Artificial general intelligence 2008 proceedings of 2008 books google com Gold E Mark 1967 Language identification in the limit PDF Information and Control 10 5 447 474 doi 10 1016 S0019 9958 67 91165 5 J Schmidhuber 2002 Hierarchies of generalized Kolmogorov complexities and nonenumerable universal measures computable in the limit PDF International Journal of Foundations of Computer Science 13 4 587 612 doi 10 1142 S0129054102001291 Sources EditAngluin Dana Smith Carl H Sep 1983 Inductive Inference Theory and Methods Computing Surveys 15 3 237 269 doi 10 1145 356914 356918 S2CID 3209224 Burgin M 2005 Super recursive Algorithms Monographs in computer science Springer ISBN 0 387 95569 0 Burgin M How We Know What Technology Can Do Communications of the ACM v 44 No 11 2001 pp 82 88 Burgin M Eberbach E Universality for Turing Machines Inductive Turing Machines and Evolutionary Algorithms Fundamenta Informaticae v 91 No 1 2009 53 77 Burgin M Eberbach E On Foundations of Evolutionary Computation An Evolutionary Automata Approach in Handbook of Research on Artificial Immune Systems and Natural Computing Applying Complex Adaptive Technologies Hongwei Mo Ed IGI Global Hershey Pennsylvania 2009 342 360 Burgin M Eberbach E Evolutionary Automata Expressiveness and Convergence of Evolutionary Computation Computer Journal v 55 No 9 2012 pp 1023 1029 Burgin M Klinger A Experience Generations and Limits in Machine Learning Theoretical Computer Science v 317 No 1 3 2004 pp 71 91 Davis Martin 2006 The Church Turing Thesis Consensus and opposition Proceedings Computability in Europe 2006 Lecture Notes in Computer Science 3988 pp 125 132 Gasarch W Smith C H 1997 A survey of inductive inference with an emphasis on queries Complexity logic and recursion theory Lecture Notes in Pure and Appl Math 187 Dekker New York pp 225 260 Hay Nick Universal Semimeasures An Introduction CDMTCS Research Report Series University of Auckland Feb 2007 Jain Sanjay Osherson Daniel Royer James Sharma Arun Systems that Learn An Introduction to Learning Theory second edition MIT Press 1999 Kleene Stephen C 1952 Introduction to Metamathematics First ed Amsterdam North Holland Li Ming Vitanyi Paul An Introduction to Kolmogorov Complexity and Its Applications 2nd Edition Springer Verlag 1997 Osherson Daniel Stob Michael Weinstein Scott Systems That Learn An Introduction to Learning Theory for Cognitive and Computer Scientists MIT Press 1986 Solomonoff Ray J 1999 Two Kinds of Probabilistic Induction PDF The Computer Journal 42 4 256 CiteSeerX 10 1 1 68 8941 doi 10 1093 comjnl 42 4 256 Solomonoff Ray March 1964 A Formal Theory of Inductive Inference Part I PDF Information and Control 7 1 1 22 doi 10 1016 S0019 9958 64 90223 2 Solomonoff Ray June 1964 A Formal Theory of Inductive Inference Part II PDF Information and Control 7 2 224 254 doi 10 1016 S0019 9958 64 90131 7 External links EditAlgorithmic probability Scholarpedia Retrieved from https en wikipedia org w index php title Solomonoff 27s theory of inductive inference amp oldid 1166921438, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.