Algorithmic Randomness and Complexity
Seiten
2016
|
Softcover reprint of the original 1st ed. 2010
Springer-Verlag New York Inc.
978-1-4939-3820-9 (ISBN)
Springer-Verlag New York Inc.
978-1-4939-3820-9 (ISBN)
The theory of algorithmic randomness uses tools from computability theory and algorithmic information theory to address questions such as these. It will be of interest to researchers and students in computability theory, algorithmic information theory, and theoretical computer science.
Intuitively, a sequence such as 101010101010101010… does not seem random, whereas 101101011101010100…, obtained using coin tosses, does. How can we reconcile this intuition with the fact that both are statistically equally likely? What does it mean to say that an individual mathematical object such as a real number is random, or to say that one real is more random than another? And what is the relationship between randomness and computational power. The theory of algorithmic randomness uses tools from computability theory and algorithmic information theory to address questions such as these. Much of this theory can be seen as exploring the relationships between three fundamental concepts: relative computability, as measured by notions such as Turing reducibility; information content, as measured by notions such as Kolmogorov complexity; and randomness of individual objects, as first successfully defined by Martin-Löf. Although algorithmic randomness has been studied for several decades, a dramatic upsurge of interest in the area, starting in the late 1990s, has led to significant advances. This is the first comprehensive treatment of this important field, designed to be both a reference tool for experts and a guide for newcomers. It surveys a broad section of work in the area, and presents most of its major results and techniques in depth. Its organization is designed to guide the reader through this large body of work, providing context for its many concepts and theorems, discussing their significance, and highlighting their interactions. It includes a discussion of effective dimension, which allows us to assign concepts like Hausdorff dimension to individual reals, and a focused but detailed introduction to computability theory. It will be of interest to researchers and students in computability theory, algorithmic information theory, and theoretical computer science.
Intuitively, a sequence such as 101010101010101010… does not seem random, whereas 101101011101010100…, obtained using coin tosses, does. How can we reconcile this intuition with the fact that both are statistically equally likely? What does it mean to say that an individual mathematical object such as a real number is random, or to say that one real is more random than another? And what is the relationship between randomness and computational power. The theory of algorithmic randomness uses tools from computability theory and algorithmic information theory to address questions such as these. Much of this theory can be seen as exploring the relationships between three fundamental concepts: relative computability, as measured by notions such as Turing reducibility; information content, as measured by notions such as Kolmogorov complexity; and randomness of individual objects, as first successfully defined by Martin-Löf. Although algorithmic randomness has been studied for several decades, a dramatic upsurge of interest in the area, starting in the late 1990s, has led to significant advances. This is the first comprehensive treatment of this important field, designed to be both a reference tool for experts and a guide for newcomers. It surveys a broad section of work in the area, and presents most of its major results and techniques in depth. Its organization is designed to guide the reader through this large body of work, providing context for its many concepts and theorems, discussing their significance, and highlighting their interactions. It includes a discussion of effective dimension, which allows us to assign concepts like Hausdorff dimension to individual reals, and a focused but detailed introduction to computability theory. It will be of interest to researchers and students in computability theory, algorithmic information theory, and theoretical computer science.
Background.- Preliminaries.- Computability Theory.- Kolmogorov Complexity of Finite Strings.- Relating Complexities.- Effective Reals.- Notions of Randomness.- Martin-Löf Randomness.- Other Notions of Algorithmic Randomness.- Algorithmic Randomness and Turing Reducibility.- Relative Randomness.- Measures of Relative Randomness.- Complexity and Relative Randomness for 1-Random Sets.- Randomness-Theoretic Weakness.- Lowness and Triviality for Other Randomness Notions.- Algorithmic Dimension.- Further Topics.- Strong Jump Traceability.- ? as an Operator.- Complexity of Computably Enumerable Sets.
Erscheinungsdatum | 08.07.2016 |
---|---|
Reihe/Serie | Theory and Applications of Computability |
Zusatzinfo | XXVIII, 855 p. |
Verlagsort | New York |
Sprache | englisch |
Maße | 155 x 235 mm |
Themenwelt | Informatik ► Software Entwicklung ► User Interfaces (HCI) |
Informatik ► Theorie / Studium ► Algorithmen | |
Mathematik / Informatik ► Mathematik ► Algebra | |
Mathematik / Informatik ► Mathematik ► Analysis | |
Mathematik / Informatik ► Mathematik ► Logik / Mengenlehre | |
Mathematik / Informatik ► Mathematik ► Wahrscheinlichkeit / Kombinatorik | |
ISBN-10 | 1-4939-3820-7 / 1493938207 |
ISBN-13 | 978-1-4939-3820-9 / 9781493938209 |
Zustand | Neuware |
Haben Sie eine Frage zum Produkt? |
Mehr entdecken
aus dem Bereich
aus dem Bereich
Aus- und Weiterbildung nach iSAQB-Standard zum Certified Professional …
Buch | Hardcover (2023)
dpunkt Verlag
34,90 €
Lean UX und Design Thinking: Teambasierte Entwicklung …
Buch | Hardcover (2022)
dpunkt (Verlag)
34,90 €
Wissensverarbeitung - Neuronale Netze
Buch | Hardcover (2023)
Carl Hanser (Verlag)
34,99 €