Information Theory for Data Science - Changho Suh

Information Theory for Data Science

(Autor)

Buch | Hardcover
250 Seiten
2023
now publishers Inc (Verlag)
978-1-63828-114-6 (ISBN)
143,40 inkl. MwSt
Demonstrates the modern roles of information theory in a widening array of data science applications. The first and second parts of the book cover the core concepts of information theory; the last part focuses on applications that arise in data science, including social networks, ranking, and machine learning.
Information theory deals with mathematical laws that govern the flow, representation and transmission of information. The most significant achievement of the field is the invention of digital communication which forms the basis of our daily-life digital products such as smart phones, laptops and any IoT devices. Recently it has also found important roles in a spotlight field that has been revolutionized during the past decades: data science.

This book aims at demonstrating modern roles of information theory in a widening array of data science applications. The first and second parts of the book covers the core concepts of information theory: basic concepts on several key notions; and celebrated source and channel coding theorems which concern the fundamental limits of communication. The last part focuses on applications that arise in data science, including social networks, ranking, and machine learning.

The book is written as a text for senior undergraduate and graduate students working on Information Theory and Communications, and it should also prove to be a valuable reference for professionals and engineers from these fields.

Changho Suh is an Associate Professor of Electrical Engineering at KAIST and an Associate Head of KAIST AI Institute. He received the B.S. and M.S. degrees in Electrical Engineering from KAIST in 2000 and 2002 respectively, and the Ph.D. degree in Electrical Engineering and Computer Sciences from UC Berkeley in 2011. From 2011 to 2012, he was a postdoctoral associate at the Research Laboratory of Electronics in MIT. From 2002 to 2006, he was with Samsung Electronics.

1 Source Coding
1.1 Overview of the book
1.2 Entropy and Python exercise
1.3 Mutual information, Kullback-Leibler (KL) divergence and Python exercise Problem Set 1
1.4 Source coding theorem for i.i.d. sources (1/3)
1.5 Source coding theorem for i.i.d. sources (2/3)
1.6 Source coding theorem for i.i.d. sources (3/3)Problem Set 2
1.7 Source code design
1.8 Source coding theorem for general sources
1.9 Huffman code and Python implementation
Problem Set 3
2 Channel Coding
2.1 Statement of channel coding theorem
2.2 Achievability proof for the binary erasure channel
2.3 Achievability proof for the binary symmetric channelProblem Set 4
2.4 Achievability proof for discrete memoryless channels
2.5 Converse proof for discrete memoryless channels
2.6 Source-channel separation theorem and feedback
Problem Set 5
2.7 Polar code: Polarization
2.8 Polar code: Implementation of polarization
2.9 Polar code: Proof of polarization and Python simulation
Problem Set 6
3 Data Science Applications
3.1 Social networks: Fundamental limits
3.2 Social networks: Achievability proof
3.3 Social networks: Converse proof
3.4 Social networks: Algorithm and Python implementationProblem Set 7
3.5 DNA sequencing: Fundamental limits
3.6 DNA sequencing: Achievability proof
3.7 DNA sequencing: Converse proof
3.8 DNA sequencing: Algorithm and Python implementationProblem Set 8
3.9 Top-K ranking: Fundamental limits
3.10 Top-K ranking: Algorithm
3.11 Top-K ranking: Python implementation
Problem Set 9
3.12 Supervised learning: Connection with information theory
3.13 Supervised learning: Logistic regression and cross entropy
3.14 Supervised learning: TensorFlow implementation
Problem Set 10
3.15 Unsupervised learning: Generative modeling
3.16 Generative Adversarial Networks (GANs) and KL divergence
3.17 GANs: TensorFlow implementation
Problem Set 11
3.18 Fair machine learning and mutual information (1/2)
3.19 Fair machine learning and mutual information (2/2)
3.20 Fair machine learning: TensorFlow implementation
Problem Set 12
Appendices
A – Python Basics
B – TensorFlow and Keras Basics
C – Note on Research

Erscheinungsdatum
Reihe/Serie NowOpen
Verlagsort Hanover
Sprache englisch
Maße 156 x 234 mm
Gewicht 774 g
Themenwelt Informatik Datenbanken Data Warehouse / Data Mining
Mathematik / Informatik Informatik Theorie / Studium
ISBN-10 1-63828-114-9 / 1638281149
ISBN-13 978-1-63828-114-6 / 9781638281146
Zustand Neuware
Haben Sie eine Frage zum Produkt?
Mehr entdecken
aus dem Bereich
Datenanalyse für Künstliche Intelligenz

von Jürgen Cleve; Uwe Lämmel

Buch | Softcover (2024)
De Gruyter Oldenbourg (Verlag)
74,95
Daten importieren, bereinigen, umformen und visualisieren

von Hadley Wickham; Mine Çetinkaya-Rundel …

Buch | Softcover (2024)
O'Reilly (Verlag)
54,90