Computer Science BooksInformation Theory Books

Information Theory in Computer Science

Information Theory in Computer Science

Information Theory in Computer Science

This note explains the following topics: Shearer's Lemma, Entropy, Relative Entropy, Hypothesis testing, total variation distance and Pinsker's lemma, Stability in Shearer's Lemma, Communication Complexity, Set Disjointness, Direct Sum in Communication Complexity and Internal Information Complexity, Data Structure Lower Bounds via Communication Complexity, Algorithmic Lovasz Local Lemma, Parallel Repetition Theorem, Graph Entropy and Sorting.

Author(s):

sNA Pages
Similar Books
An Introduction to Information Theory and Applications

An Introduction to Information Theory and Applications

This note explains the following topics: uncertainty and information, Efficient coding of information, Stationary processes and markov chains, Coding for noisy transmission, Complements to efficient coding of Information, Error correcting codes and cryptography.

s293 Pages
Lecture Notes on statistics and information Theory

Lecture Notes on statistics and information Theory

This lecture note navigates through information theory, statistics and measure theory. It covers fundamental concepts such as definitions, chain rules, data processing inequalities, and divergences and extends to optimal procedures, LeCam’s and Fano’s inequalities, and operational results like entropy and source coding. It also focus on exponential families and statistical modeling, fitting procedures, and lower bounds on testing parameters, sub-Gaussian and sub-exponential random variables, martingale methods, uniformity covering topics such as Kullback-Leibler divergence, PAC-Bayes bounds, interactive data analysis, and error bounds.

s464 Pages
Basics of information theory

Basics of information theory

This book explains basics of thermodynamics, including thermodynamic potentials, microcanonical and canonical distributions, and evolution in the phase space, The inevitability of irreversibility, basics of information theory, applications of information theory, new second law of thermodynamics and quantum information.

s165 Pages
Information Theory by Y. Polyanskiy

Information Theory by Y. Polyanskiy

This PDF covers the following topics related to Information Theory : Information measures, Lossless data compression, Binary hypothesis testing, Channel coding, Lossy data compression, Advanced topics.

s295 Pages
Information Theory and its applications in theory of computation

Information Theory and its applications in theory of computation

This note covers the following topics: Entropy, Kraft's inequality, Source coding theorem, conditional entropy, mutual information, KL-divergence and connections, KL-divergence and Chernoff bounds, Data processing and Fano's inequalities, Asymptotic Equipartition Property, Universal source coding: Lempel-Ziv algorithm and proof of its optimality, Source coding via typical sets and universality, joint typicality and joint AEP, discrete channels and channel capacity, Proof of Noisy channel coding theorem, Constructing capacity-achieving codes via concatenation, Polarization, Arikan's recursive construction of a polarizing invertible transformation, Polar codes construction, Bregman's theorem, Shearer's Lemma and applications, Source coding and Graph entropy, Monotone formula lower bounds via graph entropy, Optimal set Disjointness lower bound and applications, Compression of arbitrary communication protocols, Parallel repetition of 2-prover 1-round games.

sNA Pages
A Short Course in Information Theory (D. MacKay)

A Short Course in Information Theory (D. MacKay)

Currently this section contains no detailed description for the page, will update this page soon.

s Pages