Summary and Info
Information Theory and Statistical Learning presents theoretical and practical results about information theoretic methods used in the context of statistical learning. The book will present a comprehensive overview of the large range of different methods that have been developed in a multitude of contexts. Each chapter is written by an expert in the field. The book is intended for an interdisciplinary readership working in machine learning, applied statistics, artificial intelligence, biostatistics, computational biology, bioinformatics, web mining or related disciplines.Advance Praise for Information Theory and Statistical Learning:"A new epoch has arrived for information sciences to integrate various disciplines such as information theory, machine learning, statistical inference, data mining, model selection etc. I am enthusiastic about recommending the present book to researchers and students, because it summarizes most of these new emerging subjects and methods, which are otherwise scattered in many places."-- Shun-ichi Amari, RIKEN Brain Science Institute, Professor-Emeritus at the University of Tokyo
More About the Author
Ray Solomonoff (July 25, 1926 – December 7, 2009) was the inventor of algorithmic probability, his General Theory of Inductive Inference (also known as Universal Inductive Inference), and was a founder of algorithmic information theory.
Review and Comments
Rate the Book
Information Theory and Statistical Learning 0 out of 5 stars based on 0 ratings.