Поиcк по сайту by Google

Rambler's Top100
Образование Крыму » Физика » Entropy and information theory - Gray Robert M.

Entropy and information theory - Gray Robert M.

Название: Entropy and information theory
Автор: Gray Robert M.
Категория: Физика
Тип: Книга
Дата: 21.10.2008 21:08:39
Скачано: 69
Описание: This book is devoted to the theory of probabilistic information measures and their application to coding theorems for information sources and noisy channels. The eventual goal is a general development of Shannon's mathematical theory of communication, but much of the space is devoted to the tools and methods required to prove the Shannon coding theorems. These tools form an area common to ergodic theory and information theory and comprise several quantitative notions of the information in random variables, random processes, and dynamical systems. Examples are entropy, mutual information, conditional entropy, conditional information, and discrimination or relative entropy, along with the limiting normalized versions of these quantities such as entropy rate and information rate. Much of the book is concerned with their properties, especially the long term asymptotic behavior of sample information and expected information. The book has been strongly influenced by M. S. Pinsker's classic Information and Information Stability of Random Variables and Processes and by the seminal work of A. N. Kolmogorov, I. M. Gelfand, A. M. Yaglom, and R. L. Dobrushin on information measures for abstract alphabets and their convergence properties. Many of the results herein are extensions of their generalizations of Shannon's original results. The mathematical models of this treatment are more general than traditional treatments in that nonstationary and nonergodic information processes are treated. The models are somewhat less general than those of the Soviet school of information theory in the sense that standard alphabets rather than completely abstract alphabets are considered. This restriction, however, permits many stronger results as well as the extension to nonergodic processes. In addition, the assumption of standard spaces simplifies many proofs and such spaces include as examples virtually all examples of engineering interest. The information convergence results are combined with ergodic theorems to prove general Shannon coding theorems for sources and channels. The results are not the most general known and the converses are not the strongest available, but they are sufficently general to cover most systems encountered in applications and they provide an introduction to recent extensions requiring significant additional mathematical machinery. Several of the generalizations have not previously been treated in book form. Examples of novel topics for an information theory text include asymptotic mean stationary sources, one-sided sources as well as two-sided sources, nonergodic sources, d-continuous channels,
Файл: 1.02 МБ