کد مقاله | کد نشریه | سال انتشار | مقاله انگلیسی | نسخه تمام متن |
---|---|---|---|---|
561975 | 875344 | 2007 | 19 صفحه PDF | دانلود رایگان |
Information processing theory endeavors to quantify how well signals encode information and how well systems, by acting on signals, process information. We use information-theoretic distance measures, the Kullback–Leibler distance in particular, to quantify how well signals represent information. The ratio of distances calculated between two informationally different signals at a system's output and input quantifies the system's information processing properties. Using this approach, we derive the fundamental processing capabilities of simple system architectures that apply universally: the systems and the kinds of signals they process and produce do not affect our general results. Applications in array signal processing and in neural signal analysis illustrate how to apply the theory.
Journal: Signal Processing - Volume 87, Issue 6, June 2007, Pages 1326–1344