This preview shows pages 1–10. Sign up to view the full content.
This preview has intentionally blurred sections. Sign up to view the full version.
View Full DocumentThis preview has intentionally blurred sections. Sign up to view the full version.
View Full DocumentThis preview has intentionally blurred sections. Sign up to view the full version.
View Full DocumentThis preview has intentionally blurred sections. Sign up to view the full version.
View Full DocumentThis preview has intentionally blurred sections. Sign up to view the full version.
View Full Document
Unformatted text preview: Outline of the Lecture Outline of the Lecture Information Theory Information Theory Entropy Noise Entropy Mutual Information Entropy Noise Entropy Mutual Information Mutual Information Mutual Information Measures How Much Measures How Much Responses Tell about Responses Tell about Stimuli Stimuli In a blackbox model, we try to describe a system well enough to predict its responses without knowing what is inside the system. If the firing is different when one presents the same stimulus twice, then how does the brain know what is in the stimulus? We can also ask how much a system can tell about the ensemble of stimuli in the world; the answer comes from information theory. Let’s represent the probability of each input, P(s), and each output, P(r), by color brightness. Input Variable (s) Output Variable (r) Let’s represent the probability of an output given an input, P(rs), by an arrow thickness. Input Variable (s) Output Variable (r) Each input variable has its own probabilities to lead to different outputs. Input Variable (s) Output Variable (r) If the output distribution is sharp, then the output is not informative about the input....
View
Full
Document
This note was uploaded on 06/08/2009 for the course BME 575L taught by Professor Grzywacz during the Spring '09 term at USC.
 Spring '09
 Grzywacz

Click to edit the document details