Category:Information theory
From ETHW
The processing of information via the use of applied mathematics and electrical engineering
Subcategories
- Audio coding - the translation of auditory information into digital code
- Channel coding - code used to protect information over a channel by correcting errors resulting from noise or other interference
- Codes - rules for converting one piece of information into another
- Communication channels - a physical or logical connection between two points that allows for the exchange of an information signal
- Decoding - translating from an coded message into the original language or form
- Encoding - the process by which information from a source is changed into symbols to be communicated
- Error compensation - the encoding or transmission of extra information or code to compensate for possible errors
- Information entropy - the level of uncertainty associated with a random variable (often refers to the "Shannon entropy")
- Mutual information - occasionally called transinformation, the quantity that measures the mutual dependence of two random variables
- Rate distortion theory - the branch of information theory which explains lossy data compression and which determines the minimal amount of entropy that should be communicated over a channel
- Speech coding - the use of the data compression of digital audio signals to encode speech
Subcategories
This category has the following 11 subcategories, out of 11 total.
Pages in category "Information theory"
The following 78 pages are in this category, out of 78 total.
B
E
G
H
R
S
- Oral-History:Murray Sachs
- Oral-History:William F. Schreiber
- Manfred R. Schroeder
- Oral-History:Manfred Schroeder
- Oral-History:Mischa Schwartz
- Oral-History:Claude E. Shannon
- Claude Shannon
- Shannon Publishes on Communication Theory
- Amin Shokrollahi
- Archives:Breakout: The 1990s
- Archives:What Is Signal Processing? A Look At the CD Player
- David Slepian
- Neil J. A. Sloane