Information-Theoretic Applications of the Logarithmic Probability Comparison Bound.
Rami AtarNeri MerhavPublished in: IEEE Trans. Inf. Theory (2015)
Keyphrases
- information theoretic
- mutual information
- information theory
- theoretic framework
- worst case
- jensen shannon divergence
- information theoretic measures
- information bottleneck
- kullback leibler divergence
- upper bound
- entropy measure
- kl divergence
- multi modality
- computational learning theory
- distributional clustering
- relative entropy
- multi modal
- log likelihood
- image classification
- lower bound
- image processing
- feature selection