| Class | Description |
|---|---|
| Entropy |
Calculates entropy, joint entropy or conditional entropy, which can be used to determine the uncertainty in the states of a discrete distribution.
|
| JensenShannon |
Methods for computing the Jensen Shannon divergence, which measures the similarity between probability distributions.
|
| KullbackLeibler |
Calculate the Kullback–Leibler divergence between 2 distributions with the same variables, D(P||Q).
|
| MutualInformation |
Calculates mutual information or conditional mutual information, which measures the dependence between two variables.
|
| Enum | Description |
|---|---|
| LogarithmBase |
Determines the base of the logarithm to use during calculations such as mutual information.
|
Copyright © 2021. All rights reserved.