当前位置: 首页 > 文档资料 > Edward 中文文档 >

ed.inferences

优质
小牛编辑
134浏览
2023-12-01

Defined in edward/inferences/__init__.py.

Classes

class BiGANInference: Adversarially Learned Inference (Dumoulin et al., 2017) or

class GANInference: Parameter estimation with GAN-style training

class Gibbs: Gibbs sampling (Geman & Geman, 1984).

class HMC: Hamiltonian Monte Carlo, also known as hybrid Monte Carlo

class ImplicitKLqp: Variational inference with implicit probabilistic models

class Inference: Abstract base class for inference. All inference algorithms in

class KLpq: Variational inference with the KL divergence

class KLqp: Variational inference with the KL divergence

class Laplace: Laplace approximation (Laplace, 1986).

class MAP: Maximum a posteriori.

class MetropolisHastings: Metropolis-Hastings (Hastings, 1970; Metropolis, Rosenbluth, Rosenbluth, Teller, & Teller, 1953).

class MonteCarlo: Abstract base class for Monte Carlo. Specific Monte Carlo methods

class ReparameterizationEntropyKLqp: Variational inference with the KL divergence

class ReparameterizationKLKLqp: Variational inference with the KL divergence

class ReparameterizationKLqp: Variational inference with the KL divergence

class ReplicaExchangeMC: Replica Exchange MCMC (Hukushima & Nemoto, 1996; Swendsen & Wang, 1986).

class SGHMC: Stochastic gradient Hamiltonian Monte Carlo (Chen, Fox, & Guestrin, 2014).

class SGLD: Stochastic gradient Langevin dynamics (Welling & Teh, 2011).

class ScoreEntropyKLqp: Variational inference with the KL divergence

class ScoreKLKLqp: Variational inference with the KL divergence

class ScoreKLqp: Variational inference with the KL divergence

class ScoreRBKLqp: Variational inference with the KL divergence

class VariationalInference: Abstract base class for variational inference. Specific

class WGANInference: Parameter estimation with GAN-style training

class WakeSleep: Wake-Sleep algorithm (Hinton, Dayan, Frey, & Neal, 1995).

Functions

complete_conditional(...): Returns the conditional distribution RandomVariable

Chen, T., Fox, E. B., & Guestrin, C. (2014). Stochastic gradient Hamiltonian Monte Carlo. In International conference on machine learning.

Dumoulin, V., Belghazi, I., Poole, B., Lamb, A., Arjovsky, M., Mastropietro, O., & Courville, A. (2017). Adversarially Learned Inference. In International conference on learning representations.

Geman, S., & Geman, D. (1984). Stochastic relaxation, gibbs distributions, and the bayesian restoration of images. IEEE Transactions on Pattern Analysis and Machine Intelligence, (6), 721–741.

Hastings, W. K. (1970). Monte Carlo sampling methods using Markov chains and their applications. Biometrika, 57(1), 97–109.

Hinton, G. E., Dayan, P., Frey, B. J., & Neal, R. M. (1995). The "wake-sleep" algorithm for unsupervised neural networks. Science.

Hukushima, K., & Nemoto, K. (1996). Exchange monte carlo method and application to spin glass simulations. Journal of the Physical Society of Japan, 65(6), 1604–1608.

Laplace, P. S. (1986). Memoir on the probability of the causes of events. Statistical Science, 1(3), 364–378.

Metropolis, N., Rosenbluth, A. W., Rosenbluth, M. N., Teller, A. H., & Teller, E. (1953). Equation of state calculations by fast computing machines. The Journal of Chemical Physics, 21(6), 1087–1092.

Swendsen, R. H., & Wang, J.-S. (1986). Replica monte carlo simulation of spin-glasses. Physical Review Letters, 57(21), 2607.

Welling, M., & Teh, Y. W. (2011). Bayesian learning via stochastic gradient Langevin dynamics. In International conference on machine learning.