Skip to main content

Questions tagged [variational-inference]

1 vote
1 answer
56 views

The Google Deepmind paper "Weight Uncertainty in Neural Networks" features the following algorithm: Note that the $\frac{∂f(w,θ)}{∂w}$ term of the gradients for the mean and standard ...
user494234's user avatar
3 votes
0 answers
60 views

I am modelling the the sequence $\{(a_t,y_t)\}_t$ as follows: $$ \begin{cases} Y_{t+1} &= g_\nu(X_{t+1}) + \alpha V_{t+1}\\ X_{t+1} &= X_t + \mu_\xi(a_t) + \sigma_\psi(a_t)Z_{t+1}\\ X_0 &= ...
Uomond's user avatar
  • 51
3 votes
1 answer
129 views

I would like to perform clustering with a finite Gaussian Mixture model, however, I have missing data (some features are missing at random). I am using Variational Inference to fit my Bayesian GMM. Is ...
Tom's user avatar
  • 1,112
0 votes
0 answers
28 views

I’ve been working on implementing a binary variant of probabilistic PCA (PPCA) in Python (based on this paper), which uses variational EM for parameter estimation due to the non-conjugacy between the ...
Net_Raider's user avatar
0 votes
0 answers
128 views

I am following the Zuko "Train From Data" tutorial to train a Neural Spline Flow. My goal is to approximate a distribution over functions. Therefore, each of my function samples are actually ...
James's user avatar
  • 1
1 vote
1 answer
85 views

As mentioned in the title, I understand the mathematical derivation of equations (6-7) in Kingma's original paper. \begin{equation} \log p_\theta(\mathbf{x}, y) \geq \mathbb{E}_{q_\phi(\mathbf{z} \mid ...
Wang Jing's user avatar
2 votes
1 answer
115 views

Using normalizing flows, we can model model's posteriors $p(\theta|D)$, by feeding Gaussian noise $z$ to the NF (parametrized with $\phi$), using the output of the NF $\theta$ as model parameters, and ...
Alberto's user avatar
  • 1,561
2 votes
2 answers
169 views

I'm deeply failing to understand the first step in the ELBO derivation in VAEs. When asking my questions I'll also try to clearly state my assumptions and perhaps some of them are wrong to begin with: ...
DrPrItay's user avatar
  • 121
1 vote
0 answers
92 views

I have an estimation problem where I need to maximize the evidence lower bound: $$ \mathrm{ELBO} = -\frac{1}{2} \Bigg( \mathbb{E}_{q(\theta)} \left[ \mathrm{vec}(\mathbf{Z})^{\mathrm{H}} \mathbf{C}^{-...
CfourPiO's user avatar
  • 325
1 vote
0 answers
71 views

so confused with the derivation of elbo. In part of the derivation p(data) is intractable as it involves an integral over a high dimensional latent variable. I cant understand why the latent ...
user425635's user avatar
3 votes
1 answer
111 views

I am a bit lost with the derivation of ELBO because I dont understand why some distributions are known and some are unknown. I guess we know p(z) (the prior) because it was the last value of q(z) ...
user425635's user avatar
1 vote
0 answers
161 views

On wikipedia it says: "A simple interpretation of the KL divergence of P from Q [i.e. D_KL(P||Q)] is the expected excess surprise from using Q as a model instead of P when the actual distribution ...
profPlum's user avatar
  • 593
0 votes
1 answer
842 views

I recently trained a AE and a VAE and used the latent variables of each for a clustering task. It seemed to work well, sensible clusters. The main reason for training the VAE was too gain more ...
Nathan Thompo's user avatar
2 votes
1 answer
224 views

In Variational Autoencoders (VAE), we have: $$ \log p_\theta(x) = \log \left[ \int p_\theta(x \mid z)p(z) \, dz \right] $$ where $ p_\theta(x \mid z) = \mathcal{N}(x; \mu_\theta(z), I) $ and $ p(z) = \...
rando's user avatar
  • 360
2 votes
1 answer
150 views

I've seen in many tutorials on diffusion models refer to the distribution of the latent variables induced by the forward process as "ground truth". I wonder why. What we can actually see is ...
Daniel Mendoza's user avatar

15 30 50 per page
1
2 3 4 5
7