Skip to main content
deleted 6 characters in body
Source Link
Tim
  • 144k
  • 27
  • 277
  • 536

The problem is that you don't know the probability densities $f(x)$. With the parametric model, you would solve it by finding a parametric distribution for $f$. In the frequentist setting, we would estimate the expected value using the arithmetic average with $\phi(x_i)$$\phi_i$ weights equal to the empirical probabilities

$$ \widehat{E[X]} = \sum_{i=1}^N x_i \, \phi(x_i) $$$$ \widehat{E[X]} = \sum_{i=1}^N x_i \, \phi_i $$

The problem is that you don't know the probability densities $f(x)$. With the parametric model, you would solve it by finding a parametric distribution for $f$. In the frequentist setting, we would estimate the expected value using the arithmetic average with $\phi(x_i)$ weights equal to the empirical probabilities

$$ \widehat{E[X]} = \sum_{i=1}^N x_i \, \phi(x_i) $$

The problem is that you don't know the probability densities $f(x)$. With the parametric model, you would solve it by finding a parametric distribution for $f$. In the frequentist setting, we would estimate the expected value using the arithmetic average with $\phi_i$ weights equal to the empirical probabilities

$$ \widehat{E[X]} = \sum_{i=1}^N x_i \, \phi_i $$

added 380 characters in body
Source Link
Tim
  • 144k
  • 27
  • 277
  • 536

The problem is that you don't know the probability densities $f(x)$. With the parametric model, you would solve it by finding a parametric distribution for $f$. In the frequentist setting, it iswe would estimate the expected value estimated usingusing the arithmetic average. It is in fact a weighted average, where with $\phi(x_i)$ weights areequal to the empirical probabilities.

$$ \widehat{E[X]} = \sum_{i=1}^N x_i \, \phi(x_i) $$

Same as with Bayesian bootstrap, you could useassume a prior for them by drawing them from Dirichlet-uniform distributionprior for the $\phi_i$ weights, sampling the weights, calculating the weighted average, and repeating this many times to find the distribution of the estimates. If you think about it, It's the most trivial case of the Bayesian bootstrap for a weighted statistic. As you can read from the links provided aboveYes, this makes a number of unreasonable assumptions like approximating continuous distribution with a discrete one. Not that it isIt's also not necessarily very useful, but if you insist on a Bayesian nonparametric estimator, that's a possibility. As you can see from the example below, it gives the same results as the frequentist estimator and standard bootstrap.

set.seed(42) N <- 500 X <- rnorm(N, 53, 37) mean(x) ## [1] 51.88829 R <- 5000 mean.boot <- replicate(R, mean(sample(X, replace=TRUE))) summary(mean.boot) ## Min. 1st Qu. Median Mean 3rd Qu. Max. ## 46.41 50.81 51.90 51.91 53.03 57.88 bmean# weighted.mean itself normalizes the weights wmean.boot <- replicate(R, weighted.mean(X, w=rexp(N, 1))) summary(bmeanwmean.boot) ## Min. 1st Qu. Median Mean 3rd Qu. Max. ## 47.09 50.82 51.93 51.91 52.99 57.65 par(mfrow=c(1, 2)) hist(mean.boot, 100, freq=FALSE, main="Standard bootstrap") curve(dnorm(x, mean(X), sd(X)/sqrt(N)), 45, 60, add=T, col="red", lw=2) hist(bmeanwmean.boot, 100, freq=FALSE, main="Sampled weights") curve(dnorm(x, mean(X), sd(X)/sqrt(N)), 45, 60, add=T, col="red", lw=2) 

enter image description hereStandard bootstrap vs the described solution are the same as frequentist estimate for the sampling distribution.

In the end, it's the same as the frequentist estimator, so you could just use the frequentist estimator.

The problem is that you don't know the probability densities $f(x)$. In the frequentist setting, it is estimated using the arithmetic average. It is in fact a weighted average, where weights are empirical probabilities. Same as with Bayesian bootstrap, you could use a prior for them by drawing them from Dirichlet-uniform distribution, calculating the weighted average, and repeating this many times to find the distribution of the estimates. If you think about it, It's the most trivial case of the Bayesian bootstrap for a weighted statistic. As you can read from the links provided above, this makes a number of unreasonable assumptions like approximating continuous distribution with a discrete one. Not that it is very useful, but if you insist on a Bayesian nonparametric estimator, that's a possibility. As you can see from the example below, it gives the same results as the frequentist estimator and standard bootstrap.

set.seed(42) N <- 500 X <- rnorm(N, 53, 37) mean(x) ## [1] 51.88829 R <- 5000 mean.boot <- replicate(R, mean(sample(X, replace=TRUE))) summary(mean.boot) ## Min. 1st Qu. Median Mean 3rd Qu. Max. ## 46.41 50.81 51.90 51.91 53.03 57.88 bmean.boot <- replicate(R, weighted.mean(X, w=rexp(N, 1))) summary(bmean.boot) ## Min. 1st Qu. Median Mean 3rd Qu. Max. ## 47.09 50.82 51.93 51.91 52.99 57.65 par(mfrow=c(1, 2)) hist(mean.boot, 100, freq=FALSE, main="Standard bootstrap") curve(dnorm(x, mean(X), sd(X)/sqrt(N)), 45, 60, add=T, col="red", lw=2) hist(bmean.boot, 100, freq=FALSE, main="Sampled weights") curve(dnorm(x, mean(X), sd(X)/sqrt(N)), 45, 60, add=T, col="red", lw=2) 

enter image description here

The problem is that you don't know the probability densities $f(x)$. With the parametric model, you would solve it by finding a parametric distribution for $f$. In the frequentist setting, we would estimate the expected value using the arithmetic average with $\phi(x_i)$ weights equal to the empirical probabilities

$$ \widehat{E[X]} = \sum_{i=1}^N x_i \, \phi(x_i) $$

Same as with Bayesian bootstrap, you could assume a prior Dirichlet-uniform prior for the $\phi_i$ weights, sampling the weights, calculating the weighted average, and repeating this many times to find the distribution of the estimates. If you think about it, It's the most trivial case of the Bayesian bootstrap for a weighted statistic. Yes, this makes a number of unreasonable assumptions like approximating continuous distribution with a discrete one. It's also not necessarily very useful, but if you insist on a Bayesian nonparametric estimator, that's a possibility. As you can see from the example below, it gives the same results as the frequentist estimator and standard bootstrap.

set.seed(42) N <- 500 X <- rnorm(N, 53, 37) mean(x) ## [1] 51.88829 R <- 5000 mean.boot <- replicate(R, mean(sample(X, replace=TRUE))) summary(mean.boot) ## Min. 1st Qu. Median Mean 3rd Qu. Max. ## 46.41 50.81 51.90 51.91 53.03 57.88 # weighted.mean itself normalizes the weights wmean.boot <- replicate(R, weighted.mean(X, w=rexp(N, 1))) summary(wmean.boot) ## Min. 1st Qu. Median Mean 3rd Qu. Max. ## 47.09 50.82 51.93 51.91 52.99 57.65 par(mfrow=c(1, 2)) hist(mean.boot, 100, freq=FALSE, main="Standard bootstrap") curve(dnorm(x, mean(X), sd(X)/sqrt(N)), 45, 60, add=T, col="red", lw=2) hist(wmean.boot, 100, freq=FALSE, main="Sampled weights") curve(dnorm(x, mean(X), sd(X)/sqrt(N)), 45, 60, add=T, col="red", lw=2) 

Standard bootstrap vs the described solution are the same as frequentist estimate for the sampling distribution.

In the end, it's the same as the frequentist estimator, so you could just use the frequentist estimator.

added 112 characters in body
Source Link
Tim
  • 144k
  • 27
  • 277
  • 536

You seem to be asking about a nonparametric estimator for the mean.

First, let's make it clear: for Bayesian statistics, you always need to make distributional assumptions. You can proceed as suggested by Sextus Empiricus (+1), but this does assume Gaussian distribution. If you really didn't want to make any assumptions, in practice, you would probably just estimate the arithmetic mean.

But let's try coming up with a nonparametric solution. One thing that comes to my mind is Bayesian bootstrap also described by Rasmus Bååth who provided code example. With Bayesian bootstrap, you would assume the Dirichlet-uniform distribution for the probabilities, resample with replacement the datapoints and evaluate the statistic, so arithmetic mean on those samples. But in such a case, to approximate the frequentist estimator, you would use the same estimator on the samples to find their distribution. Not very helpful, isn't it?

Let's start again. The definition of expected value is

$$ E[X] = \int x \, f(x)\, dx $$

The problem is that you don't know the probability densities $f(x)$. In the frequentist setting, it is estimated using the arithmetic averageestimated using the arithmetic average. It is in fact a weighted average, where weights are empirical probabilities. Same as with Bayesian bootstrap, you could use a prior for them by drawing them from Dirichlet-uniform distribution, calculating the weighted average, and repeating this many times to find the distribution of the estimates. If you think about it, It's the most trivial case of the Bayesian bootstrap for a weighted statistic. As you can read from the links provided above, this makes a number of unreasonable assumptions like approximating continuous distribution with a discrete one. Not that it is very useful, but if you insist on a Bayesian nonparametric estimator, that's a possibility. As you can see from the example below, it gives the same results as the frequentist estimator and standard bootstrap.

set.seed(42) N <- 500 X <- rnorm(N, 53, 37) mean(x) ## [1] 51.88829 R <- 5000 mean.boot <- replicate(R, mean(sample(X, replace=TRUE))) summary(mean.boot) ## Min. 1st Qu. Median Mean 3rd Qu. Max. ## 46.41 50.81 51.90 51.91 53.03 57.88 bmean.boot <- replicate(R, weighted.mean(X, w=rexp(N, 1))) summary(bmean.boot) ## Min. 1st Qu. Median Mean 3rd Qu. Max. ## 47.09 50.82 51.93 51.91 52.99 57.65 par(mfrow=c(1, 2)) hist(mean.boot, 100, freq=FALSE, main="Standard bootstrap") curve(dnorm(x, mean(X), sd(X)/sqrt(N)), 45, 60, add=T, col="red", lw=2) hist(bmean.boot, 100, freq=FALSE, main="Sampled weights") curve(dnorm(x, mean(X), sd(X)/sqrt(N)), 45, 60, add=T, col="red", lw=2) 

enter image description here

You seem to be asking about a nonparametric estimator for the mean.

First, let's make it clear: for Bayesian statistics, you always need to make distributional assumptions. You can proceed as suggested by Sextus Empiricus (+1), but this does assume Gaussian distribution. If you really didn't want to make any assumptions, in practice, you would probably just estimate the arithmetic mean.

But let's try coming up with a nonparametric solution. One thing that comes to my mind is Bayesian bootstrap also described by Rasmus Bååth who provided code example. With Bayesian bootstrap, you would assume the Dirichlet-uniform distribution for the probabilities, resample with replacement the datapoints and evaluate the statistic, so arithmetic mean on those samples. But in such a case, to approximate the frequentist estimator, you would use the same estimator on the samples to find their distribution. Not very helpful, isn't it?

Let's start again. The definition of expected value is

$$ E[X] = \int x \, f(x)\, dx $$

The problem is that you don't know the probability densities $f(x)$. In the frequentist setting, it is estimated using the arithmetic average. It is in fact a weighted average, where weights are empirical probabilities. Same as with Bayesian bootstrap, you could use a prior for them by drawing them from Dirichlet-uniform distribution, calculating the weighted average, and repeating this many times to find the distribution of the estimates. If you think about it, It's the most trivial case of the Bayesian bootstrap for a weighted statistic. As you can read from the links provided above, this makes a number of unreasonable assumptions like approximating continuous distribution with a discrete one. Not that it is very useful, but if you insist on a Bayesian nonparametric estimator, that's a possibility. As you can see from the example below, it gives the same results as the frequentist estimator and standard bootstrap.

set.seed(42) N <- 500 X <- rnorm(N, 53, 37) mean(x) ## [1] 51.88829 R <- 5000 mean.boot <- replicate(R, mean(sample(X, replace=TRUE))) summary(mean.boot) ## Min. 1st Qu. Median Mean 3rd Qu. Max. ## 46.41 50.81 51.90 51.91 53.03 57.88 bmean.boot <- replicate(R, weighted.mean(X, w=rexp(N, 1))) summary(bmean.boot) ## Min. 1st Qu. Median Mean 3rd Qu. Max. ## 47.09 50.82 51.93 51.91 52.99 57.65 par(mfrow=c(1, 2)) hist(mean.boot, 100, freq=FALSE, main="Standard bootstrap") curve(dnorm(x, mean(X), sd(X)/sqrt(N)), 45, 60, add=T, col="red", lw=2) hist(bmean.boot, 100, freq=FALSE, main="Sampled weights") curve(dnorm(x, mean(X), sd(X)/sqrt(N)), 45, 60, add=T, col="red", lw=2) 

enter image description here

You seem to be asking about a nonparametric estimator for the mean.

First, let's make it clear: for Bayesian statistics, you always need to make distributional assumptions. You can proceed as suggested by Sextus Empiricus (+1), but this does assume Gaussian distribution. If you really didn't want to make any assumptions, in practice, you would probably just estimate the arithmetic mean.

But let's try coming up with a nonparametric solution. One thing that comes to my mind is Bayesian bootstrap also described by Rasmus Bååth who provided code example. With Bayesian bootstrap, you would assume the Dirichlet-uniform distribution for the probabilities, resample with replacement the datapoints and evaluate the statistic, so arithmetic mean on those samples. But in such a case, to approximate the frequentist estimator, you would use the same estimator on the samples to find their distribution. Not very helpful, isn't it?

Let's start again. The definition of expected value is

$$ E[X] = \int x \, f(x)\, dx $$

The problem is that you don't know the probability densities $f(x)$. In the frequentist setting, it is estimated using the arithmetic average. It is in fact a weighted average, where weights are empirical probabilities. Same as with Bayesian bootstrap, you could use a prior for them by drawing them from Dirichlet-uniform distribution, calculating the weighted average, and repeating this many times to find the distribution of the estimates. If you think about it, It's the most trivial case of the Bayesian bootstrap for a weighted statistic. As you can read from the links provided above, this makes a number of unreasonable assumptions like approximating continuous distribution with a discrete one. Not that it is very useful, but if you insist on a Bayesian nonparametric estimator, that's a possibility. As you can see from the example below, it gives the same results as the frequentist estimator and standard bootstrap.

set.seed(42) N <- 500 X <- rnorm(N, 53, 37) mean(x) ## [1] 51.88829 R <- 5000 mean.boot <- replicate(R, mean(sample(X, replace=TRUE))) summary(mean.boot) ## Min. 1st Qu. Median Mean 3rd Qu. Max. ## 46.41 50.81 51.90 51.91 53.03 57.88 bmean.boot <- replicate(R, weighted.mean(X, w=rexp(N, 1))) summary(bmean.boot) ## Min. 1st Qu. Median Mean 3rd Qu. Max. ## 47.09 50.82 51.93 51.91 52.99 57.65 par(mfrow=c(1, 2)) hist(mean.boot, 100, freq=FALSE, main="Standard bootstrap") curve(dnorm(x, mean(X), sd(X)/sqrt(N)), 45, 60, add=T, col="red", lw=2) hist(bmean.boot, 100, freq=FALSE, main="Sampled weights") curve(dnorm(x, mean(X), sd(X)/sqrt(N)), 45, 60, add=T, col="red", lw=2) 

enter image description here

added 989 characters in body
Source Link
Tim
  • 144k
  • 27
  • 277
  • 536
Loading
added 989 characters in body
Source Link
Tim
  • 144k
  • 27
  • 277
  • 536
Loading
Post Undeleted by Tim
Post Deleted by Tim
added 176 characters in body
Source Link
Tim
  • 144k
  • 27
  • 277
  • 536
Loading
Source Link
Tim
  • 144k
  • 27
  • 277
  • 536
Loading