Let $F_x$ be the marginal cumulative distribution function (CDF) of $x$ and let $F_{y\mid x}$ be the conditional CDF. The independence of the data (with an implicit assumption that $F_x$ is continuous) implies the CDF of $x_j$ is
$$\Pr(x_j \le x) = 1 - \left(1 - F_x(x)\right)^n.$$
Consequently the CDF of $y_j$ is obtained by averaging over $x_j,$
$$\Pr(y_j \le y) = \int_{\mathbb R} F_{y\mid x}(y)\, \mathrm d \left[1 - \left(1 - F_x(x)\right)^n\right].$$
For example, let $(x,y)$ have a Binormal distribution. Choose units of measurement for $x$ in which $F_x$ is standard Normal, $F_x = \Phi.$ Then the regression of $y$ on $x$ is linear with regression function $\hat y = \alpha + \beta x$ for parameters $\alpha$ and $\beta,$ whence (writing $\sigma^2$ for the conditional variance)
$$F_{x\mid y}(y) = \Phi\left(\frac{y - (\alpha + \beta x)}{\sigma}\right)$$
giving (with $\varphi = \Phi^\prime$ the standard normal density function)
$$\Pr(y_j \le y) = n\int_{\mathbb R} \Phi\left(\frac{y - (\alpha + \beta x)}{\sigma}\right)\left(1 - \Phi(x)\right)^{n-1}\varphi(x)\,\mathrm d x.\tag{*}$$
I believe this doesn't simplify (except for tiny values of $n$), but it is amenable to numerical integration for efficient evaluation (as shown below in the R code for the function f).
A quick simulation (in R) supports this formula. The histogram plots the empirical density of $10^5$ values of $y_j$ for $n=30,$ $\alpha = 0,$ $\beta = 1,$ and $\sigma^2 = 2/3$ while the red curve plots the derivative of $(*),$ the conditional density. They agree to within random variation.

BTW, don't let this fool you into thinking the distribution of $y_j$ is Normal or even approximately so. Although in many cases it will be (small values of $|\beta|,$ $n,$ and $\sigma$ are conducive to this appearance), when $y$ and $x$ are strongly correlated and $n$ is sufficiently large, the skewness in the distribution of the minimum of the $x_i$ will become apparent. For instance, changing $n$ to $300$ and $\beta$ to $-5$ yields this obviously skewed histogram:

n <- 30 alpha <- 0 beta <- 1 sigma <- sqrt(2/3) # # Simulation. # set.seed(17) n.sim <- 1e5 x <- apply(matrix(rnorm(n * n.sim), n), 2, min) y <- beta * x + alpha + rnorm(length(x), 0, sigma) # # Plot (a subset of) the (x, y_j) data. # j <- seq_len(min(n.sim, 1e3)) plot(x[j], y[j]) # # Integration. # f <- function(x, y, n, alpha, beta, sigma) { exp(pnorm(y, alpha + beta * x, sigma, log = TRUE) + pnorm(x, log = TRUE, lower.tail = FALSE) * (n - 1) + dnorm(x, log = TRUE) + log(n)) } f.yx <- Vectorize(function(y, ...) { integrate(\(x) f(x, y, n, alpha, beta, sigma), lower = -Inf, upper = Inf, ...)$value }) # # Numerical differentiation. # Y <- seq(min(y), max(y), length.out = 101) Z <- f.yx(Y, rel.tol = 1e-12) # (Can benefit from high precision) dZ.dY <- diff(Z) / diff(Y) y0 <- (head(Y, -1) + tail(Y, -1)) / 2 # # Plot the results. # hist(y, freq = FALSE, ylim = c(0, max(dZ.dY)), main = expression("Histogram of " * y[j]), xlab = "Value", breaks = 50) lines(y0, dZ.dY, col = "Red", lwd = 2)