How exactly are the beta and gamma distributions related?

This concerns the relationship between the Gamma and Beta distributions as opposed to the Gamma and Beta functions. Let $X \sim \mbox{Gamma}(\alpha, 1)$ and $Y \sim \mbox{Gamma}(\beta, 1)$ where the paramaterization is such that $\alpha$ is the shape parameter. Then
$$ \frac{X}{X + Y} \sim \mbox{Beta}(\alpha, \beta). $$

To prove this, write the joint pdf $f_{X, Y} (x, y) = \frac{1}{\Gamma(\alpha) \Gamma(\beta)} x^{\alpha - 1} y^{\beta - 1} e^{-(x + y)}$ (on $\mathbb R^2_+$) and make the transformation $U = \frac{X}{X + Y}$ and $V = X + Y$. The Jacobian of the transformation $X = VU, Y = V(1 - U)$ is equal to $V$ so the joint distribution of $U$ and $V$ has pdf $$ \frac{v}{\Gamma(\alpha)\Gamma(\beta)} (vu)^{\alpha - 1} (v (1 - u))^{\beta - 1} e^{-v} = \frac{1}{\Gamma(\alpha)\Gamma(\beta)}v^{\alpha + \beta - 1} e^{-v} u^{\alpha - 1} (1 - u)^{\beta - 1} $$ (on $\mathbb R_+ \times [0, 1]$) and hence $U$ and $V$ are independent (because the pdf factors over $u$ and $v$) with $V \sim \mbox{Gamma}(\alpha + \beta, 1)$ and $U \sim \mbox{Beta}(\alpha, \beta)$ which is apparent from the terms $v^{\alpha + \beta - 1} e^{-v}$ and $u^{\alpha - 1}(1 - u)^{\beta - 1}$ respectively.


Fix some $k$ and, for every $n$, let $X_n$ denote a random variable with beta distribution $\mathrm B(k,n)$ and $Y_n=nX_n$. Then, for every $s\geqslant0$, $\mathrm E(Y_n^s)=n^s\mathrm E(X_n^s)$ and one knows the value of $\mathrm E(X_n^s)$, hence $$ \mathrm E(Y_n^s)=n^s\frac{\mathrm B(k+s,n)}{\mathrm B(k,n)}=n^s\frac{\Gamma(k+s)\Gamma(k+n)}{\Gamma(k+s+n)\Gamma(k)}\longrightarrow\frac{\Gamma(k+s)}{\Gamma(k)}. $$ This is $\mathrm E(Z^s)$ for any random variable $Z$ with gamma distribution $\Gamma(k)$ hence $Y_n\to Z$ in distribution.

Let $X'_n$ denote a random variable with beta distribution $\mathrm B(k,n/b)$, and $Y'_n=nX'_n$. Then, $X'_n$ is distributed like $X_{n/b}$ hence $Y'_n$ is distributed like $bY_{n/b}$ and $Y'_n\to bZ$ in distribution.


There is another way to view the relationship between the gamma distribution and the beta distribution through the Dirichlet distribution. This post (https://math.stackexchange.com/q/190695) talks about exactly how they are related without the Dirichlet distribution, but here is a slightly broader view:

Let $Z_1, Z_2, \ldots, Z_n$ be independent random variables such that $$ Z_i \sim \text{Gamma}(\alpha_i,1), \quad i=1,\ldots, n$$ where $\alpha_i \geq 0$ are the shape parameters. If $Y_j$ is defined as follows, $$Y_j = Z_j/\sum_{i=1}^n Z_i, \quad j=1, \ldots, n$$ then $(Y_1, \ldots, Y_n)$ is Dirichlet distributed with parameter $(\alpha_1, \ldots, \alpha_n)$. When $n=2$, the Dirichlet distribution reduces to the Beta distribution, denoted by $\text{Beta}(\alpha_1, \alpha_2)$.

Here is a link to Ferguson's paper that mentions the above relationship: http://www.cis.upenn.edu/~taskar/courses/cis700-sp08/papers/ferguson.pdf