Theorem 1 (Existence of Coins): If $X=(x_i)_{i∈ℕ}$ is a countable set of infinite binary sequences, then there exists an infinite binary sequence relatively random to $X$.
Proof: Choose any decreasing real sequence $ε_n→0$. We shall construct an infinite binary sequence $α$ such that for some sequence $N_n$ we have $|δ_i(α,x_j)-\dfrac{1}{2}| < ε_n$ for all $i ≥ N_n$ and $j ≤ n$. Hence as $i→∞$ we shall have $δ_i(α,x_j)→\dfrac{1}{2}$ for arbitrary $j$, implying $δ(α,x_j)=\dfrac{1}{2}$.
We shall attain this by constructing a sequence of finite binary sequences $α_{N_m} : \{1,...,N_m\}→\{0,1\}$ that are pairwise extensions of one another, and such that $|δ_i(α_{N_m}^*,x_j)-\dfrac{1}{2}| < ε_j$ for all free variables subject to $j≤m$ and $N_m+2^{m+1}≥i≥N_j$, and for all $α_{N_m}^* : \{1,...,N_m+2^{m+1}\}→\{0,1\}$ extending $α_{N_m}$. Then the sequence $α$ defined by $α(n) = α_{N_m}(n)$ for any $N_m≥n$ will satisfy the hypothesis above.
This shall be obtained by induction on $m$. (The base case will become superfluous, but is useful for comprehension) For $m=1$, let $β_1(k)=Φ_1(k)$ where $Φ_1(k)≡|\{j~|~j < k, x_1(j)=x_1(k)\}| \mod 2$, and then since we see $δ(β_1,x_1) = \dfrac{1}{2}$, choose $N_1$ large enough so that $$|\dfrac{(\sum_{j=1}^i γ_j)+4}{i+4} - \dfrac{1}{2}| < ε_1$$ $$|\dfrac{\sum_{j=1}^i γ_j}{i+4} - \dfrac{1}{2}| < ε_1$$ for $i ≥ N_1$, where $γ_i$ is the comparison between $β_1$ and $x_1$. Then the restriction $α_{N_1}$ of $β_1$ onto the set $\{1,...,N_1\}$ satisfies the hypothesis for $m=1$.
Now suppose appropriate $α_{N_m}$ has been constructed. Define $β_{m+1}$ by $β_{m+1}(k)=α_{N_m}(k)$ for $k≤N_m$, and otherwise, $β_{m+1}(k)=Φ_{m+1}(k)$ where $$Φ_{m+1}(k)≡|\{j~|~j < k, θ(j)=θ(k)\}| \mod 2$$ $$θ(n)=[x_i(n)]_{1≤i≤m+1}$$ Notice, then, that for any $1≤i≤m+1$ and any $(m+1)×1$ binary vector $v$, the proportion of $k > N_m$ such that $θ(k)=v$ satisfying $x_i(k)=β_{m+1}(k)$ as opposed to $x_i(k)≠β_{m+1}(k)$ approaches $\dfrac{1}{2}$, with at most $1$ extra nonsatisfying $k$ given any of the limit populations observed. Since for each $k > N_m$ there is one such vector $v$ for which $θ(k)=v$, and since there are $2^{m+1}$ vectors overall, for all $1≤i≤m+1$ we observe the proportion of such $k > N_m$ for which $x_i(k)=β_i(k)$ as opposed to $x_i(k)≠β_i(k)$ approaches $\dfrac{1}{2}$, with at most $2^{m+1}$ extra nonsatisfying $k$ given any of the limit populations observed. Therefore, there is sufficiently large $N_{m+1}$ for which $|δ_i(β_{m+1},x_j)-\dfrac{1}{2}| < ε_{m+1}$ for all free variables subject to $j≤m+1$ and $i≥N_j$; all that is missing is the $α_{N_{m+1}}^*$, and the variable index of $ε$. To fix the first, we rechoose $N_{m+1}$ to satisfy $$|\dfrac{(\sum_{j=1}^i γ_j)+2^{m+2}}{i+2^{m+2}} - \dfrac{1}{2}| < ε_{m+1}$$ $$|\dfrac{\sum_{j=1}^i γ_j}{i+2^{m+2}} - \dfrac{1}{2}| < {m+1}$$ for all $i≥m$ where $γ_j$ ranges being relative over $α_{N_1},...,α_{N_{m+1}}$ toward $β_{m+1}$, and then let $α_{m+1}$ be the restriction of $β_{m+1}$ onto $\{1,...,N_{m+1}\}$. Now, observing the upper bound of $2^{m+2}$ on the lag of the previously mentioned proportions of the limit populations in conjunction with the conditions just now qualified to $N_{m+1}$, we may finally conclude $|δ_i(α_{N_{m+1}}^*,x_j)-\dfrac{1}{2}| < ε_j$ for all free variables subject to $j≤m+1$ and $N_{m+1}+2^{m+2}≥i≥N_j$, and for all $α_{N_{m+1}}^* : \{1,...,N_{m+1}+2^{m+2}\}→\{0,1\}$ extending $α_{N_{m+1}}$. This completes the induction, and the proof.$~\square$
With some further complication of the frequency by which we choose $1$ or $0$ given a finite binary pattern vector (i.e. changing $Φ$ so that it isn't merely alternating), the above proof should generalize to show the existence of arbitrarily weighted coins, yielding heads at any given probability in the unit interval.
Given an infinite binary sequence $β$ and $α : ℕ→[0,1]$, define $$δ_n = \dfrac{\sum_{i=1}^n {γ_i}}{n}$$ where $γ_i$ is the comparison of $α$ and $β$, i.e. $γ_i=1-α(i)$ if $β(i)=0$ and $γ_i=α(i)$ otherwise. If $δ_n→z$ for some $z∈[0,1]$, then we write $δ(α,β)=z$. If $z=\dfrac{1}{2}$ in this case, we say $β$ is random relative to $α$. If $A$ is a set of infinite binary sequences, then we say $β$ is random relative to $A$ if $δ(α,β)=\dfrac{1}{2}$ for all $α∈A$.
Theorem 2: If $X$ is a countable set of functions $x_i : ℕ→[0,1]$, then there exists an infinite binary sequence relatively random to $X$.
Proof (sketch): If we define $δ_n'=δ_n-\dfrac{1}{2}=\dfrac{\sum_{i=1}^n {(γ_i-\dfrac{1}{2})}}{n}$ and $δ'=\lim_{n→∞} δ_n'$, then an equivalent reformulation of the problem for the case when $A$ is of finite order $m$ becomes: "Given a sequence $(v_i)$ of vectors of the form $(w_1,...,w_m)$ where $w_i∈[-1,1]$, does there exist a sequence $β=(b_i)$ where $b_i = \pm 1$ such that when $σ_j = \sum^j b_i·v_i$, we observe $\lim_{n→∞} \dfrac{σ_n}{n} = 0$?" As well, having solved this finite case, the countably infinite case will follow in a cumbersome manner as above and will thus be omitted.
Choose a decreasing real sequence $ε_n→0$. For each $n$, let $B_n$ be a partition of $[-1,1]^m$ into finitely many regions (say $\omega_n$) of diameter less than $2ε_n$. We shall inductively construct $β$ so that for sufficiently large indices $n > N_k$, we shall have $σ_n < ε_k·n$. As such, starting with $ε_i$, follow this procedure for as long as it takes for $σ_n$ to be able to incur $\omega_{i+1}$ consecutive "errors" and still be less than $ε_i·n$: At the start, all regions of $B_i$ are "inactive." Given the choice to set $b_j = \pm 1$, if $v_j$ is located in a region of $[-1,1]^m$ that is active, then set $b_j = -1$ and deactivate that region. Otherwise, activate the region and set $b_j=1$. Move on to $v_{j+1},b_{j+1}$. This process is seen to cancel vectors in partial sums with a per-step error factor approaching less than $\dfrac{1}{2}2ε_i=ε_i$ in magnitude.$~\square$
As with the first theorem, with some complication of the region that is to be finitely partitioned, the proof should generalize to show the existence of arbitrarily weighted coins.
Fix an infinite binary sequence $θ$. For infinite binary sequences $α$ and $β$, define $P(α)=δ(α,θ)$ and $P(α~|~β)=δ(γ,θ)$ where $γ(n)=α(b_n)$, if $b_n$ is the $n^\text{th}$ smallest integer to satisfy $β(b_n)=θ(b_n)$.
Theorem 3 (Bayes' Theorem): Suppose $P(α)$ and $P(β)$ exist and are nonzero. Then $$P(α~|~β)·P(β) = P(β~|~α)·P(α)$$ where if one side is undefined, then so is the other.
Proof: Suppose without loss that $P(β~|~α)$ is well defined. Let $γ_α$ denote its helper function, so that $δ_m(γ_α,θ)$ is the fraction of values $a_n ≤ m$ such that $β(a_n)=θ(a_n)$. This latter observation leads us to the identity $$δ_n(γ_α,θ)·(n·δ_n(α,θ))=δ_n(γ_β,θ)·(n·δ_n(β,θ))$$ when $γ_β$ is the associated helper function for $P(α~|~β)$. Therefore we derive $$P(α~|~β)·P(β) = \lim_{n→∞} δ_n(γ_β,θ) · \lim_{n→∞} δ_n(β,θ) = \lim_{n→∞} δ_n(γ_β,θ) · δ_n(β,θ) = $$$$\lim_{n→∞} \dfrac{(n·δ_n(γ_β,θ)) · δ_n(β,θ)}{n} = \lim_{n→∞} \dfrac{(n·δ_n(γ_α,θ)) · δ_n(α,θ)}{n} = $$$$\lim_{n→∞} δ_n(γ_α,θ) · \lim_{n→∞} δ_n(α,θ) = P(β~|~α)·P(α)~~\square$$