\(\leftarrow\) Back to Chapters
“We verify that \(\partial{V}/\partial{y} = \partial{U}/\partial{z}\)”:
By definition \(U=G(x, v) F_1 (y, z)\), so by product rule and chain rule and using that \(v=F(y, z)\) we have
\[U_z=G_2(x,v)F_2(y,z)F_1(y,z)+ G(x,v)F_{21}(y,z).\]
Similarly \(V=G(x,v)F_2(y,z)\) so
\[V_y=G_2(x,v)F_1(y,z)F_2(y,z)+G(x,v)F_{12}(y,z).\]
Of course mixed partials are equal (under some continuity of second derivatives assumption, Clairaut’s theorem), so the two expressions are the same.
Continuing to get 2.19 from 2.18:
We have then \(G(x,y)G(y,z)=P(x,z)\). Pick any fixed \(z\). Denote \(P(x,z)=A(x)\) and \(G(y,z)=B(y)\). Then \(G(x,y)=\frac{A(x)}{B(y)}\) [and \(G(y,z)=\frac{A(y)}{B(z)}\)].
Plug this in to \(G(x,y)G(y,z)=P(x,z)\) to get \(\frac{A(x)A(y)}{B(y)B(z)}=P(x,z)\).
So \(A(y)/B(y)\) is independent of \(y\), so is constant equal to \(r\). This means
\(G(x,y)=\frac{A(x)}{B(y)}=\frac{A(x)A(y)}{A(y)B(y)}=r \frac{A(x)}{A(y)}\)
The variables \(v, y, z\) are related by \(v=F(y,z)\). One can interpret 2.22 as equality of differential 1-forms on the surface \(\Sigma=\{v,y,z|v=F(y,z)\}\) in 3D space with coordinates \(v-y-z\). The forms \(\frac{v}{H(v)}\), \(\frac{y}{H(y)}\) and \(\frac{z}{H(z)}\) are exact, meaning are differentials of functions. We can find these functions by single-variable integration because each of the 1-forms depends only on one of the variables (in forms language, is pulled back via coordinate projection from the corresponding coordinate line), meaning we can find the antiderivative of the corresponding 1-form on the coordinate line, and then “pull back” the result, which just means interpret it as a function on \(\Sigma\). Namely if we set
\[f(x)=\int_{x_0}^x \frac{1}{H(t)}dt\]
be a function well-defined up to a constant, we have
\[df(v)=\frac{dv}{H(v)}, \text{ } df(y)=\frac{dy}{H(y)} ,\text{ } df(z)=\frac{dz}{H(z)}.\]
Then, since the antiderivative of a 1-form on a connected \(\Sigma\) is well-defined up to a constant, equality of 1-forms
\[ \frac{dv}{H(v)}=\frac{dy}{H(y)}+r\frac{dz}{H(z)}\]
implies equality of antiderivatives up to an additive constant, i.e.
\[f(v)=f(y)+rf(z).\]
Exponentiating this we get, on the surface \(v=F(y, z)\) the equality, up to a multiplicative constant,
\[w(v)=w(y)w(z)^r\]
or
\[w(F(y,z))=w(y)w(z)^r.\]
TODO
Note: There is a typo in the book just above equation 2.45 on pg 31. (2.25) should be (2.40).
The domain is \(0\leq S(y)\leq x\) because \(S(y)\) is plausibility of \(\bar{B}=AD\), which is at most plausibility of \(A\), aka \(x\). for general \(A,D\) these (and \(x,y\in [0,1]\)) are the only requirements, anything else should be possible, so hence the domain).
The symmetry of the domain comes from \(S\) being self-inverse and monotone decreasing. In fact, by monotonicity we have \(S(y)\leq x \Leftrightarrow S(S(y))\geq S(x)\) and by \(SS=Id\) (Eq. 2.46) we have \(S(S(y))\geq S(x) \Leftrightarrow y\geq S(x)\).
(In general the graph of the inverse function is obtained by flipping the graph of the original function through the \(x=y\) line, and so the graph of \(S\) is symmetric to this flip preciely when \(S\) is it’s own inverse; monotonicity makes the same true for the overgraph region.)
We can slightly rewrite the above argument as: \(y\geq S(x)\) means \(w(\overline{AD})\geq w(\bar{A})\), which by monotonicity of \(S\) means \(w(AD)\geq w(A)\), i.e. \(x\geq S(y)\).
Note: There’s a missing bracket in 2.49 before the second =.
Source: stackexchange, I’ve reworded it and added detail to (hopefully) make it clearer.
We will use the Taylor series approximation, which is an approximation of \(f(t)\) around the point \(a\):
\[f(t) = f(a) + f'(a)(t-a) + O((t-a)^2)\]
Big O notation is described on Wikipedia.
The proof:
Letting \(\delta = e^{-q}\), we have from (2.48):
\[S(y) = S \left[ \frac{S(x)}{1-\delta}\right]\]
We then use a Taylor series approximation of the function \(f(\delta) = \frac{1}{1-\delta}\) around with \(a = 0\).
\[S(y) = S[S(x)(1+\delta + O(\delta^2))]\]
\[S(y) = S[S(x) + S(x)\delta + S(x)O(\delta^2)]\]
Now we want to get rid of the \(S[]\) surrounding the equation, so we will use another Taylor approximation of the function \(S(t)\). We approximate around the point \(a=S(x)\).
This gives us the approximation of \(S(t)\) as:
\[S(t) = S[S(x)]+S'[S(x)](t - S(x)) + O((t - S(x))^2)\]
Letting \(t=S(x) + S(x)\delta + S(x)O(\delta^2)\)
\[S[S(x) + S(x)\delta + S(x)O(\delta^2)] = S[S(x)]+S'[S(x)](S(x)\delta + S(x)O(\delta^2)) + O((S(x)\delta+S(x)O(\delta^2))^2)\]
\[S[S(x) + S(x)\delta + S(x)O(\delta^2)] = S[S(x)]+S'[S(x)]S(x)\delta + S'[S(x)]S(x)O(\delta^2) + O((S(x)\delta+S(x)O(\delta^2))^2)\]
With big O notation we can get rid of constant factors:
\[S[S(x) + S(x)\delta + S(x)O(\delta^2)] = S[S(x)]+S'[S(x)]S(x)\delta + O(\delta^2) + O((\delta+O(\delta^2))^2)\]
With big O notation we can also get rid terms that drop asymptotically faster than the largest term.
\[S[S(x) + S(x)\delta + S(x)O(\delta^2)] = S[S(x)]+S'[S(x)]S(x)\delta + O(\delta^2)\]
2.45 says \(S[S(x)]=x\). Differentiating in \(x\) we get \(S'[S(x)]S'(x)=1\), or \(S'[S(x)]=1/S'(x)\). Now we plug in into 2.50 to get
\[S(y)=x+\exp\{-q\} S(x)/S'(x)+O(\exp\{-2q\}) \]
Denoting by \(\alpha(x)=\log \left[\frac{-xS'(x)}{S(x)}\right]\) we get
\[S(y)=x+\exp\{-q\} (-x) \exp\{-\alpha\}+O(\exp\{-2q\}) \]
Dividing by \(x\)
\[\frac{S(y)}{x}=1-\exp\{-(q+\alpha)\}+\frac{1}{x} O(\exp\{-2q\}) \]
which is a version of 2.51.
From now on we will treat \(x\) as fixed and only vary \(q\), sending it to \(+\infty\), which in light of 2.48 means keeping \(x\) fixed and sending \(y\) to \(S(x)\) from below.
Then we can write
\[\frac{S(y)}{x}=1-\exp\{-(q+\alpha)\}+ O(\exp\{-2q\}),\]
which is 2.51.
Now we want to deduce 2.53. We make some progress but ultimately do not succeed, deriving a weaker statement, sufficient for continuing.
We start with 2.45
\[x S\left[\frac{S(y)}{x}\right]=y S\left[\frac{S(x)}{y}\right]\]
and plug in 2.51 and 2.48 to get
\[x S[1-\exp\{-(q+\alpha)\}+ O(\exp\{-2q\})]=y S[1-\exp\{-q\}] \]
Right hand side is \(y \exp\{-J(q)\}\) by definition 2.49. We also plug in 2.48 in the form \(y=S(x)/(1-\exp\{-q\})\) to get
\[RHS=S(x) \exp\{-J(q)\} /(1-\exp\{-q\})\]
Now take log of both sides to get
\[\log x +\log S[1-\exp\{-(q+\alpha)\}+ O(\exp\{-2q\})]\]
\[ =\log S(x)-J(q)-\log(1-\exp\{q\})\]
Now if we could write
\[\log S[1-\exp\{-(q+\alpha)\}+ O(\exp\{-2q\})]=\] \[\log S[1-\exp\{-(q+\alpha)\}]+ O(\exp\{-2q\})\]
we would get \(J(q+\alpha)+O(\exp\{-2q\})\) and 2.53 would follow. We don’t get that, but we get almost the same thing - just without the 2 in the last exponent.
We use the following fact: If \(f(q)=O(g(q))\) then, for any eventually non-zero \(h(q)\) one has \(f(q)=h(q) O(g(q)/h(q))\). This is simply from definition: both mean \(\lim_{q\to something} \frac{g(q)}{f(q)}=\lim_{q\to something} \frac{g(q)/h(q)}{f(q)/h(q)}=0\)
So, since \(q\) is the variable and so single \(\exp{\alpha}\) is a constant and can be absorbed into any \(O(g(q))\), we write
\[\exp\{-(q+\alpha)\}+ O(\exp\{-2q\})=\]
\[\exp\{-(q+\alpha)\}+ (\exp\{-(q+\alpha\} O(\exp\{-(q-\alpha\}) \]
\[\exp\{-(q+\alpha)\} \left[1+O(\exp\{-q\})\}\right].\]
Define \(T(s)= S(1-s)\). Then \(T(0)=0\) and Taylor expanding \(T\) around \(0\) we have, as \(s\to 0\)
\[T(s)=T'(0)s+O(s^2)=T'(0)s(1+O(s)).\]
So, as \(t\to -\infty\)
\[\log T(\exp t)=\log T'(0)+ t+ \log(1+O(\exp \{t\})\]
\[=\log T'(0)+ t+ O(\exp \{t\}) \]
Let
\[t_1=-(q+\alpha)\]
and
\[t_2 =\ln [ \exp\{-(q+\alpha)\}+ O(\exp\{-2q\})]=-(q+\alpha)+\ln(1-\exp\{-q\})\]
\[=-(q+\alpha)+O(\exp\{-q\}) \]
We plug into \(\log T'(0)+ t+ O(\exp \{t\})\) and see that the difference (as \(q\to \infty\)) is \(O(-\exp\{-q\})\). That is,
\[\log S[1-\exp\{-(q+\alpha)\}+ O(\exp\{-2q\})]=\] \[\log S[1-\exp\{-(q+\alpha)\}]+ O(\exp\{-q\})\]
so that
\[J(q+\alpha(x))-J(q)= \log \left[\frac{x}{S(x)}\right]+\log (1-\exp\{-q\}) +O( \exp\{-q\})\]
which is not quite 2.53 but is good enough for deducing 2.54 (see below).
The difficulty is getting 2.54 from 2.53. More precisely, we only need that
\[b=\alpha^{-1} \log\left[\frac{x}{S(x)}\right]= \log \left[ \frac{x}{S(x)} \right]/ \log \left[\frac{-xS'(x)}{S(x)}\right]\]
is constant. After that it’s algebra:
\[b \log \left[\frac{-xS'(x)}{S(x)}\right]=\log \left[ \frac{x}{S(x)} \right]\]
and exponentiating one gets
\[\left[ \frac{-xS'(x)}{S(x)} \right]^b = \left[ \frac{x}{S(x)} \right]\]
i.e. 2.56.
So we just need to show that \(b(x)\) is constant.
I find it simpler to do “directly”, rather than to show the asymptotic expansion 2.54.
Remark: To get 2.54 one must first make sure that \(\alpha\) actually takes “continuum of values”. Since \(\alpha\) is a continuous function of \(x\), intermediate value theorem implies that the set of values of \(\alpha\) is an interval; we just check that it is not a degenerate interval consisting of a single point. Indeed, that would mean \(\alpha(x)\) is constant, or \(S'(x)/S(x)=-c/x\), \((\ln S)'=-c/x\), \(\ln S(x)= a-c/x\), but this breaks \(S(0)=1\). Now we do know that \(\alpha\) takes ’’continuum of values". We will use this as well.
We start with 2.53 in the form
\[J(q+\alpha(x))-J(q)= \beta(x)+ O( \exp\{-q\})\]
We want to deduce that
\(b(x)=\beta(x)/\alpha(x)\) is constant.
Intuitively, \(J(q+\alpha(x))-J(q)= \beta(x)+ O(\exp\{- q\})\) does say that for every increment of \(\alpha(x)\) in the input, the output of \(J\) increases by \(\beta(x)\) (plus a small error), so (asymptotically) \(J\) must be linear with slope=rise/run=\(b(x),\) and since there can be only one slope, \(b(x)\) must be constant. The question is how to make it precise.
First, if the error term was absent
\[J(q + \alpha(x)) - J (q) = \beta(x)\]
implies that if \(\alpha(x_1)=\alpha(x_2)\) then \(\beta(x_1)=\beta(x_2)\), so \(\beta\) is a well-defined function of \(\alpha\), and if \(J(q)\) is continuous \(\beta(\alpha)\) is also continuous. Now we can write
\[J(q + \alpha) - J (q) = \beta(\alpha).\]
Now this implies by induction
\[J(q + n\alpha)= J (q) +n\beta(\alpha)\]
Then given any two \(\alpha\) values \(\alpha_0\), \(\alpha_1\), we have
\[J(q + n_0\alpha_0)=J(q)+n_0\beta(\alpha_0),\]
\[J(q + n_1\alpha_1)=J(q)+n_1\beta(\alpha_1).\]
If \(\alpha_0/\alpha_1\) is rational then \(\alpha_1=(n_0/n_1) \alpha_0\), and after plugging into the above
\[J(q+n_0\alpha_0)=J(q)+n_0\beta(\alpha_0)=J(q)+n_1\beta(\alpha_1)\]
so
\[\beta(\alpha_0)/\beta(\alpha_1)=n_0/n_1=\alpha_0/\alpha_1\]
meaning \(\beta(\alpha_0)/\alpha_0=\beta(\alpha_1)/\alpha_1\). Since \(\beta(\alpha)\) is continuous, bing constant on all rational multiples of a given \(\alpha_0\) implies that it is constant (recall \(\alpha\) varies over an interval, on which rational multiples of \(\alpha_0\) are dense).
Now we want to repeat this argument with error terms.
Suppose \(x_1\) and \(x_2\) are such that \(\alpha(x_1)=\alpha(x_2)\).
Then \(J(q+\alpha(x))-J(q)= \beta(x)+ O(\exp\{- q \})\) implies, by plugging in sufficiently large \(q\), that the difference \(\beta(x_1)-\beta(x_2)\) is smaller than any positive number, so is zero. Thus, as before \(\beta(\alpha)\) is well-defined.
To see that \(\beta(\alpha)\) is continuous, given \(x_0\) corresponding to some \(\alpha_0\) and any \(\varepsilon>0\) pick \(q\) such that \(|O(\exp\{-q\})| <\varepsilon/2\) and, using continuity of \(J\) at \(q+\alpha_0\), pick \(\delta\) such that \(|\alpha-\alpha_0|<\delta\) implies \(|J(q+\alpha)-J(q+\alpha_0)|<\varepsilon/2\). Then \(|\beta(\alpha)-\beta(\alpha_0)|<\varepsilon\) on the same interval \(|\alpha-\alpha_0|<\delta\), meaning that \(\beta(\alpha)\) is continuous at \(\alpha_0\), as wanted.
We know that for each \(x\) and each \(C>0\) there exists \(Q(x, C)\) such that that for \(q\geq Q(x, C)\) we have \(|O(\exp\{-q \})|< C\exp\{ -q \}\). Pick any \(q(x,C)\geq Q(x, C)\).
Now we have by induction (with everything depending on \(x\))
\[|J(q+n\alpha)-(J (q)+n\beta)|\]
\[\leq C \exp\{-q\}(1+\exp\{- \alpha \}+...+\exp\{-(n-1) \alpha \})\]
\[<\frac{C}{1-\exp\{- \alpha \}} \exp\{- q \} \]
As before is \(\alpha_1=(n_0/n_1)\alpha_0\) writing the above and picking sufficiently large \(q\) we get
\[\beta(\alpha_0)/\beta(\alpha_1)=n_0/n_1=\alpha_0/\alpha_1.\] The rest is the sam as in the “error-less” case.
This shows that
\[b(x)=\beta{x}/\alpha{x}=\log \left[ \frac{x}{S(x)} \right]/ \log \left[\frac{-xS'(x)}{S(x)}\right]\]
is constant, and thus establishing 2.56.
## Proof of 2.57
Start with 2.56 and doing a bit of manipulation to isolate \(S^{\prime}(x)\)
\[ \begin{aligned} \frac{x}{S(x)} &= \left[\frac{-x S^{\prime}(x)}{S(x)}\right]^{b}\\ \frac{x^{\frac{1}{b}}}{S(x)^{\frac{1}{b}}} &= -\frac{x S^{\prime}(x)}{S(x)}\\ S^{\prime}(x) &= -\frac{x^{\frac{1}{b}} S(x)}{x S(x)^{\frac{1}{b}}}\\ &= - x^{\frac{1}{b} - 1} S(x)^{1 - \frac{1}{b}} \end{aligned} \] Expanding \(S^{\prime}(x)\) into the actual derivative, and treating them as differentials. \[ \begin{aligned} \frac{dS(x)}{d x} &= - x^{\frac{1}{b} - 1} S(x)^{1 - \frac{1}{b}}\\ S(x)^{\frac{1}{b} - 1} dS &= -x^{\frac{1}{b} - 1} dx \\ S(x)^{\frac{1}{b} - 1} dS + x^{\frac{1}{b} - 1} dx &= 0\\ S(x)^{m - 1} dS + x^{m-1} dx &=0 \end{aligned} \]
\(S^{m-1}S'=-x^{m-1}\) is equivalent to \((S^m)'=-mx^{m-1}\), so that \(S^m=C-x^m\). Initial value \(S(0)=1\) fixes \(C=1\) and \(S(x)=(1-x^m)^{1/m}\) as wanted.
Cox (1946), page 12, starting at equation (15).
I think this problem is ambiguous and can be interpreted in multiple ways, see here for a different interpretation. But I think the following interpretation makes more sense.
With \(X\) representing any background information: \[ \begin{aligned} p(C|(A+B)X) &= \frac{p(A+B|CX)p(C|X)}{p(A+B|X)}\\ &= \frac{[p(A|CX)+p(B|CX)-p(AB|CX)]p(C|X)}{p(A|X)+p(B|X)-p(AB|X)}\\ &= \frac{p(AC|X)+p(BC|X)-p(ABC|X)}{p(A|X)+p(B|X)-p(AB|X)} \end{aligned} \]
We will use convention that all \(P\) are conditioned on \(X\). So \(P(A|C)\) actually stands for \(P(A|CX)\).
First we do a bunch of lemmas about mutually exclusive propositions.
If \(A_i\) and are mutually exclusive, and \(C\) is arbitrary, then
Proof: \(P(A_i+A_j)=P(A_i)+P(A_j)-P(A_iA_j)=P(A_i)+P(A_j).\)
Proof: If \(i\neq j\) then \(P(A_iCA_jC)=P(A_iA_j)P(C|A_iA_j)=0.\)
c)\(A_i|C\) are mutually exclusive
Proof: If \(i\neq j\) then \(P(A_i|C)P(A_j|C)=P(A_iC)P(A_jC)/P(C)^2=0.\)
If \(A_1, A_2, A_3\) are mutually exclusive, then \(A_1+A_2\) and \(A_3\) are mutually exclusive.
First of all \(P(A_1A_2A_3)=P(A_1|A_2A_3)P(A_2 A_3)=0\). Then,
\[P((A_1+A_2)A_3)=P(A_1A_3+A_2A_3)=\]
\[P(A_1A_2)+P(A_2A_3)-P(A_1A_2A_3)=0.\]
With this in place, we can use induction to see
\[P(\sum A_i)=\sum P(A_i)\]
and
\[P(C(\sum A_i))=\sum P(CA_i).\]
Finally,
\[P(C(\sum A_i))=P(C|(\sum A_i))P(\sum A_i)\]
and plugging in we get
\[P(C|(\sum A_i))=\frac{P(C(\sum A_i))}{P(\sum A_i)}=\frac{\sum P(C A_i)}{\sum P(A_i)}=\frac{\sum P(A_i)P(C| A_i)}{\sum P(A_i)}.\]
Again, everything is conditional on \(C\), but we don’t write it.
Then
\[P(AB)=P(B|A)P(A)\leq P(A)=a,\]
\[P(A+B)=P(A)+P(B)-P(AB)=a+b-P(AB)\geq b.\]
Also
\[P(AB)=P(A)+P(B)-P(A+B)\geq a+b-1\]
and
\[P(A+B)=P(A)+P(B)-P(AB)\leq a+b.\]