<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://tcs.nju.edu.cn/wiki/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Zhangxy</id>
	<title>TCS Wiki - User contributions [en]</title>
	<link rel="self" type="application/atom+xml" href="https://tcs.nju.edu.cn/wiki/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Zhangxy"/>
	<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=Special:Contributions/Zhangxy"/>
	<updated>2026-04-28T19:26:04Z</updated>
	<subtitle>User contributions</subtitle>
	<generator>MediaWiki 1.45.3</generator>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_4&amp;diff=13125</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 4</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_4&amp;diff=13125"/>
		<updated>2025-05-07T08:33:49Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: /* Problem 3 (Strong Law of Large Numbers, 15 points, Bonus Problem) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
*Bonus problem为附加题（选做）。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 4==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Continuous Random Variables, 30 points)==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Density function&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Determine the value of &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;f(x) = C\exp(-x-e^{-x}), x\in \mathbb{R}&amp;lt;/math&amp;gt; is a probability density function (PDF) for a continuous random variable.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Independence&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be independent and identically distributed continuous random variables with cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; and probability density function (PDF) &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt;. Find out the density functions of &amp;lt;math&amp;gt;V = \max\{X,Y\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;U = \min\{X,Y\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Correlation&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be uniformly distributed on &amp;lt;math&amp;gt;(-1,1)&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_k = \cos(k \pi X)&amp;lt;/math&amp;gt; for &amp;lt;math&amp;gt;k=1,2,\ldots,n&amp;lt;/math&amp;gt;. Are the random variables &amp;lt;math&amp;gt;Y_1, Y_2, \ldots, Y_n&amp;lt;/math&amp;gt; correlated? independent? You should prove your claim rigorously.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Expectation of random variables (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a continuous random variable with mean &amp;lt;math&amp;gt;\mu&amp;lt;/math&amp;gt; and cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X \ge 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\int_{0}^a F(x) dx = \int_{a}^{\infty} [1-F(x)] dx&amp;lt;/math&amp;gt; if and only if &amp;lt;math&amp;gt;a = \mu&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; has finite variance. Show that &amp;lt;math&amp;gt;g(a) = \mathbb{E}((X-a)^2)&amp;lt;/math&amp;gt; achieves the minimum when &amp;lt;math&amp;gt;a = \mu&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Expectation of random variables (II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X, Y&amp;lt;/math&amp;gt; be two independent and identically distributed continuous random variables with cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;. Furthermore, &amp;lt;math&amp;gt;X,Y \ge 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbb{E}[|X-Y|] = 2 \left(\mathbb{E}[X] - \int_{0}^{\infty} (1-F(x))^2 dx\right)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Conditional distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be two random variables. The joint density of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; is given by &amp;lt;math&amp;gt;f(x,y) = c(x^2 - y^2)e^{-x}&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;0\leq x &amp;lt;\infty&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-x\leq y \leq x&amp;lt;/math&amp;gt;. Here, &amp;lt;math&amp;gt;c\in \mathbb{R}_+&amp;lt;/math&amp;gt; is a constant. Find out the conditional distribution of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;, given &amp;lt;math&amp;gt;X = x&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Uniform Distribution (I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;P_i = (X_i,Y_i), 1\leq i\leq n&amp;lt;/math&amp;gt;, be independent, uniformly distributed points in the unit square &amp;lt;math&amp;gt;[0,1]^2&amp;lt;/math&amp;gt;. A point &amp;lt;math&amp;gt;P_i&amp;lt;/math&amp;gt; is called &amp;quot;peripheral&amp;quot; if, for all &amp;lt;math&amp;gt;r = 1,2,\cdots,n&amp;lt;/math&amp;gt;, either &amp;lt;math&amp;gt;X_r \leq X_i&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;Y_r \leq Y_i&amp;lt;/math&amp;gt;, or both. Find out the expected number of peripheral points.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Uniform Distribution (II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X,Y&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; be independent and uniformly distributed on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt;. Find the joint density function of &amp;lt;math&amp;gt;XY&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Z^2&amp;lt;/math&amp;gt;, and show that &amp;lt;math&amp;gt;\textbf{Pr}[XY&amp;lt;Z^2] = 5/9&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Exponential distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; have an exponential distribution. Show that &amp;lt;math&amp;gt;\textbf{Pr}[X&amp;gt;s+x|X&amp;gt;s] = \textbf{Pr}[X&amp;gt;x]&amp;lt;/math&amp;gt;, for &amp;lt;math&amp;gt;x,s\geq 0&amp;lt;/math&amp;gt;. This is the memoryless property. Show that the exponential distribution is the only continuous distribution with this property.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Normal distribution(I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X,Y\sim N(0,1)&amp;lt;/math&amp;gt; be two independent and identically distributed normal random variables. Let &amp;lt;math&amp;gt;Z = X-Y&amp;lt;/math&amp;gt;. Find the density function of &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;|Z|&amp;lt;/math&amp;gt; respectively.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Normal distribution(II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; have the &amp;lt;math&amp;gt;N(0,1)&amp;lt;/math&amp;gt; distribution and let &amp;lt;math&amp;gt;a&amp;gt;0&amp;lt;/math&amp;gt;. Show that the random variable &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; given by&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{equation*}&lt;br /&gt;
Y = \begin{cases}&lt;br /&gt;
X, &amp;amp; |X|&amp;lt; a \\&lt;br /&gt;
-X, &amp;amp; |X|\geq a&lt;br /&gt;
\end{cases}&lt;br /&gt;
\end{equation*}&amp;lt;/math&amp;gt;&lt;br /&gt;
has the &amp;lt;math&amp;gt;N(0,1)&amp;lt;/math&amp;gt; distribution, and find an expression for &amp;lt;math&amp;gt;\rho(a) = \textbf{Cov}(X,Y)&amp;lt;/math&amp;gt; in terms of the density function &amp;lt;math&amp;gt;\phi&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random process&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Given a real number &amp;lt;math&amp;gt;U&amp;lt;1&amp;lt;/math&amp;gt; as input of the following process, find out the expected returning value.&lt;br /&gt;
{{Theorem|&#039;&#039;Process&#039;&#039;|&lt;br /&gt;
:&#039;&#039;&#039;Input:&#039;&#039;&#039;  real numbers &amp;lt;math&amp;gt;U &amp;lt; 1&amp;lt;/math&amp;gt;;&lt;br /&gt;
----&lt;br /&gt;
:initialize &amp;lt;math&amp;gt;x = 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;count = 0&amp;lt;/math&amp;gt;;&lt;br /&gt;
:while &amp;lt;math&amp;gt; x &amp;lt; U &amp;lt;/math&amp;gt; do&lt;br /&gt;
:* choose &amp;lt;math&amp;gt;y \in (0,1)&amp;lt;/math&amp;gt; uniformly at random;&lt;br /&gt;
:* update &amp;lt;math&amp;gt;x = x + y&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;count = count + 1&amp;lt;/math&amp;gt;;&lt;br /&gt;
:return &amp;lt;math&amp;gt;count&amp;lt;/math&amp;gt;;&lt;br /&gt;
}}&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Random semicircle&amp;lt;/strong&amp;gt;] We sample &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; points within a circle &amp;lt;math&amp;gt;C=\{(x,y) \in \mathbb{R}^2 \mid x^2+y^2 \le 1\}&amp;lt;/math&amp;gt; independently and uniformly at random (i.e., the density function &amp;lt;math&amp;gt;f(x,y) \propto 1_{(x,y) \in C}&amp;lt;/math&amp;gt;). Find out the probability that they all lie within some semicircle of the original circle &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;. (Hint: you may apply the technique of change of variables, see [https://en.wikipedia.org/wiki/Random_variable#Functions_of_random_variables function of random variables] or Chapter 4.7 in [GS])&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Stochastic domination&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X, Y&amp;lt;/math&amp;gt; be continuous random variables. Show that &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; dominates &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; stochastically if and only if &amp;lt;math&amp;gt;\mathbb{E}[f(X)]\geq \mathbb{E}[f(Y)]&amp;lt;/math&amp;gt; for any non-decreasing function &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; for which the expectations exist.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Concentration of measure, 10 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Tossing coins&amp;lt;/strong&amp;gt;] We repeatedly toss a fair coin (with an equal probability of heads and tails). Let the random variable &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of throws required to obtain a total of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; heads. Show that &amp;lt;math&amp;gt;\textbf{Pr}[X &amp;gt; 2n + 2\sqrt{n\log n}]\leq O(1/n)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Chernoff vs Chebyshev&amp;lt;/strong&amp;gt;] We have a standard six-sided die. Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of times a 6 occurs in &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; throws off the die. Compare the best upper bounds on &amp;lt;math&amp;gt;\textbf{Pr}[X\geq n/4]&amp;lt;/math&amp;gt; that you can obtain using Chebyshev&#039;s inequality and Chernoff bounds.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;&amp;lt;strong&amp;gt;-th moment bound]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a random variable with expectation &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; such that moment generating function &amp;lt;math&amp;gt;\mathbf{E}[\exp(t|X|)]&amp;lt;/math&amp;gt; is finite for some &amp;lt;math&amp;gt; t &amp;gt; 0 &amp;lt;/math&amp;gt;. We can use the following two kinds of tail inequalities for &amp;lt;math&amp;gt; X &amp;lt;/math&amp;gt;: &lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&amp;lt;nowiki&amp;gt;     &amp;lt;/nowiki&amp;gt;&#039;&#039;&#039;&#039;&#039;Chernoff Bound&#039;&#039;&#039;&#039;&#039;&lt;br /&gt;
:&amp;lt;math&amp;gt;&lt;br /&gt;
\begin{align}&lt;br /&gt;
\mathbf{Pr}[|X| \geq \delta] \leq \min_{t \geq 0} \frac{\mathbf{E}[e^{t|X|}]}{e^{t\delta}}&lt;br /&gt;
\end{align}&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&#039;&#039;&#039;&#039;&#039;&amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;th-Moment Bound&#039;&#039;&#039;&#039;&#039;&lt;br /&gt;
:&amp;lt;math&amp;gt;&lt;br /&gt;
\begin{align}&lt;br /&gt;
\mathbf{Pr}[|X| \geq \delta] \leq \frac{\mathbf{E}[|X|^k]}{\delta^k}&lt;br /&gt;
\end{align}&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
# Show that for each &amp;lt;math&amp;gt;\delta&amp;lt;/math&amp;gt;, there exists a choice of &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; such that the &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;th-moment bound is no weaker than the Chernoff bound. (Hint: Use the probabilistic method.)&lt;br /&gt;
# Why would we still prefer the Chernoff bound to the (seemingly) stronger &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;-th moment bound?&lt;br /&gt;
&lt;br /&gt;
* [&#039;&#039;&#039;Cut size in random graph&#039;&#039;&#039;] Show that with probability at least &amp;lt;math&amp;gt;2/3&amp;lt;/math&amp;gt;, the size of the max-cut in [[wikipedia:Erdős–Rényi_model#Definition|Erdős–Rényi random graph]] &amp;lt;math&amp;gt;G(n,1/2)&amp;lt;/math&amp;gt; is at most &amp;lt;math&amp;gt;n^2/8 + O(n^{1.5})&amp;lt;/math&amp;gt;. In the &amp;lt;math&amp;gt;G(n,1/2)&amp;lt;/math&amp;gt; model, each edge is included in the graph with probability &amp;lt;math&amp;gt;1/2&amp;lt;/math&amp;gt;, independently of every other edge.&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Strong Law of Large Numbers, 15 points, &#039;&#039;&#039;Bonus Problem&#039;&#039;&#039;)==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
Throughout this problem, we assume &amp;lt;math&amp;gt;X_1,X_2,\ldots,&amp;lt;/math&amp;gt; be jointly independent square-integrable real random variables of mean zero. We will prove the strong law of large numbers by Kolmogorov maximal inequality.&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Kolmogorov maximal inequality&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;S_n = \sum_{i=1}^n X_i&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{Pr}\left(\max_{1 \le i \le n} |S_i| \ge t \right) \le \frac{\mathbf{Var}(S_n)}{t^2}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Convergence of random series&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Var}(X_i) &amp;lt; \infty&amp;lt;/math&amp;gt;. Prove that the series &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} X_i&amp;lt;/math&amp;gt; is almost surely convergent.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Strong law of large numbers&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Prove the strong law of large numbers using previous propositions.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 4 (LLN and CLT, 15 points + 10 points) ==&lt;br /&gt;
&amp;lt;strong&amp;gt;In this problem, you may apply the results of Laws of Large Numbers (LLN) and the Central Limit Theorem (CLT) to solve the problems.&amp;lt;/strong&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;St. Petersburg paradox&amp;lt;/strong&amp;gt;] Consider the well-known game involving a fair coin. In this game, if it takes &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; tosses to obtain a head, you will win &amp;lt;math&amp;gt;2^k&amp;lt;/math&amp;gt; dollars as the reward. Despite the game&#039;s expected reward being infinite, people tend to offer relatively modest amounts to participate. The following provides a mathematical explanation for this phenomenon.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
For each &amp;lt;math&amp;gt;n \ge 1&amp;lt;/math&amp;gt;, let &amp;lt;math&amp;gt;X_{n,1}, X_{n,2},\ldots, X_{n,k}&amp;lt;/math&amp;gt; be independent random variables. Furthermore, let &amp;lt;math&amp;gt;b_n &amp;gt; 0&amp;lt;/math&amp;gt; be real numbers with &amp;lt;math&amp;gt;b_n \to \infty&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widetilde{X}_{n,k} = X_{n,k} \mathbf{1}_{|X_{n,k}| \le b_n}&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;1 \le k \le n&amp;lt;/math&amp;gt;. If &amp;lt;math&amp;gt;\sum_{k=1}^n \mathbf{Pr}(|X_{n,k}| &amp;gt; b_n) \to 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;b_n^{-2} \sum_{k=1}^n \mathbf{E}[\widetilde{X}_{n,k}^2] \to 0&amp;lt;/math&amp;gt; when &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, then &amp;lt;math&amp;gt;(S_n-a_n)/b_n \overset{P}{\to} 0 &amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;S_n = \sum_{k=1}^n X_{n,k}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;a_n = \sum_{k=1}^n \mathbf{E}[\widetilde{X}_{n,k}]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;S_n&amp;lt;/math&amp;gt; be the total winnings after playing &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds of the game. Prove that &amp;lt;math&amp;gt;\frac{S_n}{n \log_2 n} \overset{P}{\to} 1&amp;lt;/math&amp;gt;. (Therefore, a fair price to play this game &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; times is roughly &amp;lt;math&amp;gt;n \log_2 n&amp;lt;/math&amp;gt; dollars)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; (&amp;lt;strong&amp;gt;Bonus problem, 5 points&amp;lt;/strong&amp;gt;)&lt;br /&gt;
Let &amp;lt;math&amp;gt;S_n&amp;lt;/math&amp;gt; be the total winnings after playing &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds of the game. Prove that &amp;lt;math&amp;gt; \limsup_{n \to \infty} \frac{S_n}{n \log_2 n} = \infty&amp;lt;/math&amp;gt; almost surely. (Hint: You may use Borel-Cantelli lemmas)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; (&amp;lt;strong&amp;gt;Bonus problem, 5 points&amp;lt;/strong&amp;gt;)&lt;br /&gt;
Let &amp;lt;math&amp;gt;X_1,X_2,\ldots,X_n&amp;lt;/math&amp;gt; are i.i.d. copies of a random variable &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; with the standard Cauchy distribution (i.e., the probability density function of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; is &amp;lt;math&amp;gt;\frac{1}{\pi} \frac{1}{1+x^2}&amp;lt;/math&amp;gt;), show that &amp;lt;math&amp;gt;\frac{S_n}{n \log n}&amp;lt;/math&amp;gt; converges in probability to a constant &amp;lt;math&amp;gt;\frac{2}{\pi}&amp;lt;/math&amp;gt; but is almost surely unbounded, where &amp;lt;math&amp;gt;S_n = \sum_{i=1}^n |X_i|&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Monte Carlo integration&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; be a continuous function on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;U_1,U_2,\ldots,U_n \sim U(0,1)&amp;lt;/math&amp;gt; be independent random variables. Show that &amp;lt;math&amp;gt;I = \frac{1}{n}\sum_{i=1}^n f(U_i) \to \int_0^1 f(x) \mathrm{d}x&amp;lt;/math&amp;gt; in probability. (Remark: This holds so long as &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; is a measurable function on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;\int_0^1 |f(x)| \mathrm{d}x &amp;lt; \infty&amp;lt;/math&amp;gt;)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Stirling&#039;s formula&amp;lt;/strong&amp;gt;]&lt;br /&gt;
By considering the central limit theorem for the sum of independent Poisson-distributed random variables, show that&lt;br /&gt;
&amp;lt;math&amp;gt;n! \sim \sqrt{2\pi n} \cdot \left(\frac{n}{\mathrm{e}}\right)^n&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_4&amp;diff=13102</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 4</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_4&amp;diff=13102"/>
		<updated>2025-04-27T15:34:19Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: /* Problem 4 (LLN and CLT, 15 points + 10 points) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
*Bonus problem为附加题（选做）。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 4==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Continuous Random Variables, 30 points)==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Density function&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Determine the value of &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;f(x) = C\exp(-x-e^{-x}), x\in \mathbb{R}&amp;lt;/math&amp;gt; is a probability density function (PDF) for a continuous random variable.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Independence&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be independent and identically distributed continuous random variables with cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; and probability density function (PDF) &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt;. Find out the density functions of &amp;lt;math&amp;gt;V = \max\{X,Y\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;U = \min\{X,Y\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Correlation&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be uniformly distributed on &amp;lt;math&amp;gt;(-1,1)&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_k = \cos(k \pi X)&amp;lt;/math&amp;gt; for &amp;lt;math&amp;gt;k=1,2,\ldots,n&amp;lt;/math&amp;gt;. Are the random variables &amp;lt;math&amp;gt;Y_1, Y_2, \ldots, Y_n&amp;lt;/math&amp;gt; correlated? independent? You should prove your claim rigorously.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Expectation of random variables (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a continuous random variable with mean &amp;lt;math&amp;gt;\mu&amp;lt;/math&amp;gt; and cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X \ge 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\int_{0}^a F(x) dx = \int_{a}^{\infty} [1-F(x)] dx&amp;lt;/math&amp;gt; if and only if &amp;lt;math&amp;gt;a = \mu&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; has finite variance. Show that &amp;lt;math&amp;gt;g(a) = \mathbb{E}((X-a)^2)&amp;lt;/math&amp;gt; achieves the minimum when &amp;lt;math&amp;gt;a = \mu&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Expectation of random variables (II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X, Y&amp;lt;/math&amp;gt; be two independent and identically distributed continuous random variables with cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;. Furthermore, &amp;lt;math&amp;gt;X,Y \ge 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbb{E}[|X-Y|] = 2 \left(\mathbb{E}[X] - \int_{0}^{\infty} (1-F(x))^2 dx\right)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Conditional distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be two random variables. The joint density of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; is given by &amp;lt;math&amp;gt;f(x,y) = c(x^2 - y^2)e^{-x}&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;0\leq x &amp;lt;\infty&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-x\leq y \leq x&amp;lt;/math&amp;gt;. Here, &amp;lt;math&amp;gt;c\in \mathbb{R}_+&amp;lt;/math&amp;gt; is a constant. Find out the conditional distribution of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;, given &amp;lt;math&amp;gt;X = x&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Uniform Distribution (I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;P_i = (X_i,Y_i), 1\leq i\leq n&amp;lt;/math&amp;gt;, be independent, uniformly distributed points in the unit square &amp;lt;math&amp;gt;[0,1]^2&amp;lt;/math&amp;gt;. A point &amp;lt;math&amp;gt;P_i&amp;lt;/math&amp;gt; is called &amp;quot;peripheral&amp;quot; if, for all &amp;lt;math&amp;gt;r = 1,2,\cdots,n&amp;lt;/math&amp;gt;, either &amp;lt;math&amp;gt;X_r \leq X_i&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;Y_r \leq Y_i&amp;lt;/math&amp;gt;, or both. Find out the expected number of peripheral points.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Uniform Distribution (II)&amp;lt;/strong&amp;gt;] Derive the moment generating function of the standard uniform distribution, i.e., uniform distribution on &amp;lt;math&amp;gt;(0,1)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Exponential distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; have an exponential distribution. Show that &amp;lt;math&amp;gt;\textbf{Pr}[X&amp;gt;s+x|X&amp;gt;s] = \textbf{Pr}[X&amp;gt;x]&amp;lt;/math&amp;gt;, for &amp;lt;math&amp;gt;x,s\geq 0&amp;lt;/math&amp;gt;. This is the memoryless property. Show that the exponential distribution is the only continuous distribution with this property.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Normal distribution(I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X,Y\sim N(0,1)&amp;lt;/math&amp;gt; be two independent and identically distributed normal random variables. Let &amp;lt;math&amp;gt;Z = X-Y&amp;lt;/math&amp;gt;. Find the density function of &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;|Z|&amp;lt;/math&amp;gt; respectively.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Normal distribution(II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; have the &amp;lt;math&amp;gt;N(0,1)&amp;lt;/math&amp;gt; distribution and let &amp;lt;math&amp;gt;a&amp;gt;0&amp;lt;/math&amp;gt;. Show that the random variable &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; given by&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{equation*}&lt;br /&gt;
Y = \begin{cases}&lt;br /&gt;
X, &amp;amp; |X|&amp;lt; a \\&lt;br /&gt;
-X, &amp;amp; |X|\geq a&lt;br /&gt;
\end{cases}&lt;br /&gt;
\end{equation*}&amp;lt;/math&amp;gt;&lt;br /&gt;
has the &amp;lt;math&amp;gt;N(0,1)&amp;lt;/math&amp;gt; distribution, and find an expression for &amp;lt;math&amp;gt;\rho(a) = \textbf{Cov}(X,Y)&amp;lt;/math&amp;gt; in terms of the density function &amp;lt;math&amp;gt;\phi&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random process&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Given a real number &amp;lt;math&amp;gt;U&amp;lt;1&amp;lt;/math&amp;gt; as input of the following process, find out the expected returning value.&lt;br /&gt;
{{Theorem|&#039;&#039;Process&#039;&#039;|&lt;br /&gt;
:&#039;&#039;&#039;Input:&#039;&#039;&#039;  real numbers &amp;lt;math&amp;gt;U &amp;lt; 1&amp;lt;/math&amp;gt;;&lt;br /&gt;
----&lt;br /&gt;
:initialize &amp;lt;math&amp;gt;x = 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;count = 0&amp;lt;/math&amp;gt;;&lt;br /&gt;
:while &amp;lt;math&amp;gt; x &amp;lt; U &amp;lt;/math&amp;gt; do&lt;br /&gt;
:* choose &amp;lt;math&amp;gt;y \in (0,1)&amp;lt;/math&amp;gt; uniformly at random;&lt;br /&gt;
:* update &amp;lt;math&amp;gt;x = x + y&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;count = count + 1&amp;lt;/math&amp;gt;;&lt;br /&gt;
:return &amp;lt;math&amp;gt;count&amp;lt;/math&amp;gt;;&lt;br /&gt;
}}&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Random semicircle&amp;lt;/strong&amp;gt;] We sample &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; points within a circle &amp;lt;math&amp;gt;C=\{(x,y) \in \mathbb{R}^2 \mid x^2+y^2 \le 1\}&amp;lt;/math&amp;gt; independently and uniformly at random (i.e., the density function &amp;lt;math&amp;gt;f(x,y) \propto 1_{(x,y) \in C}&amp;lt;/math&amp;gt;). Find out the probability that they all lie within some semicircle of the original circle &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;. (Hint: you may apply the technique of change of variables, see [https://en.wikipedia.org/wiki/Random_variable#Functions_of_random_variables function of random variables] or Chapter 4.7 in [GS])&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Stochastic domination&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X, Y&amp;lt;/math&amp;gt; be continuous random variables. Show that &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; dominates &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; stochastically if and only if &amp;lt;math&amp;gt;\mathbb{E}[f(X)]\geq \mathbb{E}[f(Y)]&amp;lt;/math&amp;gt; for any non-decreasing function &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; for which the expectations exist.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Concentration of measure, 10 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Tossing coins&amp;lt;/strong&amp;gt;] We repeatedly toss a fair coin (with an equal probability of heads and tails). Let the random variable &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of throws required to obtain a total of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; heads. Show that &amp;lt;math&amp;gt;\textbf{Pr}[X &amp;gt; 2n + 2\sqrt{n\log n}]\leq O(1/n)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Chernoff vs Chebyshev&amp;lt;/strong&amp;gt;] We have a standard six-sided die. Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of times a 6 occurs in &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; throws off the die. Compare the best upper bounds on &amp;lt;math&amp;gt;\textbf{Pr}[X\geq n/4]&amp;lt;/math&amp;gt; that you can obtain using Chebyshev&#039;s inequality and Chernoff bounds.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;&amp;lt;strong&amp;gt;-th moment bound]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a random variable with expectation &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; such that moment generating function &amp;lt;math&amp;gt;\mathbf{E}[\exp(t|X|)]&amp;lt;/math&amp;gt; is finite for some &amp;lt;math&amp;gt; t &amp;gt; 0 &amp;lt;/math&amp;gt;. We can use the following two kinds of tail inequalities for &amp;lt;math&amp;gt; X &amp;lt;/math&amp;gt;: &lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&amp;lt;nowiki&amp;gt;     &amp;lt;/nowiki&amp;gt;&#039;&#039;&#039;&#039;&#039;Chernoff Bound&#039;&#039;&#039;&#039;&#039;&lt;br /&gt;
:&amp;lt;math&amp;gt;&lt;br /&gt;
\begin{align}&lt;br /&gt;
\mathbf{Pr}[|X| \geq \delta] \leq \min_{t \geq 0} \frac{\mathbf{E}[e^{t|X|}]}{e^{t\delta}}&lt;br /&gt;
\end{align}&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&#039;&#039;&#039;&#039;&#039;&amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;th-Moment Bound&#039;&#039;&#039;&#039;&#039;&lt;br /&gt;
:&amp;lt;math&amp;gt;&lt;br /&gt;
\begin{align}&lt;br /&gt;
\mathbf{Pr}[|X| \geq \delta] \leq \frac{\mathbf{E}[|X|^k]}{\delta^k}&lt;br /&gt;
\end{align}&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
# Show that for each &amp;lt;math&amp;gt;\delta&amp;lt;/math&amp;gt;, there exists a choice of &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; such that the &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;th-moment bound is no weaker than the Chernoff bound. (Hint: Use the probabilistic method.)&lt;br /&gt;
# Why would we still prefer the Chernoff bound to the (seemingly) stronger &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;-th moment bound?&lt;br /&gt;
&lt;br /&gt;
* [&#039;&#039;&#039;Cut size in random graph&#039;&#039;&#039;] Show that with probability at least &amp;lt;math&amp;gt;2/3&amp;lt;/math&amp;gt;, the size of the max-cut in [[wikipedia:Erdős–Rényi_model#Definition|Erdős–Rényi random graph]] &amp;lt;math&amp;gt;G(n,1/2)&amp;lt;/math&amp;gt; is at most &amp;lt;math&amp;gt;n^2/8 + O(n^{1.5})&amp;lt;/math&amp;gt;. In the &amp;lt;math&amp;gt;G(n,1/2)&amp;lt;/math&amp;gt; model, each edge is included in the graph with probability &amp;lt;math&amp;gt;1/2&amp;lt;/math&amp;gt;, independently of every other edge.&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Strong Law of Large Numbers, 15 points, &#039;&#039;&#039;Bonus Problem&#039;&#039;&#039;)==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
Throughout this problem, we assume &amp;lt;math&amp;gt;X_1,X_2,\ldots,&amp;lt;/math&amp;gt; be jointly independent square-integrable real random variables of mean zero. We will prove the strong law of large numbers by Kolmogorov maximal inequality.&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Kolmogorov maximal inequality&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;S_n = \sum_{i=1}^n X_i&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{Pr}\left(\max_{1 \le i \le n} |S_i| \ge t \right) \le \frac{n \mathbf{Var}(X_1)}{t^2}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Convergence of random series&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Var}(X_i) &amp;lt; \infty&amp;lt;/math&amp;gt;. Prove that the series &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} X_i&amp;lt;/math&amp;gt; is almost surely convergent.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Strong law of large numbers&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Prove the strong law of large numbers using previous propositions.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 4 (LLN and CLT, 15 points + 10 points) ==&lt;br /&gt;
&amp;lt;strong&amp;gt;In this problem, you may apply the results of Laws of Large Numbers (LLN) and the Central Limit Theorem (CLT) to solve the problems.&amp;lt;/strong&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;St. Petersburg paradox&amp;lt;/strong&amp;gt;] Consider the well-known game involving a fair coin. In this game, if it takes &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; tosses to obtain a head, you will win &amp;lt;math&amp;gt;2^k&amp;lt;/math&amp;gt; dollars as the reward. Despite the game&#039;s expected reward being infinite, people tend to offer relatively modest amounts to participate. The following provides a mathematical explanation for this phenomenon.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
For each &amp;lt;math&amp;gt;n \ge 1&amp;lt;/math&amp;gt;, let &amp;lt;math&amp;gt;X_{n,1}, X_{n,2},\ldots, X_{n,k}&amp;lt;/math&amp;gt; be independent random variables. Furthermore, let &amp;lt;math&amp;gt;b_n &amp;gt; 0&amp;lt;/math&amp;gt; be real numbers with &amp;lt;math&amp;gt;b_n \to \infty&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widetilde{X}_{n,k} = X_{n,k} \mathbf{1}_{|X_{n,k}| \le b_n}&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;1 \le k \le n&amp;lt;/math&amp;gt;. If &amp;lt;math&amp;gt;\sum_{k=1}^n \mathbf{Pr}(|X_{n,k}| &amp;gt; b_n) \to 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;b_n^{-2} \sum_{k=1}^n \mathbf{E}[\widetilde{X}_{n,k}^2] \to 0&amp;lt;/math&amp;gt; when &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, then &amp;lt;math&amp;gt;(S_n-a_n)/b_n \overset{P}{\to} 0 &amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;S_n = \sum_{k=1}^n X_{n,k}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;a_n = \sum_{k=1}^n \mathbf{E}[\widetilde{X}_{n,k}]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;S_n&amp;lt;/math&amp;gt; be the total winnings after playing &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds of the game. Prove that &amp;lt;math&amp;gt;\frac{S_n}{n \log_2 n} \overset{P}{\to} 1&amp;lt;/math&amp;gt;. (Therefore, a fair price to play this game &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; times is roughly &amp;lt;math&amp;gt;n \log_2 n&amp;lt;/math&amp;gt; dollars)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; (&amp;lt;strong&amp;gt;Bonus problem, 5 points&amp;lt;/strong&amp;gt;)&lt;br /&gt;
Let &amp;lt;math&amp;gt;S_n&amp;lt;/math&amp;gt; be the total winnings after playing &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds of the game. Prove that &amp;lt;math&amp;gt; \limsup_{n \to \infty} \frac{S_n}{n \log_2 n} = \infty&amp;lt;/math&amp;gt; almost surely. (Hint: You may use Borel-Cantelli lemmas)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; (&amp;lt;strong&amp;gt;Bonus problem, 5 points&amp;lt;/strong&amp;gt;)&lt;br /&gt;
Let &amp;lt;math&amp;gt;X_1,X_2,\ldots,X_n&amp;lt;/math&amp;gt; are i.i.d. copies of a random variable &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; with the standard Cauchy distribution (i.e., the probability density function of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; is &amp;lt;math&amp;gt;\frac{1}{\pi} \frac{1}{1+x^2}&amp;lt;/math&amp;gt;), show that &amp;lt;math&amp;gt;\frac{S_n}{n \log n}&amp;lt;/math&amp;gt; converges in probability to a constant &amp;lt;math&amp;gt;\frac{2}{\pi}&amp;lt;/math&amp;gt; but is almost surely unbounded, where &amp;lt;math&amp;gt;S_n = \sum_{i=1}^n |X_i|&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Monte Carlo integration&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; be a continuous function on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;U_1,U_2,\ldots,U_n \sim U(0,1)&amp;lt;/math&amp;gt; be independent random variables. Show that &amp;lt;math&amp;gt;I = \frac{1}{n}\sum_{i=1}^n f(U_i) \to \int_0^1 f(x) \mathrm{d}x&amp;lt;/math&amp;gt; in probability. (Remark: This holds so long as &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; is a measurable function on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;\int_0^1 |f(x)| \mathrm{d}x &amp;lt; \infty&amp;lt;/math&amp;gt;)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Stirling&#039;s formula&amp;lt;/strong&amp;gt;]&lt;br /&gt;
By considering the central limit theorem for the sum of independent Poisson-distributed random variables, show that&lt;br /&gt;
&amp;lt;math&amp;gt;n! \sim \sqrt{2\pi n} \cdot \left(\frac{n}{\mathrm{e}}\right)^n&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_4&amp;diff=13101</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 4</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_4&amp;diff=13101"/>
		<updated>2025-04-27T15:34:06Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: /* Problem 4 (LLN and CLT, 15 points + 10 points) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
*Bonus problem为附加题（选做）。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 4==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Continuous Random Variables, 30 points)==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Density function&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Determine the value of &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;f(x) = C\exp(-x-e^{-x}), x\in \mathbb{R}&amp;lt;/math&amp;gt; is a probability density function (PDF) for a continuous random variable.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Independence&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be independent and identically distributed continuous random variables with cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; and probability density function (PDF) &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt;. Find out the density functions of &amp;lt;math&amp;gt;V = \max\{X,Y\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;U = \min\{X,Y\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Correlation&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be uniformly distributed on &amp;lt;math&amp;gt;(-1,1)&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_k = \cos(k \pi X)&amp;lt;/math&amp;gt; for &amp;lt;math&amp;gt;k=1,2,\ldots,n&amp;lt;/math&amp;gt;. Are the random variables &amp;lt;math&amp;gt;Y_1, Y_2, \ldots, Y_n&amp;lt;/math&amp;gt; correlated? independent? You should prove your claim rigorously.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Expectation of random variables (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a continuous random variable with mean &amp;lt;math&amp;gt;\mu&amp;lt;/math&amp;gt; and cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X \ge 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\int_{0}^a F(x) dx = \int_{a}^{\infty} [1-F(x)] dx&amp;lt;/math&amp;gt; if and only if &amp;lt;math&amp;gt;a = \mu&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; has finite variance. Show that &amp;lt;math&amp;gt;g(a) = \mathbb{E}((X-a)^2)&amp;lt;/math&amp;gt; achieves the minimum when &amp;lt;math&amp;gt;a = \mu&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Expectation of random variables (II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X, Y&amp;lt;/math&amp;gt; be two independent and identically distributed continuous random variables with cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;. Furthermore, &amp;lt;math&amp;gt;X,Y \ge 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbb{E}[|X-Y|] = 2 \left(\mathbb{E}[X] - \int_{0}^{\infty} (1-F(x))^2 dx\right)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Conditional distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be two random variables. The joint density of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; is given by &amp;lt;math&amp;gt;f(x,y) = c(x^2 - y^2)e^{-x}&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;0\leq x &amp;lt;\infty&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-x\leq y \leq x&amp;lt;/math&amp;gt;. Here, &amp;lt;math&amp;gt;c\in \mathbb{R}_+&amp;lt;/math&amp;gt; is a constant. Find out the conditional distribution of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;, given &amp;lt;math&amp;gt;X = x&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Uniform Distribution (I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;P_i = (X_i,Y_i), 1\leq i\leq n&amp;lt;/math&amp;gt;, be independent, uniformly distributed points in the unit square &amp;lt;math&amp;gt;[0,1]^2&amp;lt;/math&amp;gt;. A point &amp;lt;math&amp;gt;P_i&amp;lt;/math&amp;gt; is called &amp;quot;peripheral&amp;quot; if, for all &amp;lt;math&amp;gt;r = 1,2,\cdots,n&amp;lt;/math&amp;gt;, either &amp;lt;math&amp;gt;X_r \leq X_i&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;Y_r \leq Y_i&amp;lt;/math&amp;gt;, or both. Find out the expected number of peripheral points.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Uniform Distribution (II)&amp;lt;/strong&amp;gt;] Derive the moment generating function of the standard uniform distribution, i.e., uniform distribution on &amp;lt;math&amp;gt;(0,1)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Exponential distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; have an exponential distribution. Show that &amp;lt;math&amp;gt;\textbf{Pr}[X&amp;gt;s+x|X&amp;gt;s] = \textbf{Pr}[X&amp;gt;x]&amp;lt;/math&amp;gt;, for &amp;lt;math&amp;gt;x,s\geq 0&amp;lt;/math&amp;gt;. This is the memoryless property. Show that the exponential distribution is the only continuous distribution with this property.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Normal distribution(I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X,Y\sim N(0,1)&amp;lt;/math&amp;gt; be two independent and identically distributed normal random variables. Let &amp;lt;math&amp;gt;Z = X-Y&amp;lt;/math&amp;gt;. Find the density function of &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;|Z|&amp;lt;/math&amp;gt; respectively.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Normal distribution(II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; have the &amp;lt;math&amp;gt;N(0,1)&amp;lt;/math&amp;gt; distribution and let &amp;lt;math&amp;gt;a&amp;gt;0&amp;lt;/math&amp;gt;. Show that the random variable &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; given by&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{equation*}&lt;br /&gt;
Y = \begin{cases}&lt;br /&gt;
X, &amp;amp; |X|&amp;lt; a \\&lt;br /&gt;
-X, &amp;amp; |X|\geq a&lt;br /&gt;
\end{cases}&lt;br /&gt;
\end{equation*}&amp;lt;/math&amp;gt;&lt;br /&gt;
has the &amp;lt;math&amp;gt;N(0,1)&amp;lt;/math&amp;gt; distribution, and find an expression for &amp;lt;math&amp;gt;\rho(a) = \textbf{Cov}(X,Y)&amp;lt;/math&amp;gt; in terms of the density function &amp;lt;math&amp;gt;\phi&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random process&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Given a real number &amp;lt;math&amp;gt;U&amp;lt;1&amp;lt;/math&amp;gt; as input of the following process, find out the expected returning value.&lt;br /&gt;
{{Theorem|&#039;&#039;Process&#039;&#039;|&lt;br /&gt;
:&#039;&#039;&#039;Input:&#039;&#039;&#039;  real numbers &amp;lt;math&amp;gt;U &amp;lt; 1&amp;lt;/math&amp;gt;;&lt;br /&gt;
----&lt;br /&gt;
:initialize &amp;lt;math&amp;gt;x = 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;count = 0&amp;lt;/math&amp;gt;;&lt;br /&gt;
:while &amp;lt;math&amp;gt; x &amp;lt; U &amp;lt;/math&amp;gt; do&lt;br /&gt;
:* choose &amp;lt;math&amp;gt;y \in (0,1)&amp;lt;/math&amp;gt; uniformly at random;&lt;br /&gt;
:* update &amp;lt;math&amp;gt;x = x + y&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;count = count + 1&amp;lt;/math&amp;gt;;&lt;br /&gt;
:return &amp;lt;math&amp;gt;count&amp;lt;/math&amp;gt;;&lt;br /&gt;
}}&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Random semicircle&amp;lt;/strong&amp;gt;] We sample &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; points within a circle &amp;lt;math&amp;gt;C=\{(x,y) \in \mathbb{R}^2 \mid x^2+y^2 \le 1\}&amp;lt;/math&amp;gt; independently and uniformly at random (i.e., the density function &amp;lt;math&amp;gt;f(x,y) \propto 1_{(x,y) \in C}&amp;lt;/math&amp;gt;). Find out the probability that they all lie within some semicircle of the original circle &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;. (Hint: you may apply the technique of change of variables, see [https://en.wikipedia.org/wiki/Random_variable#Functions_of_random_variables function of random variables] or Chapter 4.7 in [GS])&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Stochastic domination&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X, Y&amp;lt;/math&amp;gt; be continuous random variables. Show that &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; dominates &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; stochastically if and only if &amp;lt;math&amp;gt;\mathbb{E}[f(X)]\geq \mathbb{E}[f(Y)]&amp;lt;/math&amp;gt; for any non-decreasing function &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; for which the expectations exist.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Concentration of measure, 10 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Tossing coins&amp;lt;/strong&amp;gt;] We repeatedly toss a fair coin (with an equal probability of heads and tails). Let the random variable &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of throws required to obtain a total of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; heads. Show that &amp;lt;math&amp;gt;\textbf{Pr}[X &amp;gt; 2n + 2\sqrt{n\log n}]\leq O(1/n)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Chernoff vs Chebyshev&amp;lt;/strong&amp;gt;] We have a standard six-sided die. Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of times a 6 occurs in &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; throws off the die. Compare the best upper bounds on &amp;lt;math&amp;gt;\textbf{Pr}[X\geq n/4]&amp;lt;/math&amp;gt; that you can obtain using Chebyshev&#039;s inequality and Chernoff bounds.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;&amp;lt;strong&amp;gt;-th moment bound]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a random variable with expectation &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; such that moment generating function &amp;lt;math&amp;gt;\mathbf{E}[\exp(t|X|)]&amp;lt;/math&amp;gt; is finite for some &amp;lt;math&amp;gt; t &amp;gt; 0 &amp;lt;/math&amp;gt;. We can use the following two kinds of tail inequalities for &amp;lt;math&amp;gt; X &amp;lt;/math&amp;gt;: &lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&amp;lt;nowiki&amp;gt;     &amp;lt;/nowiki&amp;gt;&#039;&#039;&#039;&#039;&#039;Chernoff Bound&#039;&#039;&#039;&#039;&#039;&lt;br /&gt;
:&amp;lt;math&amp;gt;&lt;br /&gt;
\begin{align}&lt;br /&gt;
\mathbf{Pr}[|X| \geq \delta] \leq \min_{t \geq 0} \frac{\mathbf{E}[e^{t|X|}]}{e^{t\delta}}&lt;br /&gt;
\end{align}&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&#039;&#039;&#039;&#039;&#039;&amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;th-Moment Bound&#039;&#039;&#039;&#039;&#039;&lt;br /&gt;
:&amp;lt;math&amp;gt;&lt;br /&gt;
\begin{align}&lt;br /&gt;
\mathbf{Pr}[|X| \geq \delta] \leq \frac{\mathbf{E}[|X|^k]}{\delta^k}&lt;br /&gt;
\end{align}&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
# Show that for each &amp;lt;math&amp;gt;\delta&amp;lt;/math&amp;gt;, there exists a choice of &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; such that the &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;th-moment bound is no weaker than the Chernoff bound. (Hint: Use the probabilistic method.)&lt;br /&gt;
# Why would we still prefer the Chernoff bound to the (seemingly) stronger &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;-th moment bound?&lt;br /&gt;
&lt;br /&gt;
* [&#039;&#039;&#039;Cut size in random graph&#039;&#039;&#039;] Show that with probability at least &amp;lt;math&amp;gt;2/3&amp;lt;/math&amp;gt;, the size of the max-cut in [[wikipedia:Erdős–Rényi_model#Definition|Erdős–Rényi random graph]] &amp;lt;math&amp;gt;G(n,1/2)&amp;lt;/math&amp;gt; is at most &amp;lt;math&amp;gt;n^2/8 + O(n^{1.5})&amp;lt;/math&amp;gt;. In the &amp;lt;math&amp;gt;G(n,1/2)&amp;lt;/math&amp;gt; model, each edge is included in the graph with probability &amp;lt;math&amp;gt;1/2&amp;lt;/math&amp;gt;, independently of every other edge.&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Strong Law of Large Numbers, 15 points, &#039;&#039;&#039;Bonus Problem&#039;&#039;&#039;)==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
Throughout this problem, we assume &amp;lt;math&amp;gt;X_1,X_2,\ldots,&amp;lt;/math&amp;gt; be jointly independent square-integrable real random variables of mean zero. We will prove the strong law of large numbers by Kolmogorov maximal inequality.&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Kolmogorov maximal inequality&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;S_n = \sum_{i=1}^n X_i&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{Pr}\left(\max_{1 \le i \le n} |S_i| \ge t \right) \le \frac{n \mathbf{Var}(X_1)}{t^2}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Convergence of random series&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Var}(X_i) &amp;lt; \infty&amp;lt;/math&amp;gt;. Prove that the series &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} X_i&amp;lt;/math&amp;gt; is almost surely convergent.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Strong law of large numbers&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Prove the strong law of large numbers using previous propositions.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 4 (LLN and CLT, 15 points + 10 points) ==&lt;br /&gt;
&amp;lt;strong&amp;gt;In this problem, you may apply the results of Laws of Large Numbers (LLN) and the Central Limit Theorem (CLT) to solve the problems.&amp;lt;/strong&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;St. Petersburg paradox&amp;lt;/strong&amp;gt;] Consider the well-known game involving a fair coin. In this game, if it takes &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; tosses to obtain a head, you will win &amp;lt;math&amp;gt;2^k&amp;lt;/math&amp;gt; dollars as the reward. Despite the game&#039;s expected reward being infinite, people tend to offer relatively modest amounts to participate. The following provides a mathematical explanation for this phenomenon.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
For each &amp;lt;math&amp;gt;n \ge 1&amp;lt;/math&amp;gt;, let &amp;lt;math&amp;gt;X_{n,1}, X_{n,2},\ldots, X_{n,k}&amp;lt;/math&amp;gt; be independent random variables. Furthermore, let &amp;lt;math&amp;gt;b_n &amp;gt; 0&amp;lt;/math&amp;gt; be real numbers with &amp;lt;math&amp;gt;b_n \to \infty&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widetilde{X}_{n,k} = X_{n,k} \mathbf{1}_{|X_{n,k}| \le b_n}&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;1 \le k \le n&amp;lt;/math&amp;gt;. If &amp;lt;math&amp;gt;\sum_{k=1}^n \mathbf{Pr}(|X_{n,k}| &amp;gt; b_n) \to 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;b_n^{-2} \sum_{k=1}^n \mathbf{E}[\widetilde{X}_{n,k}^2] \to 0&amp;lt;/math&amp;gt; when &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, then &amp;lt;math&amp;gt;(S_n-a_n)/b_n \overset{P}{\to} 0 &amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;S_n = \sum_{k=1}^n X_{n,k}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;a_n = \sum_{k=1}^n \mathbf{E}[\widetilde{X}_{n,k}]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;S_n&amp;lt;/math&amp;gt; be the total winnings after playing &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds of the game. Prove that &amp;lt;math&amp;gt;\frac{S_n}{n \log_2 n} \overset{P}{\to} 1&amp;lt;/math&amp;gt;. (Therefore, a fair price to play this game &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; times is roughly &amp;lt;math&amp;gt;n \log_2 n&amp;lt;/math&amp;gt; dollars)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; (&amp;lt;strong&amp;gt;Bonus problem, 5 points&amp;lt;/strong&amp;gt;)&lt;br /&gt;
Let &amp;lt;math&amp;gt;S_n&amp;lt;/math&amp;gt; be the total winnings after playing &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds of the game. Prove that &amp;lt;math&amp;gt; \limsup_{n \to \infty} \frac{S_n}{n \log_2 n} = \infty&amp;lt;/math&amp;gt; almost surely. (Hint: You may use Borel-Cantelli lemmas)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; (&amp;lt;strong&amp;gt;Bonus problem, 5 points&amp;lt;/strong&amp;gt;)&lt;br /&gt;
Let &amp;lt;math&amp;gt;X_1,X_2,\ldots,X_n&amp;lt;/math&amp;gt; are i.i.d. copies of a random variable &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; with the standard Cauchy distribution (i.e., the probability density function of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; is &amp;lt;math&amp;gt;\frac{1}{\pi} \frac{1}{1+x^2}&amp;lt;/math&amp;gt;), show that &amp;lt;math&amp;gt;\frac{S_n}{n \log n}&amp;lt;/math&amp;gt; converges in probability to a constant &amp;lt;math&amp;gt;\frac{2}{\pi}&amp;lt;/math&amp;gt; but is almost surely unbounded, where &amp;lt;math&amp;gt;S_n = \sum_{i=1}^n |X_i|&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Monte Carlo integration&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; be a continuous function on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;U_1,U_2,\ldots,U_n \sim U(0,1)&amp;lt;/math&amp;gt; be independent random variables. Show that &amp;lt;math&amp;gt;I = \frac{1}{n}\sum_{i=1}^n f(U_i) \to \int_0^1 f(x) \mathrm{d}x&amp;lt;/math&amp;gt; in probability. (Remark: This holds so long as &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; is a measurable function on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;\int_0^1 |f(x)| \mathrm{d}x &amp;lt; \infty&amp;lt;/math&amp;gt;)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Stirling&#039;s formula&amp;lt;/strong&amp;gt;]&lt;br /&gt;
By considering the central limit theorem for the sum of independent Poisson-distributed random variables, show that&lt;br /&gt;
&amp;lt;math&amp;gt;n! \sim \sqrt{2\pi n} \left(\frac{n}{\mathrm{e}}\right)^n&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_4&amp;diff=13100</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 4</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_4&amp;diff=13100"/>
		<updated>2025-04-27T15:26:17Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: /* Problem 3 (Modes of Convergence and Characteristic Function, 15 points) (Bonus problem) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
*Bonus problem为附加题（选做）。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 4==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Continuous Random Variables, 30 points)==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Density function&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Determine the value of &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;f(x) = C\exp(-x-e^{-x}), x\in \mathbb{R}&amp;lt;/math&amp;gt; is a probability density function (PDF) for a continuous random variable.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Independence&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be independent and identically distributed continuous random variables with cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; and probability density function (PDF) &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt;. Find out the density functions of &amp;lt;math&amp;gt;V = \max\{X,Y\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;U = \min\{X,Y\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Correlation&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be uniformly distributed on &amp;lt;math&amp;gt;(-1,1)&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_k = \cos(k \pi X)&amp;lt;/math&amp;gt; for &amp;lt;math&amp;gt;k=1,2,\ldots,n&amp;lt;/math&amp;gt;. Are the random variables &amp;lt;math&amp;gt;Y_1, Y_2, \ldots, Y_n&amp;lt;/math&amp;gt; correlated? independent? You should prove your claim rigorously.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Expectation of random variables (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a continuous random variable with mean &amp;lt;math&amp;gt;\mu&amp;lt;/math&amp;gt; and cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X \ge 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\int_{0}^a F(x) dx = \int_{a}^{\infty} [1-F(x)] dx&amp;lt;/math&amp;gt; if and only if &amp;lt;math&amp;gt;a = \mu&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; has finite variance. Show that &amp;lt;math&amp;gt;g(a) = \mathbb{E}((X-a)^2)&amp;lt;/math&amp;gt; achieves the minimum when &amp;lt;math&amp;gt;a = \mu&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Expectation of random variables (II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X, Y&amp;lt;/math&amp;gt; be two independent and identically distributed continuous random variables with cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;. Furthermore, &amp;lt;math&amp;gt;X,Y \ge 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbb{E}[|X-Y|] = 2 \left(\mathbb{E}[X] - \int_{0}^{\infty} (1-F(x))^2 dx\right)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Conditional distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be two random variables. The joint density of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; is given by &amp;lt;math&amp;gt;f(x,y) = c(x^2 - y^2)e^{-x}&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;0\leq x &amp;lt;\infty&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-x\leq y \leq x&amp;lt;/math&amp;gt;. Here, &amp;lt;math&amp;gt;c\in \mathbb{R}_+&amp;lt;/math&amp;gt; is a constant. Find out the conditional distribution of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;, given &amp;lt;math&amp;gt;X = x&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Uniform Distribution (I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;P_i = (X_i,Y_i), 1\leq i\leq n&amp;lt;/math&amp;gt;, be independent, uniformly distributed points in the unit square &amp;lt;math&amp;gt;[0,1]^2&amp;lt;/math&amp;gt;. A point &amp;lt;math&amp;gt;P_i&amp;lt;/math&amp;gt; is called &amp;quot;peripheral&amp;quot; if, for all &amp;lt;math&amp;gt;r = 1,2,\cdots,n&amp;lt;/math&amp;gt;, either &amp;lt;math&amp;gt;X_r \leq X_i&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;Y_r \leq Y_i&amp;lt;/math&amp;gt;, or both. Find out the expected number of peripheral points.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Uniform Distribution (II)&amp;lt;/strong&amp;gt;] Derive the moment generating function of the standard uniform distribution, i.e., uniform distribution on &amp;lt;math&amp;gt;(0,1)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Exponential distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; have an exponential distribution. Show that &amp;lt;math&amp;gt;\textbf{Pr}[X&amp;gt;s+x|X&amp;gt;s] = \textbf{Pr}[X&amp;gt;x]&amp;lt;/math&amp;gt;, for &amp;lt;math&amp;gt;x,s\geq 0&amp;lt;/math&amp;gt;. This is the memoryless property. Show that the exponential distribution is the only continuous distribution with this property.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Normal distribution(I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X,Y\sim N(0,1)&amp;lt;/math&amp;gt; be two independent and identically distributed normal random variables. Let &amp;lt;math&amp;gt;Z = X-Y&amp;lt;/math&amp;gt;. Find the density function of &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;|Z|&amp;lt;/math&amp;gt; respectively.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Normal distribution(II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; have the &amp;lt;math&amp;gt;N(0,1)&amp;lt;/math&amp;gt; distribution and let &amp;lt;math&amp;gt;a&amp;gt;0&amp;lt;/math&amp;gt;. Show that the random variable &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; given by&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{equation*}&lt;br /&gt;
Y = \begin{cases}&lt;br /&gt;
X, &amp;amp; |X|&amp;lt; a \\&lt;br /&gt;
-X, &amp;amp; |X|\geq a&lt;br /&gt;
\end{cases}&lt;br /&gt;
\end{equation*}&amp;lt;/math&amp;gt;&lt;br /&gt;
has the &amp;lt;math&amp;gt;N(0,1)&amp;lt;/math&amp;gt; distribution, and find an expression for &amp;lt;math&amp;gt;\rho(a) = \textbf{Cov}(X,Y)&amp;lt;/math&amp;gt; in terms of the density function &amp;lt;math&amp;gt;\phi&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random process&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Given a real number &amp;lt;math&amp;gt;U&amp;lt;1&amp;lt;/math&amp;gt; as input of the following process, find out the expected returning value.&lt;br /&gt;
{{Theorem|&#039;&#039;Process&#039;&#039;|&lt;br /&gt;
:&#039;&#039;&#039;Input:&#039;&#039;&#039;  real numbers &amp;lt;math&amp;gt;U &amp;lt; 1&amp;lt;/math&amp;gt;;&lt;br /&gt;
----&lt;br /&gt;
:initialize &amp;lt;math&amp;gt;x = 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;count = 0&amp;lt;/math&amp;gt;;&lt;br /&gt;
:while &amp;lt;math&amp;gt; x &amp;lt; U &amp;lt;/math&amp;gt; do&lt;br /&gt;
:* choose &amp;lt;math&amp;gt;y \in (0,1)&amp;lt;/math&amp;gt; uniformly at random;&lt;br /&gt;
:* update &amp;lt;math&amp;gt;x = x + y&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;count = count + 1&amp;lt;/math&amp;gt;;&lt;br /&gt;
:return &amp;lt;math&amp;gt;count&amp;lt;/math&amp;gt;;&lt;br /&gt;
}}&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Random semicircle&amp;lt;/strong&amp;gt;] We sample &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; points within a circle &amp;lt;math&amp;gt;C=\{(x,y) \in \mathbb{R}^2 \mid x^2+y^2 \le 1\}&amp;lt;/math&amp;gt; independently and uniformly at random (i.e., the density function &amp;lt;math&amp;gt;f(x,y) \propto 1_{(x,y) \in C}&amp;lt;/math&amp;gt;). Find out the probability that they all lie within some semicircle of the original circle &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;. (Hint: you may apply the technique of change of variables, see [https://en.wikipedia.org/wiki/Random_variable#Functions_of_random_variables function of random variables] or Chapter 4.7 in [GS])&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Stochastic domination&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X, Y&amp;lt;/math&amp;gt; be continuous random variables. Show that &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; dominates &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; stochastically if and only if &amp;lt;math&amp;gt;\mathbb{E}[f(X)]\geq \mathbb{E}[f(Y)]&amp;lt;/math&amp;gt; for any non-decreasing function &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; for which the expectations exist.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Concentration of measure, 10 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Tossing coins&amp;lt;/strong&amp;gt;] We repeatedly toss a fair coin (with an equal probability of heads and tails). Let the random variable &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of throws required to obtain a total of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; heads. Show that &amp;lt;math&amp;gt;\textbf{Pr}[X &amp;gt; 2n + 2\sqrt{n\log n}]\leq O(1/n)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Chernoff vs Chebyshev&amp;lt;/strong&amp;gt;] We have a standard six-sided die. Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of times a 6 occurs in &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; throws off the die. Compare the best upper bounds on &amp;lt;math&amp;gt;\textbf{Pr}[X\geq n/4]&amp;lt;/math&amp;gt; that you can obtain using Chebyshev&#039;s inequality and Chernoff bounds.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;&amp;lt;strong&amp;gt;-th moment bound]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a random variable with expectation &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; such that moment generating function &amp;lt;math&amp;gt;\mathbf{E}[\exp(t|X|)]&amp;lt;/math&amp;gt; is finite for some &amp;lt;math&amp;gt; t &amp;gt; 0 &amp;lt;/math&amp;gt;. We can use the following two kinds of tail inequalities for &amp;lt;math&amp;gt; X &amp;lt;/math&amp;gt;: &lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&amp;lt;nowiki&amp;gt;     &amp;lt;/nowiki&amp;gt;&#039;&#039;&#039;&#039;&#039;Chernoff Bound&#039;&#039;&#039;&#039;&#039;&lt;br /&gt;
:&amp;lt;math&amp;gt;&lt;br /&gt;
\begin{align}&lt;br /&gt;
\mathbf{Pr}[|X| \geq \delta] \leq \min_{t \geq 0} \frac{\mathbf{E}[e^{t|X|}]}{e^{t\delta}}&lt;br /&gt;
\end{align}&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&#039;&#039;&#039;&#039;&#039;&amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;th-Moment Bound&#039;&#039;&#039;&#039;&#039;&lt;br /&gt;
:&amp;lt;math&amp;gt;&lt;br /&gt;
\begin{align}&lt;br /&gt;
\mathbf{Pr}[|X| \geq \delta] \leq \frac{\mathbf{E}[|X|^k]}{\delta^k}&lt;br /&gt;
\end{align}&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
# Show that for each &amp;lt;math&amp;gt;\delta&amp;lt;/math&amp;gt;, there exists a choice of &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; such that the &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;th-moment bound is no weaker than the Chernoff bound. (Hint: Use the probabilistic method.)&lt;br /&gt;
# Why would we still prefer the Chernoff bound to the (seemingly) stronger &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;-th moment bound?&lt;br /&gt;
&lt;br /&gt;
* [&#039;&#039;&#039;Cut size in random graph&#039;&#039;&#039;] Show that with probability at least &amp;lt;math&amp;gt;2/3&amp;lt;/math&amp;gt;, the size of the max-cut in [[wikipedia:Erdős–Rényi_model#Definition|Erdős–Rényi random graph]] &amp;lt;math&amp;gt;G(n,1/2)&amp;lt;/math&amp;gt; is at most &amp;lt;math&amp;gt;n^2/8 + O(n^{1.5})&amp;lt;/math&amp;gt;. In the &amp;lt;math&amp;gt;G(n,1/2)&amp;lt;/math&amp;gt; model, each edge is included in the graph with probability &amp;lt;math&amp;gt;1/2&amp;lt;/math&amp;gt;, independently of every other edge.&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Strong Law of Large Numbers, 15 points, &#039;&#039;&#039;Bonus Problem&#039;&#039;&#039;)==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
Throughout this problem, we assume &amp;lt;math&amp;gt;X_1,X_2,\ldots,&amp;lt;/math&amp;gt; be jointly independent square-integrable real random variables of mean zero. We will prove the strong law of large numbers by Kolmogorov maximal inequality.&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Kolmogorov maximal inequality&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;S_n = \sum_{i=1}^n X_i&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{Pr}\left(\max_{1 \le i \le n} |S_i| \ge t \right) \le \frac{n \mathbf{Var}(X_1)}{t^2}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Convergence of random series&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Var}(X_i) &amp;lt; \infty&amp;lt;/math&amp;gt;. Prove that the series &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} X_i&amp;lt;/math&amp;gt; is almost surely convergent.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Strong law of large numbers&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Prove the strong law of large numbers using previous propositions.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 4 (LLN and CLT, 15 points + 10 points) ==&lt;br /&gt;
&amp;lt;strong&amp;gt;In this problem, you may apply the results of Laws of Large Numbers (LLN) and the Central Limit Theorem (CLT) to solve the problems.&amp;lt;/strong&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;St. Petersburg paradox&amp;lt;/strong&amp;gt;] Consider the well-known game involving a fair coin. In this game, if it takes &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; tosses to obtain a head, you will win &amp;lt;math&amp;gt;2^k&amp;lt;/math&amp;gt; dollars as the reward. Despite the game&#039;s expected reward being infinite, people tend to offer relatively modest amounts to participate. The following provides a mathematical explanation for this phenomenon.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
For each &amp;lt;math&amp;gt;n \ge 1&amp;lt;/math&amp;gt;, let &amp;lt;math&amp;gt;X_{n,1}, X_{n,2},\ldots, X_{n,k}&amp;lt;/math&amp;gt; be independent random variables. Furthermore, let &amp;lt;math&amp;gt;b_n &amp;gt; 0&amp;lt;/math&amp;gt; be real numbers with &amp;lt;math&amp;gt;b_n \to \infty&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widetilde{X}_{n,k} = X_{n,k} \mathbf{1}_{|X_{n,k}| \le b_n}&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;1 \le k \le n&amp;lt;/math&amp;gt;. If &amp;lt;math&amp;gt;\sum_{k=1}^n \mathbf{Pr}(|X_{n,k}| &amp;gt; b_n) \to 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;b_n^{-2} \sum_{k=1}^n \mathbf{E}[\widetilde{X}_{n,k}^2] \to 0&amp;lt;/math&amp;gt; when &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, then &amp;lt;math&amp;gt;(S_n-a_n)/b_n \overset{P}{\to} 0 &amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;S_n = \sum_{k=1}^n X_{n,k}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;a_n = \sum_{k=1}^n \mathbf{E}[\widetilde{X}_{n,k}]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;S_n&amp;lt;/math&amp;gt; be the total winnings after playing &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds of the game. Prove that &amp;lt;math&amp;gt;\frac{S_n}{n \log_2 n} \overset{P}{\to} 1&amp;lt;/math&amp;gt;. (Therefore, a fair price to play this game &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; times is roughly &amp;lt;math&amp;gt;n \log_2 n&amp;lt;/math&amp;gt; dollars)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; (&amp;lt;strong&amp;gt;Bonus problem, 5 points&amp;lt;/strong&amp;gt;)&lt;br /&gt;
Let &amp;lt;math&amp;gt;S_n&amp;lt;/math&amp;gt; be the total winnings after playing &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds of the game. Prove that &amp;lt;math&amp;gt; \limsup_{n \to \infty} \frac{S_n}{n \log_2 n} = \infty&amp;lt;/math&amp;gt; almost surely. (Hint: You may use Borel-Cantelli lemmas)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; (&amp;lt;strong&amp;gt;Bonus problem, 5 points&amp;lt;/strong&amp;gt;)&lt;br /&gt;
Let &amp;lt;math&amp;gt;X_1,X_2,\ldots,X_n&amp;lt;/math&amp;gt; are i.i.d. copies of a random variable &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; with the standard Cauchy distribution (i.e., the probability density function of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; is &amp;lt;math&amp;gt;\frac{1}{\pi} \frac{1}{1+x^2}&amp;lt;/math&amp;gt;), show that &amp;lt;math&amp;gt;\frac{S_n}{n \log n}&amp;lt;/math&amp;gt; converges in probability to a constant &amp;lt;math&amp;gt;\frac{2}{\pi}&amp;lt;/math&amp;gt; but is almost surely unbounded, where &amp;lt;math&amp;gt;S_n = \sum_{i=1}^n |X_i|&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Monte Carlo Integration&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; be a continuous function on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;U_1,U_2,\ldots,U_n \sim U(0,1)&amp;lt;/math&amp;gt; be independent random variables. Show that &amp;lt;math&amp;gt;I = \frac{1}{n}\sum_{i=1}^n f(U_i) \to \int_0^1 f(x) \mathrm{d}x&amp;lt;/math&amp;gt; in probability. (Remark: This holds so long as &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; is a measurable function on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;\int_0^1 |f(x)| \mathrm{d}x &amp;lt; \infty&amp;lt;/math&amp;gt;)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Square root&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X_1,X_2,\ldots&amp;lt;/math&amp;gt; be i.i.d. random variables with &amp;lt;math&amp;gt;X_i \ge 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{E}[X_1] = 1&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{Var}[X_1] = \sigma^2 \in (0,+\infty)&amp;lt;/math&amp;gt;. Show &amp;lt;math&amp;gt;2(\sqrt{S_n} - \sqrt{n}) \overset{D}{\to} \sigma N(0,1)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;S_n = \sum_{i=1}^n X_i&amp;lt;/math&amp;gt;. (Hint: You may use the statement in Problem 3.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_4&amp;diff=13099</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 4</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_4&amp;diff=13099"/>
		<updated>2025-04-27T15:11:31Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: /* Problem 3 (Modes of Convergence and Characteristic Function, 15 points) (Bonus problem) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
*Bonus problem为附加题（选做）。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 4==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Continuous Random Variables, 30 points)==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Density function&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Determine the value of &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;f(x) = C\exp(-x-e^{-x}), x\in \mathbb{R}&amp;lt;/math&amp;gt; is a probability density function (PDF) for a continuous random variable.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Independence&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be independent and identically distributed continuous random variables with cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; and probability density function (PDF) &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt;. Find out the density functions of &amp;lt;math&amp;gt;V = \max\{X,Y\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;U = \min\{X,Y\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Correlation&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be uniformly distributed on &amp;lt;math&amp;gt;(-1,1)&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_k = \cos(k \pi X)&amp;lt;/math&amp;gt; for &amp;lt;math&amp;gt;k=1,2,\ldots,n&amp;lt;/math&amp;gt;. Are the random variables &amp;lt;math&amp;gt;Y_1, Y_2, \ldots, Y_n&amp;lt;/math&amp;gt; correlated? independent? You should prove your claim rigorously.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Expectation of random variables (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a continuous random variable with mean &amp;lt;math&amp;gt;\mu&amp;lt;/math&amp;gt; and cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X \ge 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\int_{0}^a F(x) dx = \int_{a}^{\infty} [1-F(x)] dx&amp;lt;/math&amp;gt; if and only if &amp;lt;math&amp;gt;a = \mu&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; has finite variance. Show that &amp;lt;math&amp;gt;g(a) = \mathbb{E}((X-a)^2)&amp;lt;/math&amp;gt; achieves the minimum when &amp;lt;math&amp;gt;a = \mu&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Expectation of random variables (II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X, Y&amp;lt;/math&amp;gt; be two independent and identically distributed continuous random variables with cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;. Furthermore, &amp;lt;math&amp;gt;X,Y \ge 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbb{E}[|X-Y|] = 2 \left(\mathbb{E}[X] - \int_{0}^{\infty} (1-F(x))^2 dx\right)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Conditional distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be two random variables. The joint density of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; is given by &amp;lt;math&amp;gt;f(x,y) = c(x^2 - y^2)e^{-x}&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;0\leq x &amp;lt;\infty&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-x\leq y \leq x&amp;lt;/math&amp;gt;. Here, &amp;lt;math&amp;gt;c\in \mathbb{R}_+&amp;lt;/math&amp;gt; is a constant. Find out the conditional distribution of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;, given &amp;lt;math&amp;gt;X = x&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Uniform Distribution (I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;P_i = (X_i,Y_i), 1\leq i\leq n&amp;lt;/math&amp;gt;, be independent, uniformly distributed points in the unit square &amp;lt;math&amp;gt;[0,1]^2&amp;lt;/math&amp;gt;. A point &amp;lt;math&amp;gt;P_i&amp;lt;/math&amp;gt; is called &amp;quot;peripheral&amp;quot; if, for all &amp;lt;math&amp;gt;r = 1,2,\cdots,n&amp;lt;/math&amp;gt;, either &amp;lt;math&amp;gt;X_r \leq X_i&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;Y_r \leq Y_i&amp;lt;/math&amp;gt;, or both. Find out the expected number of peripheral points.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Uniform Distribution (II)&amp;lt;/strong&amp;gt;] Derive the moment generating function of the standard uniform distribution, i.e., uniform distribution on &amp;lt;math&amp;gt;(0,1)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Exponential distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; have an exponential distribution. Show that &amp;lt;math&amp;gt;\textbf{Pr}[X&amp;gt;s+x|X&amp;gt;s] = \textbf{Pr}[X&amp;gt;x]&amp;lt;/math&amp;gt;, for &amp;lt;math&amp;gt;x,s\geq 0&amp;lt;/math&amp;gt;. This is the memoryless property. Show that the exponential distribution is the only continuous distribution with this property.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Normal distribution(I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X,Y\sim N(0,1)&amp;lt;/math&amp;gt; be two independent and identically distributed normal random variables. Let &amp;lt;math&amp;gt;Z = X-Y&amp;lt;/math&amp;gt;. Find the density function of &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;|Z|&amp;lt;/math&amp;gt; respectively.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Normal distribution(II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; have the &amp;lt;math&amp;gt;N(0,1)&amp;lt;/math&amp;gt; distribution and let &amp;lt;math&amp;gt;a&amp;gt;0&amp;lt;/math&amp;gt;. Show that the random variable &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; given by&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{equation*}&lt;br /&gt;
Y = \begin{cases}&lt;br /&gt;
X, &amp;amp; |X|&amp;lt; a \\&lt;br /&gt;
-X, &amp;amp; |X|\geq a&lt;br /&gt;
\end{cases}&lt;br /&gt;
\end{equation*}&amp;lt;/math&amp;gt;&lt;br /&gt;
has the &amp;lt;math&amp;gt;N(0,1)&amp;lt;/math&amp;gt; distribution, and find an expression for &amp;lt;math&amp;gt;\rho(a) = \textbf{Cov}(X,Y)&amp;lt;/math&amp;gt; in terms of the density function &amp;lt;math&amp;gt;\phi&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random process&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Given a real number &amp;lt;math&amp;gt;U&amp;lt;1&amp;lt;/math&amp;gt; as input of the following process, find out the expected returning value.&lt;br /&gt;
{{Theorem|&#039;&#039;Process&#039;&#039;|&lt;br /&gt;
:&#039;&#039;&#039;Input:&#039;&#039;&#039;  real numbers &amp;lt;math&amp;gt;U &amp;lt; 1&amp;lt;/math&amp;gt;;&lt;br /&gt;
----&lt;br /&gt;
:initialize &amp;lt;math&amp;gt;x = 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;count = 0&amp;lt;/math&amp;gt;;&lt;br /&gt;
:while &amp;lt;math&amp;gt; x &amp;lt; U &amp;lt;/math&amp;gt; do&lt;br /&gt;
:* choose &amp;lt;math&amp;gt;y \in (0,1)&amp;lt;/math&amp;gt; uniformly at random;&lt;br /&gt;
:* update &amp;lt;math&amp;gt;x = x + y&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;count = count + 1&amp;lt;/math&amp;gt;;&lt;br /&gt;
:return &amp;lt;math&amp;gt;count&amp;lt;/math&amp;gt;;&lt;br /&gt;
}}&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Random semicircle&amp;lt;/strong&amp;gt;] We sample &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; points within a circle &amp;lt;math&amp;gt;C=\{(x,y) \in \mathbb{R}^2 \mid x^2+y^2 \le 1\}&amp;lt;/math&amp;gt; independently and uniformly at random (i.e., the density function &amp;lt;math&amp;gt;f(x,y) \propto 1_{(x,y) \in C}&amp;lt;/math&amp;gt;). Find out the probability that they all lie within some semicircle of the original circle &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;. (Hint: you may apply the technique of change of variables, see [https://en.wikipedia.org/wiki/Random_variable#Functions_of_random_variables function of random variables] or Chapter 4.7 in [GS])&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Stochastic domination&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X, Y&amp;lt;/math&amp;gt; be continuous random variables. Show that &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; dominates &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; stochastically if and only if &amp;lt;math&amp;gt;\mathbb{E}[f(X)]\geq \mathbb{E}[f(Y)]&amp;lt;/math&amp;gt; for any non-decreasing function &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; for which the expectations exist.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Concentration of measure, 10 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Tossing coins&amp;lt;/strong&amp;gt;] We repeatedly toss a fair coin (with an equal probability of heads and tails). Let the random variable &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of throws required to obtain a total of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; heads. Show that &amp;lt;math&amp;gt;\textbf{Pr}[X &amp;gt; 2n + 2\sqrt{n\log n}]\leq O(1/n)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Chernoff vs Chebyshev&amp;lt;/strong&amp;gt;] We have a standard six-sided die. Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of times a 6 occurs in &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; throws off the die. Compare the best upper bounds on &amp;lt;math&amp;gt;\textbf{Pr}[X\geq n/4]&amp;lt;/math&amp;gt; that you can obtain using Chebyshev&#039;s inequality and Chernoff bounds.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;&amp;lt;strong&amp;gt;-th moment bound]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a random variable with expectation &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; such that moment generating function &amp;lt;math&amp;gt;\mathbf{E}[\exp(t|X|)]&amp;lt;/math&amp;gt; is finite for some &amp;lt;math&amp;gt; t &amp;gt; 0 &amp;lt;/math&amp;gt;. We can use the following two kinds of tail inequalities for &amp;lt;math&amp;gt; X &amp;lt;/math&amp;gt;: &lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&amp;lt;nowiki&amp;gt;     &amp;lt;/nowiki&amp;gt;&#039;&#039;&#039;&#039;&#039;Chernoff Bound&#039;&#039;&#039;&#039;&#039;&lt;br /&gt;
:&amp;lt;math&amp;gt;&lt;br /&gt;
\begin{align}&lt;br /&gt;
\mathbf{Pr}[|X| \geq \delta] \leq \min_{t \geq 0} \frac{\mathbf{E}[e^{t|X|}]}{e^{t\delta}}&lt;br /&gt;
\end{align}&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&#039;&#039;&#039;&#039;&#039;&amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;th-Moment Bound&#039;&#039;&#039;&#039;&#039;&lt;br /&gt;
:&amp;lt;math&amp;gt;&lt;br /&gt;
\begin{align}&lt;br /&gt;
\mathbf{Pr}[|X| \geq \delta] \leq \frac{\mathbf{E}[|X|^k]}{\delta^k}&lt;br /&gt;
\end{align}&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
# Show that for each &amp;lt;math&amp;gt;\delta&amp;lt;/math&amp;gt;, there exists a choice of &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; such that the &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;th-moment bound is no weaker than the Chernoff bound. (Hint: Use the probabilistic method.)&lt;br /&gt;
# Why would we still prefer the Chernoff bound to the (seemingly) stronger &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;-th moment bound?&lt;br /&gt;
&lt;br /&gt;
* [&#039;&#039;&#039;Cut size in random graph&#039;&#039;&#039;] Show that with probability at least &amp;lt;math&amp;gt;2/3&amp;lt;/math&amp;gt;, the size of the max-cut in [[wikipedia:Erdős–Rényi_model#Definition|Erdős–Rényi random graph]] &amp;lt;math&amp;gt;G(n,1/2)&amp;lt;/math&amp;gt; is at most &amp;lt;math&amp;gt;n^2/8 + O(n^{1.5})&amp;lt;/math&amp;gt;. In the &amp;lt;math&amp;gt;G(n,1/2)&amp;lt;/math&amp;gt; model, each edge is included in the graph with probability &amp;lt;math&amp;gt;1/2&amp;lt;/math&amp;gt;, independently of every other edge.&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Modes of Convergence and Characteristic Function, 15 points) (&amp;lt;strong&amp;gt;Bonus problem&amp;lt;/strong&amp;gt;)==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Connection of convergence modes (I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;(X_n)_{n \ge 1}, (Y_n)_{n \ge 1}, X, Y&amp;lt;/math&amp;gt; be random variables and &amp;lt;math&amp;gt;c\in\mathbb{R}&amp;lt;/math&amp;gt; be a real number. &lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; Suppose &amp;lt;math&amp;gt;X_n \overset{D}{\to} X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_n \overset{D}{\to} c&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;X_nY_n \overset{D}{\to} cX&amp;lt;/math&amp;gt;. &lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Construct an example such that &amp;lt;math&amp;gt;X_n \overset{D}{\to} X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_n \overset{D}{\to} Y&amp;lt;/math&amp;gt; but &amp;lt;math&amp;gt;X_nY_n&amp;lt;/math&amp;gt; does not converge to &amp;lt;math&amp;gt;XY&amp;lt;/math&amp;gt; in distribution.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Connection of convergence modes (II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;(X_n)_{n \ge 1}, X&amp;lt;/math&amp;gt; be random variables. Prove that &amp;lt;math&amp;gt;X_n \overset{P}{\to} X&amp;lt;/math&amp;gt; if and only if for every subsequence &amp;lt;math&amp;gt;X_{n(m)}&amp;lt;/math&amp;gt;, there exists a further subsequence &amp;lt;math&amp;gt;Y_k = X_{n(m_k)}&amp;lt;/math&amp;gt; that converges almost surely to &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;. (Hint: you may use the first Borel-Cantelli lemma.)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Convergence and Independence&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X_1,X_2,\ldots,&amp;lt;/math&amp;gt; be a sequence of scalar random variables converging in probability to another random variable &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;. Suppose there is a random variable that is independent of &amp;lt;math&amp;gt;X_i&amp;lt;/math&amp;gt; for each individual &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; is also independent of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 4 (LLN and CLT, 15 points + 10 points) ==&lt;br /&gt;
&amp;lt;strong&amp;gt;In this problem, you may apply the results of Laws of Large Numbers (LLN) and the Central Limit Theorem (CLT) to solve the problems.&amp;lt;/strong&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;St. Petersburg paradox&amp;lt;/strong&amp;gt;] Consider the well-known game involving a fair coin. In this game, if it takes &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; tosses to obtain a head, you will win &amp;lt;math&amp;gt;2^k&amp;lt;/math&amp;gt; dollars as the reward. Despite the game&#039;s expected reward being infinite, people tend to offer relatively modest amounts to participate. The following provides a mathematical explanation for this phenomenon.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
For each &amp;lt;math&amp;gt;n \ge 1&amp;lt;/math&amp;gt;, let &amp;lt;math&amp;gt;X_{n,1}, X_{n,2},\ldots, X_{n,k}&amp;lt;/math&amp;gt; be independent random variables. Furthermore, let &amp;lt;math&amp;gt;b_n &amp;gt; 0&amp;lt;/math&amp;gt; be real numbers with &amp;lt;math&amp;gt;b_n \to \infty&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widetilde{X}_{n,k} = X_{n,k} \mathbf{1}_{|X_{n,k}| \le b_n}&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;1 \le k \le n&amp;lt;/math&amp;gt;. If &amp;lt;math&amp;gt;\sum_{k=1}^n \mathbf{Pr}(|X_{n,k}| &amp;gt; b_n) \to 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;b_n^{-2} \sum_{k=1}^n \mathbf{E}[\widetilde{X}_{n,k}^2] \to 0&amp;lt;/math&amp;gt; when &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, then &amp;lt;math&amp;gt;(S_n-a_n)/b_n \overset{P}{\to} 0 &amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;S_n = \sum_{k=1}^n X_{n,k}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;a_n = \sum_{k=1}^n \mathbf{E}[\widetilde{X}_{n,k}]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;S_n&amp;lt;/math&amp;gt; be the total winnings after playing &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds of the game. Prove that &amp;lt;math&amp;gt;\frac{S_n}{n \log_2 n} \overset{P}{\to} 1&amp;lt;/math&amp;gt;. (Therefore, a fair price to play this game &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; times is roughly &amp;lt;math&amp;gt;n \log_2 n&amp;lt;/math&amp;gt; dollars)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; (&amp;lt;strong&amp;gt;Bonus problem, 5 points&amp;lt;/strong&amp;gt;)&lt;br /&gt;
Let &amp;lt;math&amp;gt;S_n&amp;lt;/math&amp;gt; be the total winnings after playing &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds of the game. Prove that &amp;lt;math&amp;gt; \limsup_{n \to \infty} \frac{S_n}{n \log_2 n} = \infty&amp;lt;/math&amp;gt; almost surely. (Hint: You may use Borel-Cantelli lemmas)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; (&amp;lt;strong&amp;gt;Bonus problem, 5 points&amp;lt;/strong&amp;gt;)&lt;br /&gt;
Let &amp;lt;math&amp;gt;X_1,X_2,\ldots,X_n&amp;lt;/math&amp;gt; are i.i.d. copies of a random variable &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; with the standard Cauchy distribution (i.e., the probability density function of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; is &amp;lt;math&amp;gt;\frac{1}{\pi} \frac{1}{1+x^2}&amp;lt;/math&amp;gt;), show that &amp;lt;math&amp;gt;\frac{S_n}{n \log n}&amp;lt;/math&amp;gt; converges in probability to a constant &amp;lt;math&amp;gt;\frac{2}{\pi}&amp;lt;/math&amp;gt; but is almost surely unbounded, where &amp;lt;math&amp;gt;S_n = \sum_{i=1}^n |X_i|&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Monte Carlo Integration&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; be a continuous function on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;U_1,U_2,\ldots,U_n \sim U(0,1)&amp;lt;/math&amp;gt; be independent random variables. Show that &amp;lt;math&amp;gt;I = \frac{1}{n}\sum_{i=1}^n f(U_i) \to \int_0^1 f(x) \mathrm{d}x&amp;lt;/math&amp;gt; in probability. (Remark: This holds so long as &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; is a measurable function on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;\int_0^1 |f(x)| \mathrm{d}x &amp;lt; \infty&amp;lt;/math&amp;gt;)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Square root&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X_1,X_2,\ldots&amp;lt;/math&amp;gt; be i.i.d. random variables with &amp;lt;math&amp;gt;X_i \ge 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{E}[X_1] = 1&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{Var}[X_1] = \sigma^2 \in (0,+\infty)&amp;lt;/math&amp;gt;. Show &amp;lt;math&amp;gt;2(\sqrt{S_n} - \sqrt{n}) \overset{D}{\to} \sigma N(0,1)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;S_n = \sum_{i=1}^n X_i&amp;lt;/math&amp;gt;. (Hint: You may use the statement in Problem 3.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_3&amp;diff=13056</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 3</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_3&amp;diff=13056"/>
		<updated>2025-04-07T07:04:01Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: /* Problem 3 (Probability meets distinct sums) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 3==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Warm-up Problems) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
      Let &amp;lt;math&amp;gt;X_1,X_2,...,X_n&amp;lt;/math&amp;gt; be independent random variables, and suppose that &amp;lt;math&amp;gt;X_k&amp;lt;/math&amp;gt; is Bernoulli with parameter &amp;lt;math&amp;gt;p_k&amp;lt;/math&amp;gt;. Let &amp;lt;math&amp;gt;Y= X_1 + X_2 + \dots + X_n&amp;lt;/math&amp;gt;. Show that, for &amp;lt;math&amp;gt;\mathbb E[Y]&amp;lt;/math&amp;gt; fixed, &amp;lt;math&amp;gt;\mathrm{Var}(Y )&amp;lt;/math&amp;gt; is a maximized when &amp;lt;math&amp;gt;p_1 = p_2 = \dots = p_n&amp;lt;/math&amp;gt;. That is to say, the variation in the sum is greatest when individuals are most alike.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Each member of a group of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; players rolls a (fair) 6-sided die. For any pair of players who throw the same number, the group scores &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; point. Find the mean and variance of the total score of the group.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        An urn contains &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; balls numbered &amp;lt;math&amp;gt;1, 2, \ldots, n&amp;lt;/math&amp;gt;. We select &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; balls uniformly at random &amp;lt;strong&amp;gt;without replacement&amp;lt;/strong&amp;gt; and add up their numbers. Find the mean and variance of the sum.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (IV)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
      Let &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; be an integer-valued, positive random variable and let &amp;lt;math&amp;gt;\{X_i\}_{i=1}^{\infty}&amp;lt;/math&amp;gt; be indepedently identically distributed random variables that are independent of &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt;, too. &lt;br /&gt;
Precisely, for any finite subset &amp;lt;math&amp;gt;I \subseteq\mathbb{N}_+&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\{X_i\}_{i \in I}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; are mutually independent. Let &amp;lt;math&amp;gt;X = \sum_{i=1}^N X_i&amp;lt;/math&amp;gt;, show that &amp;lt;math&amp;gt;\textbf{Var}[X] = \textbf{Var}[X_1] \mathbb{E}[N] + \mathbb{E}[X_1]^2 \textbf{Var}[N]&amp;lt;/math&amp;gt;.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Moments (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Show that [math]G(t) = \frac{e^t}{4} + \frac{e^{-t}}{2} + \frac{1}{4}[/math] is a moment-generating function of a random variable, and write the probability mass function of this random variable.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Moments (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X\sim \text{Geo}(p)&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt;. Find &amp;lt;math&amp;gt;\mathbb{E}[X^3]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbb{E}[X^4]&amp;lt;/math&amp;gt;. &lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Moments (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X\sim \text{Pois}(\lambda)&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;\lambda &amp;gt;0 &amp;lt;/math&amp;gt;. Find &amp;lt;math&amp;gt;\mathbb{E}[X^3]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbb{E}[X^4]&amp;lt;/math&amp;gt;. &lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be discrete random variables with correlation &amp;lt;math&amp;gt;\rho&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;|\rho|= 1&amp;lt;/math&amp;gt; if and only if &amp;lt;math&amp;gt;X=aY+b&amp;lt;/math&amp;gt; for some real numbers &amp;lt;math&amp;gt;a,b&amp;lt;/math&amp;gt;.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
    Let [math]X[/math] and [math]Y[/math] be discrete random variables with mean &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;, variance &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;, and correlation [math]\rho[/math]. Show that [math]\mathbb{E}(\max\{X^2,Y^2\})\leq 1+\sqrt{1-\rho^2}[/math]. (Hint: use the identity [math]\max\{a,b\} = \frac{1}{2}(a+b+|a-b|)[/math].)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
   Let [math]X[/math] and [math]Y[/math] be independent Bernoulli random variables with parameter [math]1/2[/math]. Show that [math]X+Y[/math] and [math]|X-Y|[/math] are dependent though uncorrelated.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Inequalities) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Reverse Markov&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable with bounded range &amp;lt;math&amp;gt;0 \le X \le U&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;U &amp;gt; 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbf{Pr}(X \le a) \le \frac{U-\mathbf{E}[X]}{U-a}&amp;lt;/math&amp;gt; for any &amp;lt;math&amp;gt;0 &amp;lt; a &amp;lt; U&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Markov&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable. Show that for all &amp;lt;math&amp;gt;\beta \geq 0&amp;lt;/math&amp;gt; and all &amp;lt;math&amp;gt;x &amp;gt; 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}(X\geq x)\leq \mathbb{E}(e^{\beta X})e^{-\beta x}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Cantelli&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable with mean &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; and variance &amp;lt;math&amp;gt;\sigma^2&amp;lt;/math&amp;gt;. Prove that for any &amp;lt;math&amp;gt;\lambda &amp;gt; 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}[X \ge \lambda] \le \frac{\sigma^2}{\lambda^2+\sigma^2}&amp;lt;/math&amp;gt;. (Hint: You may first show that &amp;lt;math&amp;gt;\mathbf{Pr}[X \ge \lambda] \le \frac{\sigma^2 + u^2}{(\lambda + u)^2}&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;u &amp;gt; 0&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; &amp;lt;strong&amp;gt;[Union of events]&amp;lt;/strong&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;A_1,A_2,\ldots,A_n&amp;lt;/math&amp;gt; be events with &amp;lt;math&amp;gt;\mathbf{Pr}[A_i] &amp;gt; 0&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;1 \le i \le n&amp;lt;/math&amp;gt;. Define &amp;lt;math&amp;gt;a = \sum_{i=1}^n \mathbf{Pr}[A_i]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;b = \sum_{1 \le i&amp;lt;j\le n} \mathbf{Pr}[A_i \cap A_j]&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbf{Pr}[A_1 \cup A_2 \cup \ldots \cup A_n] \ge \max\left\{2-\frac{a+2b}{a^2}, \frac{a^2}{a+2b}\right\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Probability meets distinct sums) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;    &lt;br /&gt;
&lt;br /&gt;
Let &amp;lt;math&amp;gt;f(n)&amp;lt;/math&amp;gt; denote the maximal &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt; such that there exists a set of &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt; distinct numbers &amp;lt;math&amp;gt;\{x_1,x_2,\ldots,x_m\}&amp;lt;/math&amp;gt;&lt;br /&gt;
in &amp;lt;math&amp;gt;[n] = \{1,2,\ldots,n\}&amp;lt;/math&amp;gt; all of whose sums are distinct. Namely, &amp;lt;math&amp;gt;\sum_{i \in S} x_i&amp;lt;/math&amp;gt; are distinct for all &amp;lt;math&amp;gt;S \subseteq \{1,2,\ldots,m\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
Use the second moment method (i.e., Chebyshev&#039;s inequality) to show that &amp;lt;math&amp;gt;f(n) \le \log_2 n + \frac{1}{2} \log_2 \log_2 n + O(1)&amp;lt;/math&amp;gt;. (Remark: Erdös&#039; [https://www.erdosproblems.com/1 first open problem] asks if &amp;lt;math&amp;gt;f(n) \le \log_2 n + C&amp;lt;/math&amp;gt; for some universal constant &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_3&amp;diff=13052</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 3</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_3&amp;diff=13052"/>
		<updated>2025-04-06T11:43:01Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: /* Problem 3 (Probability meets distinct sums) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 3==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Warm-up Problems) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
      Let &amp;lt;math&amp;gt;X_1,X_2,...,X_n&amp;lt;/math&amp;gt; be independent random variables, and suppose that &amp;lt;math&amp;gt;X_k&amp;lt;/math&amp;gt; is Bernoulli with parameter &amp;lt;math&amp;gt;p_k&amp;lt;/math&amp;gt;. Let &amp;lt;math&amp;gt;Y= X_1 + X_2 + \dots + X_n&amp;lt;/math&amp;gt;. Show that, for &amp;lt;math&amp;gt;\mathbb E[Y]&amp;lt;/math&amp;gt; fixed, &amp;lt;math&amp;gt;\mathrm{Var}(Y )&amp;lt;/math&amp;gt; is a maximized when &amp;lt;math&amp;gt;p_1 = p_2 = \dots = p_n&amp;lt;/math&amp;gt;. That is to say, the variation in the sum is greatest when individuals are most alike.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Each member of a group of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; players rolls a (fair) 6-sided die. For any pair of players who throw the same number, the group scores &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; point. Find the mean and variance of the total score of the group.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        An urn contains &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; balls numbered &amp;lt;math&amp;gt;1, 2, \ldots, n&amp;lt;/math&amp;gt;. We select &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; balls uniformly at random &amp;lt;strong&amp;gt;without replacement&amp;lt;/strong&amp;gt; and add up their numbers. Find the mean and variance of the sum.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (IV)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
      Let &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; be an integer-valued, positive random variable and let &amp;lt;math&amp;gt;\{X_i\}_{i=1}^{\infty}&amp;lt;/math&amp;gt; be indepedently identically distributed random variables that are independent of &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt;, too. &lt;br /&gt;
Precisely, for any finite subset &amp;lt;math&amp;gt;I \subseteq\mathbb{N}_+&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\{X_i\}_{i \in I}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; are mutually independent. Let &amp;lt;math&amp;gt;X = \sum_{i=1}^N X_i&amp;lt;/math&amp;gt;, show that &amp;lt;math&amp;gt;\textbf{Var}[X] = \textbf{Var}[X_1] \mathbb{E}[N] + \mathbb{E}[X_1]^2 \textbf{Var}[N]&amp;lt;/math&amp;gt;.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Moments (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Show that [math]G(t) = \frac{e^t}{4} + \frac{e^{-t}}{2} + \frac{1}{4}[/math] is a moment-generating function of a random variable, and write the probability mass function of this random variable.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Moments (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X\sim \text{Geo}(p)&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt;. Find &amp;lt;math&amp;gt;\mathbb{E}[X^3]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbb{E}[X^4]&amp;lt;/math&amp;gt;. &lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Moments (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X\sim \text{Pois}(\lambda)&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;\lambda &amp;gt;0 &amp;lt;/math&amp;gt;. Find &amp;lt;math&amp;gt;\mathbb{E}[X^3]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbb{E}[X^4]&amp;lt;/math&amp;gt;. &lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be discrete random variables with correlation &amp;lt;math&amp;gt;\rho&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;|\rho|= 1&amp;lt;/math&amp;gt; if and only if &amp;lt;math&amp;gt;X=aY+b&amp;lt;/math&amp;gt; for some real numbers &amp;lt;math&amp;gt;a,b&amp;lt;/math&amp;gt;.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
    Let [math]X[/math] and [math]Y[/math] be discrete random variables with mean &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;, variance &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;, and correlation [math]\rho[/math]. Show that [math]\mathbb{E}(\max\{X^2,Y^2\})\leq 1+\sqrt{1-\rho^2}[/math]. (Hint: use the identity [math]\max\{a,b\} = \frac{1}{2}(a+b+|a-b|)[/math].)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
   Let [math]X[/math] and [math]Y[/math] be independent Bernoulli random variables with parameter [math]1/2[/math]. Show that [math]X+Y[/math] and [math]|X-Y|[/math] are dependent though uncorrelated.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Inequalities) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Reverse Markov&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable with bounded range &amp;lt;math&amp;gt;0 \le X \le U&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;U &amp;gt; 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbf{Pr}(X \le a) \le \frac{U-\mathbf{E}[X]}{U-a}&amp;lt;/math&amp;gt; for any &amp;lt;math&amp;gt;0 &amp;lt; a &amp;lt; U&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Markov&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable. Show that for all &amp;lt;math&amp;gt;\beta \geq 0&amp;lt;/math&amp;gt; and all &amp;lt;math&amp;gt;x &amp;gt; 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}(X\geq x)\leq \mathbb{E}(e^{\beta X})e^{-\beta x}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Cantelli&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable with mean &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; and variance &amp;lt;math&amp;gt;\sigma^2&amp;lt;/math&amp;gt;. Prove that for any &amp;lt;math&amp;gt;\lambda &amp;gt; 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}[X \ge \lambda] \le \frac{\sigma^2}{\lambda^2+\sigma^2}&amp;lt;/math&amp;gt;. (Hint: You may first show that &amp;lt;math&amp;gt;\mathbf{Pr}[X \ge \lambda] \le \frac{\sigma^2 + u^2}{(\lambda + u)^2}&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;u &amp;gt; 0&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; &amp;lt;strong&amp;gt;[Union of events]&amp;lt;/strong&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;A_1,A_2,\ldots,A_n&amp;lt;/math&amp;gt; be events with &amp;lt;math&amp;gt;\mathbf{Pr}[A_i] &amp;gt; 0&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;1 \le i \le n&amp;lt;/math&amp;gt;. Define &amp;lt;math&amp;gt;a = \sum_{i=1}^n \mathbf{Pr}[A_i]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;b = \sum_{1 \le i&amp;lt;j\le n} \mathbf{Pr}[A_i \cap A_j]&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbf{Pr}[A_1 \cup A_2 \cup \ldots \cup A_n] \ge \max\left\{2-\frac{a+2b}{a^2}, \frac{a^2}{a+2b}\right\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Probability meets distinct sums) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;    &lt;br /&gt;
&lt;br /&gt;
Let &amp;lt;math&amp;gt;f(n)&amp;lt;/math&amp;gt; denote the maximal &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt; such that there exists a set of &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt; distinct numbers &amp;lt;math&amp;gt;\{x_1,x_2,\ldots,x_m\}&amp;lt;/math&amp;gt;&lt;br /&gt;
in &amp;lt;math&amp;gt;[n] = \{1,2,\ldots,n\}&amp;lt;/math&amp;gt; all of whose sums are distinct. Namely, &amp;lt;math&amp;gt;\sum_{i \in S} x_i&amp;lt;/math&amp;gt; are distinct for all &amp;lt;math&amp;gt;S \subseteq \{1,2,\ldots,m\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
Use the second moment method to show that &amp;lt;math&amp;gt;f(n) \le \log_2 n + \frac{1}{2} \log_2 \log_2 n + O(1)&amp;lt;/math&amp;gt;. (Remark: Erdös&#039; [https://www.erdosproblems.com/1 first open problem] asks if &amp;lt;math&amp;gt;f(n) \le \log_2 n + C&amp;lt;/math&amp;gt; for some universal constant &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_3&amp;diff=13051</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 3</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_3&amp;diff=13051"/>
		<updated>2025-04-06T11:39:18Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: /* Problem 3 (Probability meets distinct sums) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 3==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Warm-up Problems) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
      Let &amp;lt;math&amp;gt;X_1,X_2,...,X_n&amp;lt;/math&amp;gt; be independent random variables, and suppose that &amp;lt;math&amp;gt;X_k&amp;lt;/math&amp;gt; is Bernoulli with parameter &amp;lt;math&amp;gt;p_k&amp;lt;/math&amp;gt;. Let &amp;lt;math&amp;gt;Y= X_1 + X_2 + \dots + X_n&amp;lt;/math&amp;gt;. Show that, for &amp;lt;math&amp;gt;\mathbb E[Y]&amp;lt;/math&amp;gt; fixed, &amp;lt;math&amp;gt;\mathrm{Var}(Y )&amp;lt;/math&amp;gt; is a maximized when &amp;lt;math&amp;gt;p_1 = p_2 = \dots = p_n&amp;lt;/math&amp;gt;. That is to say, the variation in the sum is greatest when individuals are most alike.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Each member of a group of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; players rolls a (fair) 6-sided die. For any pair of players who throw the same number, the group scores &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; point. Find the mean and variance of the total score of the group.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        An urn contains &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; balls numbered &amp;lt;math&amp;gt;1, 2, \ldots, n&amp;lt;/math&amp;gt;. We select &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; balls uniformly at random &amp;lt;strong&amp;gt;without replacement&amp;lt;/strong&amp;gt; and add up their numbers. Find the mean and variance of the sum.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (IV)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
      Let &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; be an integer-valued, positive random variable and let &amp;lt;math&amp;gt;\{X_i\}_{i=1}^{\infty}&amp;lt;/math&amp;gt; be indepedently identically distributed random variables that are independent of &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt;, too. &lt;br /&gt;
Precisely, for any finite subset &amp;lt;math&amp;gt;I \subseteq\mathbb{N}_+&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\{X_i\}_{i \in I}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; are mutually independent. Let &amp;lt;math&amp;gt;X = \sum_{i=1}^N X_i&amp;lt;/math&amp;gt;, show that &amp;lt;math&amp;gt;\textbf{Var}[X] = \textbf{Var}[X_1] \mathbb{E}[N] + \mathbb{E}[X_1]^2 \textbf{Var}[N]&amp;lt;/math&amp;gt;.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Moments (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Show that [math]G(t) = \frac{e^t}{4} + \frac{e^{-t}}{2} + \frac{1}{4}[/math] is a moment-generating function of a random variable, and write the probability mass function of this random variable.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Moments (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X\sim \text{Geo}(p)&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt;. Find &amp;lt;math&amp;gt;\mathbb{E}[X^3]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbb{E}[X^4]&amp;lt;/math&amp;gt;. &lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Moments (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X\sim \text{Pois}(\lambda)&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;\lambda &amp;gt;0 &amp;lt;/math&amp;gt;. Find &amp;lt;math&amp;gt;\mathbb{E}[X^3]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbb{E}[X^4]&amp;lt;/math&amp;gt;. &lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be discrete random variables with correlation &amp;lt;math&amp;gt;\rho&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;|\rho|= 1&amp;lt;/math&amp;gt; if and only if &amp;lt;math&amp;gt;X=aY+b&amp;lt;/math&amp;gt; for some real numbers &amp;lt;math&amp;gt;a,b&amp;lt;/math&amp;gt;.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
    Let [math]X[/math] and [math]Y[/math] be discrete random variables with mean &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;, variance &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;, and correlation [math]\rho[/math]. Show that [math]\mathbb{E}(\max\{X^2,Y^2\})\leq 1+\sqrt{1-\rho^2}[/math]. (Hint: use the identity [math]\max\{a,b\} = \frac{1}{2}(a+b+|a-b|)[/math].)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
   Let [math]X[/math] and [math]Y[/math] be independent Bernoulli random variables with parameter [math]1/2[/math]. Show that [math]X+Y[/math] and [math]|X-Y|[/math] are dependent though uncorrelated.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Inequalities) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Reverse Markov&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable with bounded range &amp;lt;math&amp;gt;0 \le X \le U&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;U &amp;gt; 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbf{Pr}(X \le a) \le \frac{U-\mathbf{E}[X]}{U-a}&amp;lt;/math&amp;gt; for any &amp;lt;math&amp;gt;0 &amp;lt; a &amp;lt; U&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Markov&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable. Show that for all &amp;lt;math&amp;gt;\beta \geq 0&amp;lt;/math&amp;gt; and all &amp;lt;math&amp;gt;x &amp;gt; 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}(X\geq x)\leq \mathbb{E}(e^{\beta X})e^{-\beta x}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Cantelli&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable with mean &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; and variance &amp;lt;math&amp;gt;\sigma^2&amp;lt;/math&amp;gt;. Prove that for any &amp;lt;math&amp;gt;\lambda &amp;gt; 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}[X \ge \lambda] \le \frac{\sigma^2}{\lambda^2+\sigma^2}&amp;lt;/math&amp;gt;. (Hint: You may first show that &amp;lt;math&amp;gt;\mathbf{Pr}[X \ge \lambda] \le \frac{\sigma^2 + u^2}{(\lambda + u)^2}&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;u &amp;gt; 0&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; &amp;lt;strong&amp;gt;[Union of events]&amp;lt;/strong&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;A_1,A_2,\ldots,A_n&amp;lt;/math&amp;gt; be events with &amp;lt;math&amp;gt;\mathbf{Pr}[A_i] &amp;gt; 0&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;1 \le i \le n&amp;lt;/math&amp;gt;. Define &amp;lt;math&amp;gt;a = \sum_{i=1}^n \mathbf{Pr}[A_i]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;b = \sum_{1 \le i&amp;lt;j\le n} \mathbf{Pr}[A_i \cap A_j]&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbf{Pr}[A_1 \cup A_2 \cup \ldots \cup A_n] \ge \max\left\{2-\frac{a+2b}{a^2}, \frac{a^2}{a+2b}\right\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Probability meets distinct sums) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;    &lt;br /&gt;
&lt;br /&gt;
Let &amp;lt;math&amp;gt;f(n)&amp;lt;/math&amp;gt; denote the maximal &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt; such that there exists a set of &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt; distinct numbers&lt;br /&gt;
in &amp;lt;math&amp;gt;[n] = \{1,2,\ldots,n\}&amp;lt;/math&amp;gt; all of whose sums are distinct. &lt;br /&gt;
Use the second moment to show that &amp;lt;math&amp;gt;f(n) \le \log_2 n + \frac{1}{2} \log_2 \log_2 n + O(1)&amp;lt;/math&amp;gt;. (Remark: Erdös&#039; [https://www.erdosproblems.com/1 first open problem] asks if &amp;lt;math&amp;gt;f(n) \le \log_2 n + C&amp;lt;/math&amp;gt; for some universal constant &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_3&amp;diff=13050</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 3</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_3&amp;diff=13050"/>
		<updated>2025-04-06T11:36:12Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: /* Problem 3 (Probability meets distinct sums) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 3==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Warm-up Problems) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
      Let &amp;lt;math&amp;gt;X_1,X_2,...,X_n&amp;lt;/math&amp;gt; be independent random variables, and suppose that &amp;lt;math&amp;gt;X_k&amp;lt;/math&amp;gt; is Bernoulli with parameter &amp;lt;math&amp;gt;p_k&amp;lt;/math&amp;gt;. Let &amp;lt;math&amp;gt;Y= X_1 + X_2 + \dots + X_n&amp;lt;/math&amp;gt;. Show that, for &amp;lt;math&amp;gt;\mathbb E[Y]&amp;lt;/math&amp;gt; fixed, &amp;lt;math&amp;gt;\mathrm{Var}(Y )&amp;lt;/math&amp;gt; is a maximized when &amp;lt;math&amp;gt;p_1 = p_2 = \dots = p_n&amp;lt;/math&amp;gt;. That is to say, the variation in the sum is greatest when individuals are most alike.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Each member of a group of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; players rolls a (fair) 6-sided die. For any pair of players who throw the same number, the group scores &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; point. Find the mean and variance of the total score of the group.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        An urn contains &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; balls numbered &amp;lt;math&amp;gt;1, 2, \ldots, n&amp;lt;/math&amp;gt;. We select &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; balls uniformly at random &amp;lt;strong&amp;gt;without replacement&amp;lt;/strong&amp;gt; and add up their numbers. Find the mean and variance of the sum.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (IV)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
      Let &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; be an integer-valued, positive random variable and let &amp;lt;math&amp;gt;\{X_i\}_{i=1}^{\infty}&amp;lt;/math&amp;gt; be indepedently identically distributed random variables that are independent of &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt;, too. &lt;br /&gt;
Precisely, for any finite subset &amp;lt;math&amp;gt;I \subseteq\mathbb{N}_+&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\{X_i\}_{i \in I}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; are mutually independent. Let &amp;lt;math&amp;gt;X = \sum_{i=1}^N X_i&amp;lt;/math&amp;gt;, show that &amp;lt;math&amp;gt;\textbf{Var}[X] = \textbf{Var}[X_1] \mathbb{E}[N] + \mathbb{E}[X_1]^2 \textbf{Var}[N]&amp;lt;/math&amp;gt;.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Moments (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Show that [math]G(t) = \frac{e^t}{4} + \frac{e^{-t}}{2} + \frac{1}{4}[/math] is a moment-generating function of a random variable, and write the probability mass function of this random variable.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Moments (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X\sim \text{Geo}(p)&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt;. Find &amp;lt;math&amp;gt;\mathbb{E}[X^3]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbb{E}[X^4]&amp;lt;/math&amp;gt;. &lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Moments (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X\sim \text{Pois}(\lambda)&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;\lambda &amp;gt;0 &amp;lt;/math&amp;gt;. Find &amp;lt;math&amp;gt;\mathbb{E}[X^3]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbb{E}[X^4]&amp;lt;/math&amp;gt;. &lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be discrete random variables with correlation &amp;lt;math&amp;gt;\rho&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;|\rho|= 1&amp;lt;/math&amp;gt; if and only if &amp;lt;math&amp;gt;X=aY+b&amp;lt;/math&amp;gt; for some real numbers &amp;lt;math&amp;gt;a,b&amp;lt;/math&amp;gt;.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
    Let [math]X[/math] and [math]Y[/math] be discrete random variables with mean &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;, variance &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;, and correlation [math]\rho[/math]. Show that [math]\mathbb{E}(\max\{X^2,Y^2\})\leq 1+\sqrt{1-\rho^2}[/math]. (Hint: use the identity [math]\max\{a,b\} = \frac{1}{2}(a+b+|a-b|)[/math].)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
   Let [math]X[/math] and [math]Y[/math] be independent Bernoulli random variables with parameter [math]1/2[/math]. Show that [math]X+Y[/math] and [math]|X-Y|[/math] are dependent though uncorrelated.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Inequalities) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Reverse Markov&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable with bounded range &amp;lt;math&amp;gt;0 \le X \le U&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;U &amp;gt; 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbf{Pr}(X \le a) \le \frac{U-\mathbf{E}[X]}{U-a}&amp;lt;/math&amp;gt; for any &amp;lt;math&amp;gt;0 &amp;lt; a &amp;lt; U&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Markov&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable. Show that for all &amp;lt;math&amp;gt;\beta \geq 0&amp;lt;/math&amp;gt; and all &amp;lt;math&amp;gt;x &amp;gt; 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}(X\geq x)\leq \mathbb{E}(e^{\beta X})e^{-\beta x}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Cantelli&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable with mean &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; and variance &amp;lt;math&amp;gt;\sigma^2&amp;lt;/math&amp;gt;. Prove that for any &amp;lt;math&amp;gt;\lambda &amp;gt; 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}[X \ge \lambda] \le \frac{\sigma^2}{\lambda^2+\sigma^2}&amp;lt;/math&amp;gt;. (Hint: You may first show that &amp;lt;math&amp;gt;\mathbf{Pr}[X \ge \lambda] \le \frac{\sigma^2 + u^2}{(\lambda + u)^2}&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;u &amp;gt; 0&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; &amp;lt;strong&amp;gt;[Union of events]&amp;lt;/strong&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;A_1,A_2,\ldots,A_n&amp;lt;/math&amp;gt; be events with &amp;lt;math&amp;gt;\mathbf{Pr}[A_i] &amp;gt; 0&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;1 \le i \le n&amp;lt;/math&amp;gt;. Define &amp;lt;math&amp;gt;a = \sum_{i=1}^n \mathbf{Pr}[A_i]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;b = \sum_{1 \le i&amp;lt;j\le n} \mathbf{Pr}[A_i \cap A_j]&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbf{Pr}[A_1 \cup A_2 \cup \ldots \cup A_n] \ge \max\left\{2-\frac{a+2b}{a^2}, \frac{a^2}{a+2b}\right\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Probability meets distinct sums) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;    &lt;br /&gt;
&lt;br /&gt;
Let &amp;lt;math&amp;gt;f(n)&amp;lt;/math&amp;gt; denote the maximal &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt; such that there exists a set of &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt; distinct numbers&lt;br /&gt;
in &amp;lt;math&amp;gt;[n] = \{1,2,\ldots,n\}&amp;lt;/math&amp;gt; all of whose sums are distinct. &lt;br /&gt;
Show that &amp;lt;math&amp;gt;f(n) \le \log_2 n + \frac{1}{2} \log_2 \log_2 n + O(1)&amp;lt;/math&amp;gt;. (Remark: Erdös&#039; [https://www.erdosproblems.com/1 first open problem] asks if &amp;lt;math&amp;gt;f(n) \le \log_2 n + C&amp;lt;/math&amp;gt; for some universal constant &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_3&amp;diff=13049</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 3</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_3&amp;diff=13049"/>
		<updated>2025-04-06T11:35:39Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 3==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Warm-up Problems) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
      Let &amp;lt;math&amp;gt;X_1,X_2,...,X_n&amp;lt;/math&amp;gt; be independent random variables, and suppose that &amp;lt;math&amp;gt;X_k&amp;lt;/math&amp;gt; is Bernoulli with parameter &amp;lt;math&amp;gt;p_k&amp;lt;/math&amp;gt;. Let &amp;lt;math&amp;gt;Y= X_1 + X_2 + \dots + X_n&amp;lt;/math&amp;gt;. Show that, for &amp;lt;math&amp;gt;\mathbb E[Y]&amp;lt;/math&amp;gt; fixed, &amp;lt;math&amp;gt;\mathrm{Var}(Y )&amp;lt;/math&amp;gt; is a maximized when &amp;lt;math&amp;gt;p_1 = p_2 = \dots = p_n&amp;lt;/math&amp;gt;. That is to say, the variation in the sum is greatest when individuals are most alike.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Each member of a group of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; players rolls a (fair) 6-sided die. For any pair of players who throw the same number, the group scores &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; point. Find the mean and variance of the total score of the group.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        An urn contains &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; balls numbered &amp;lt;math&amp;gt;1, 2, \ldots, n&amp;lt;/math&amp;gt;. We select &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; balls uniformly at random &amp;lt;strong&amp;gt;without replacement&amp;lt;/strong&amp;gt; and add up their numbers. Find the mean and variance of the sum.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (IV)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
      Let &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; be an integer-valued, positive random variable and let &amp;lt;math&amp;gt;\{X_i\}_{i=1}^{\infty}&amp;lt;/math&amp;gt; be indepedently identically distributed random variables that are independent of &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt;, too. &lt;br /&gt;
Precisely, for any finite subset &amp;lt;math&amp;gt;I \subseteq\mathbb{N}_+&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\{X_i\}_{i \in I}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; are mutually independent. Let &amp;lt;math&amp;gt;X = \sum_{i=1}^N X_i&amp;lt;/math&amp;gt;, show that &amp;lt;math&amp;gt;\textbf{Var}[X] = \textbf{Var}[X_1] \mathbb{E}[N] + \mathbb{E}[X_1]^2 \textbf{Var}[N]&amp;lt;/math&amp;gt;.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Moments (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Show that [math]G(t) = \frac{e^t}{4} + \frac{e^{-t}}{2} + \frac{1}{4}[/math] is a moment-generating function of a random variable, and write the probability mass function of this random variable.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Moments (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X\sim \text{Geo}(p)&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt;. Find &amp;lt;math&amp;gt;\mathbb{E}[X^3]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbb{E}[X^4]&amp;lt;/math&amp;gt;. &lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Moments (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X\sim \text{Pois}(\lambda)&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;\lambda &amp;gt;0 &amp;lt;/math&amp;gt;. Find &amp;lt;math&amp;gt;\mathbb{E}[X^3]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbb{E}[X^4]&amp;lt;/math&amp;gt;. &lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be discrete random variables with correlation &amp;lt;math&amp;gt;\rho&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;|\rho|= 1&amp;lt;/math&amp;gt; if and only if &amp;lt;math&amp;gt;X=aY+b&amp;lt;/math&amp;gt; for some real numbers &amp;lt;math&amp;gt;a,b&amp;lt;/math&amp;gt;.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
    Let [math]X[/math] and [math]Y[/math] be discrete random variables with mean &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;, variance &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;, and correlation [math]\rho[/math]. Show that [math]\mathbb{E}(\max\{X^2,Y^2\})\leq 1+\sqrt{1-\rho^2}[/math]. (Hint: use the identity [math]\max\{a,b\} = \frac{1}{2}(a+b+|a-b|)[/math].)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
   Let [math]X[/math] and [math]Y[/math] be independent Bernoulli random variables with parameter [math]1/2[/math]. Show that [math]X+Y[/math] and [math]|X-Y|[/math] are dependent though uncorrelated.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Inequalities) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Reverse Markov&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable with bounded range &amp;lt;math&amp;gt;0 \le X \le U&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;U &amp;gt; 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbf{Pr}(X \le a) \le \frac{U-\mathbf{E}[X]}{U-a}&amp;lt;/math&amp;gt; for any &amp;lt;math&amp;gt;0 &amp;lt; a &amp;lt; U&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Markov&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable. Show that for all &amp;lt;math&amp;gt;\beta \geq 0&amp;lt;/math&amp;gt; and all &amp;lt;math&amp;gt;x &amp;gt; 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}(X\geq x)\leq \mathbb{E}(e^{\beta X})e^{-\beta x}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Cantelli&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable with mean &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; and variance &amp;lt;math&amp;gt;\sigma^2&amp;lt;/math&amp;gt;. Prove that for any &amp;lt;math&amp;gt;\lambda &amp;gt; 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}[X \ge \lambda] \le \frac{\sigma^2}{\lambda^2+\sigma^2}&amp;lt;/math&amp;gt;. (Hint: You may first show that &amp;lt;math&amp;gt;\mathbf{Pr}[X \ge \lambda] \le \frac{\sigma^2 + u^2}{(\lambda + u)^2}&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;u &amp;gt; 0&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; &amp;lt;strong&amp;gt;[Union of events]&amp;lt;/strong&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;A_1,A_2,\ldots,A_n&amp;lt;/math&amp;gt; be events with &amp;lt;math&amp;gt;\mathbf{Pr}[A_i] &amp;gt; 0&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;1 \le i \le n&amp;lt;/math&amp;gt;. Define &amp;lt;math&amp;gt;a = \sum_{i=1}^n \mathbf{Pr}[A_i]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;b = \sum_{1 \le i&amp;lt;j\le n} \mathbf{Pr}[A_i \cap A_j]&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbf{Pr}[A_1 \cup A_2 \cup \ldots \cup A_n] \ge \max\left\{2-\frac{a+2b}{a^2}, \frac{a^2}{a+2b}\right\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Probability meets distinct sums) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;    &lt;br /&gt;
&lt;br /&gt;
Let &amp;lt;math&amp;gt;f(n)&amp;lt;/math&amp;gt; denote the maximal &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt; such that there exists a set of &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt; distinct numbers&lt;br /&gt;
in &amp;lt;math&amp;gt;[n] = \{1,2,\ldots,n\}&amp;lt;/math&amp;gt; all of whose sums are distinct. &lt;br /&gt;
Show that &amp;lt;math&amp;gt;f(n) \le \log_2 n + \frac{1}{2} \log_2 \log_2 n + O(1)&amp;lt;/math&amp;gt;. &lt;br /&gt;
&lt;br /&gt;
(Remark: Erdös&#039; [https://www.erdosproblems.com/1 first open problem] asks if &amp;lt;math&amp;gt;f(n) \le \log_2 n + C&amp;lt;/math&amp;gt; for some universal constant &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_3&amp;diff=13048</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 3</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_3&amp;diff=13048"/>
		<updated>2025-04-06T11:29:14Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: /* Problem 3 (Probability meets graph theory) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 3==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Warm-up Problems) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
      Let &amp;lt;math&amp;gt;X_1,X_2,...,X_n&amp;lt;/math&amp;gt; be independent random variables, and suppose that &amp;lt;math&amp;gt;X_k&amp;lt;/math&amp;gt; is Bernoulli with parameter &amp;lt;math&amp;gt;p_k&amp;lt;/math&amp;gt;. Let &amp;lt;math&amp;gt;Y= X_1 + X_2 + \dots + X_n&amp;lt;/math&amp;gt;. Show that, for &amp;lt;math&amp;gt;\mathbb E[Y]&amp;lt;/math&amp;gt; fixed, &amp;lt;math&amp;gt;\mathrm{Var}(Y )&amp;lt;/math&amp;gt; is a maximized when &amp;lt;math&amp;gt;p_1 = p_2 = \dots = p_n&amp;lt;/math&amp;gt;. That is to say, the variation in the sum is greatest when individuals are most alike.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Each member of a group of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; players rolls a (fair) 6-sided die. For any pair of players who throw the same number, the group scores &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; point. Find the mean and variance of the total score of the group.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        An urn contains &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; balls numbered &amp;lt;math&amp;gt;1, 2, \ldots, n&amp;lt;/math&amp;gt;. We select &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; balls uniformly at random &amp;lt;strong&amp;gt;without replacement&amp;lt;/strong&amp;gt; and add up their numbers. Find the mean and variance of the sum.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (IV)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
      Let &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; be an integer-valued, positive random variable and let &amp;lt;math&amp;gt;\{X_i\}_{i=1}^{\infty}&amp;lt;/math&amp;gt; be indepedently identically distributed random variables that are independent of &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt;, too. &lt;br /&gt;
Precisely, for any finite subset &amp;lt;math&amp;gt;I \subseteq\mathbb{N}_+&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\{X_i\}_{i \in I}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; are mutually independent. Let &amp;lt;math&amp;gt;X = \sum_{i=1}^N X_i&amp;lt;/math&amp;gt;, show that &amp;lt;math&amp;gt;\textbf{Var}[X] = \textbf{Var}[X_1] \mathbb{E}[N] + \mathbb{E}[X_1]^2 \textbf{Var}[N]&amp;lt;/math&amp;gt;.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Moments (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Show that [math]G(t) = \frac{e^t}{4} + \frac{e^{-t}}{2} + \frac{1}{4}[/math] is a moment-generating function of a random variable, and write the probability mass function of this random variable.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Moments (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X\sim \text{Geo}(p)&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt;. Find &amp;lt;math&amp;gt;\mathbb{E}[X^3]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbb{E}[X^4]&amp;lt;/math&amp;gt;. &lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Moments (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X\sim \text{Pois}(\lambda)&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;\lambda &amp;gt;0 &amp;lt;/math&amp;gt;. Find &amp;lt;math&amp;gt;\mathbb{E}[X^3]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbb{E}[X^4]&amp;lt;/math&amp;gt;. &lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be discrete random variables with correlation &amp;lt;math&amp;gt;\rho&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;|\rho|= 1&amp;lt;/math&amp;gt; if and only if &amp;lt;math&amp;gt;X=aY+b&amp;lt;/math&amp;gt; for some real numbers &amp;lt;math&amp;gt;a,b&amp;lt;/math&amp;gt;.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
    Let [math]X[/math] and [math]Y[/math] be discrete random variables with mean &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;, variance &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;, and correlation [math]\rho[/math]. Show that [math]\mathbb{E}(\max\{X^2,Y^2\})\leq 1+\sqrt{1-\rho^2}[/math]. (Hint: use the identity [math]\max\{a,b\} = \frac{1}{2}(a+b+|a-b|)[/math].)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
   Let [math]X[/math] and [math]Y[/math] be independent Bernoulli random variables with parameter [math]1/2[/math]. Show that [math]X+Y[/math] and [math]|X-Y|[/math] are dependent though uncorrelated.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Inequalities) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Reverse Markov&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable with bounded range &amp;lt;math&amp;gt;0 \le X \le U&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;U &amp;gt; 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbf{Pr}(X \le a) \le \frac{U-\mathbf{E}[X]}{U-a}&amp;lt;/math&amp;gt; for any &amp;lt;math&amp;gt;0 &amp;lt; a &amp;lt; U&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Markov&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable. Show that for all &amp;lt;math&amp;gt;\beta \geq 0&amp;lt;/math&amp;gt; and all &amp;lt;math&amp;gt;x &amp;gt; 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}(X\geq x)\leq \mathbb{E}(e^{\beta X})e^{-\beta x}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Cantelli&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable with mean &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; and variance &amp;lt;math&amp;gt;\sigma^2&amp;lt;/math&amp;gt;. Prove that for any &amp;lt;math&amp;gt;\lambda &amp;gt; 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}[X \ge \lambda] \le \frac{\sigma^2}{\lambda^2+\sigma^2}&amp;lt;/math&amp;gt;. (Hint: You may first show that &amp;lt;math&amp;gt;\mathbf{Pr}[X \ge \lambda] \le \frac{\sigma^2 + u^2}{(\lambda + u)^2}&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;u &amp;gt; 0&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; &amp;lt;strong&amp;gt;[Union of events]&amp;lt;/strong&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;A_1,A_2,\ldots,A_n&amp;lt;/math&amp;gt; be events with &amp;lt;math&amp;gt;\mathbf{Pr}[A_i] &amp;gt; 0&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;1 \le i \le n&amp;lt;/math&amp;gt;. Define &amp;lt;math&amp;gt;a = \sum_{i=1}^n \mathbf{Pr}[A_i]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;b = \sum_{1 \le i&amp;lt;j\le n} \mathbf{Pr}[A_i \cap A_j]&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbf{Pr}[A_1 \cup A_2 \cup \ldots \cup A_n] \ge \max\left\{2-\frac{a+2b}{a^2}, \frac{a^2}{a+2b}\right\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Probability meets distinct sums) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;    &lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;f(n)&amp;lt;/math&amp;gt; denote the maximal &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt; such that there exists a set &amp;lt;math&amp;gt;&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt; distinct numbers&lt;br /&gt;
in &amp;lt;math&amp;gt;[n] = \{1,2,\ldots,n\}&amp;lt;/math&amp;gt; all of whose sums are distinct. &lt;br /&gt;
Show that &amp;lt;math&amp;gt;f(n) \le \log_2 n + \frac{1}{2} \log_2 \log_2 n + O(1)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_3&amp;diff=13047</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 3</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_3&amp;diff=13047"/>
		<updated>2025-04-06T11:10:23Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: /* Problem 2 (Inequalities) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 3==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Warm-up Problems) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
      Let &amp;lt;math&amp;gt;X_1,X_2,...,X_n&amp;lt;/math&amp;gt; be independent random variables, and suppose that &amp;lt;math&amp;gt;X_k&amp;lt;/math&amp;gt; is Bernoulli with parameter &amp;lt;math&amp;gt;p_k&amp;lt;/math&amp;gt;. Let &amp;lt;math&amp;gt;Y= X_1 + X_2 + \dots + X_n&amp;lt;/math&amp;gt;. Show that, for &amp;lt;math&amp;gt;\mathbb E[Y]&amp;lt;/math&amp;gt; fixed, &amp;lt;math&amp;gt;\mathrm{Var}(Y )&amp;lt;/math&amp;gt; is a maximized when &amp;lt;math&amp;gt;p_1 = p_2 = \dots = p_n&amp;lt;/math&amp;gt;. That is to say, the variation in the sum is greatest when individuals are most alike.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Each member of a group of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; players rolls a (fair) 6-sided die. For any pair of players who throw the same number, the group scores &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; point. Find the mean and variance of the total score of the group.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        An urn contains &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; balls numbered &amp;lt;math&amp;gt;1, 2, \ldots, n&amp;lt;/math&amp;gt;. We select &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; balls uniformly at random &amp;lt;strong&amp;gt;without replacement&amp;lt;/strong&amp;gt; and add up their numbers. Find the mean and variance of the sum.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (IV)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
      Let &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; be an integer-valued, positive random variable and let &amp;lt;math&amp;gt;\{X_i\}_{i=1}^{\infty}&amp;lt;/math&amp;gt; be indepedently identically distributed random variables that are independent of &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt;, too. &lt;br /&gt;
Precisely, for any finite subset &amp;lt;math&amp;gt;I \subseteq\mathbb{N}_+&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\{X_i\}_{i \in I}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; are mutually independent. Let &amp;lt;math&amp;gt;X = \sum_{i=1}^N X_i&amp;lt;/math&amp;gt;, show that &amp;lt;math&amp;gt;\textbf{Var}[X] = \textbf{Var}[X_1] \mathbb{E}[N] + \mathbb{E}[X_1]^2 \textbf{Var}[N]&amp;lt;/math&amp;gt;.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Moments (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Show that [math]G(t) = \frac{e^t}{4} + \frac{e^{-t}}{2} + \frac{1}{4}[/math] is a moment-generating function of a random variable, and write the probability mass function of this random variable.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Moments (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X\sim \text{Geo}(p)&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt;. Find &amp;lt;math&amp;gt;\mathbb{E}[X^3]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbb{E}[X^4]&amp;lt;/math&amp;gt;. &lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Moments (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X\sim \text{Pois}(\lambda)&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;\lambda &amp;gt;0 &amp;lt;/math&amp;gt;. Find &amp;lt;math&amp;gt;\mathbb{E}[X^3]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbb{E}[X^4]&amp;lt;/math&amp;gt;. &lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be discrete random variables with correlation &amp;lt;math&amp;gt;\rho&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;|\rho|= 1&amp;lt;/math&amp;gt; if and only if &amp;lt;math&amp;gt;X=aY+b&amp;lt;/math&amp;gt; for some real numbers &amp;lt;math&amp;gt;a,b&amp;lt;/math&amp;gt;.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
    Let [math]X[/math] and [math]Y[/math] be discrete random variables with mean &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;, variance &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;, and correlation [math]\rho[/math]. Show that [math]\mathbb{E}(\max\{X^2,Y^2\})\leq 1+\sqrt{1-\rho^2}[/math]. (Hint: use the identity [math]\max\{a,b\} = \frac{1}{2}(a+b+|a-b|)[/math].)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
   Let [math]X[/math] and [math]Y[/math] be independent Bernoulli random variables with parameter [math]1/2[/math]. Show that [math]X+Y[/math] and [math]|X-Y|[/math] are dependent though uncorrelated.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Inequalities) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Reverse Markov&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable with bounded range &amp;lt;math&amp;gt;0 \le X \le U&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;U &amp;gt; 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbf{Pr}(X \le a) \le \frac{U-\mathbf{E}[X]}{U-a}&amp;lt;/math&amp;gt; for any &amp;lt;math&amp;gt;0 &amp;lt; a &amp;lt; U&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Markov&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable. Show that for all &amp;lt;math&amp;gt;\beta \geq 0&amp;lt;/math&amp;gt; and all &amp;lt;math&amp;gt;x &amp;gt; 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}(X\geq x)\leq \mathbb{E}(e^{\beta X})e^{-\beta x}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Cantelli&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable with mean &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; and variance &amp;lt;math&amp;gt;\sigma^2&amp;lt;/math&amp;gt;. Prove that for any &amp;lt;math&amp;gt;\lambda &amp;gt; 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}[X \ge \lambda] \le \frac{\sigma^2}{\lambda^2+\sigma^2}&amp;lt;/math&amp;gt;. (Hint: You may first show that &amp;lt;math&amp;gt;\mathbf{Pr}[X \ge \lambda] \le \frac{\sigma^2 + u^2}{(\lambda + u)^2}&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;u &amp;gt; 0&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; &amp;lt;strong&amp;gt;[Union of events]&amp;lt;/strong&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;A_1,A_2,\ldots,A_n&amp;lt;/math&amp;gt; be events with &amp;lt;math&amp;gt;\mathbf{Pr}[A_i] &amp;gt; 0&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;1 \le i \le n&amp;lt;/math&amp;gt;. Define &amp;lt;math&amp;gt;a = \sum_{i=1}^n \mathbf{Pr}[A_i]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;b = \sum_{1 \le i&amp;lt;j\le n} \mathbf{Pr}[A_i \cap A_j]&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbf{Pr}[A_1 \cup A_2 \cup \ldots \cup A_n] \ge \max\left\{2-\frac{a+2b}{a^2}, \frac{a^2}{a+2b}\right\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Probability meets graph theory) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;    &lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;4-clique threshold&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Prove that &amp;lt;math&amp;gt;p = n^{-2/3}&amp;lt;/math&amp;gt; is the threshold probability for the existence of 4-clique. &lt;br /&gt;
        Formally, you are required to show that &lt;br /&gt;
        &amp;lt;ol type=&amp;quot;a&amp;quot;&amp;gt;&lt;br /&gt;
            &amp;lt;li&amp;gt;&lt;br /&gt;
                with a probability approaching to &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; (as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity) the Erdős–Rényi random graph &amp;lt;math&amp;gt;\mathbf{G} = \mathbf{G}(n,p)&amp;lt;/math&amp;gt; contains a 4-clique when &amp;lt;math&amp;gt;p = \omega(n^{-2/3})&amp;lt;/math&amp;gt;; (Hint: use Chebyshev&#039;s inequality.)&lt;br /&gt;
            &amp;lt;/li&amp;gt;&lt;br /&gt;
            &amp;lt;li&amp;gt;&lt;br /&gt;
                with a probability approaching to &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; (as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity) the Erdős–Rényi random graph &amp;lt;math&amp;gt;\mathbf{G} = \mathbf{G}(n,p)&amp;lt;/math&amp;gt; contains a 4-clique when &amp;lt;math&amp;gt;p = o(n^{-2/3})&amp;lt;/math&amp;gt;. (Hint: use Markov&#039;s inequality.)&lt;br /&gt;
            &amp;lt;/li&amp;gt;&lt;br /&gt;
        &amp;lt;/ol&amp;gt;&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_3&amp;diff=13046</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 3</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_3&amp;diff=13046"/>
		<updated>2025-04-06T10:57:27Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: /* Problem 2 (Inequalities) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 3==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Warm-up Problems) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
      Let &amp;lt;math&amp;gt;X_1,X_2,...,X_n&amp;lt;/math&amp;gt; be independent random variables, and suppose that &amp;lt;math&amp;gt;X_k&amp;lt;/math&amp;gt; is Bernoulli with parameter &amp;lt;math&amp;gt;p_k&amp;lt;/math&amp;gt;. Let &amp;lt;math&amp;gt;Y= X_1 + X_2 + \dots + X_n&amp;lt;/math&amp;gt;. Show that, for &amp;lt;math&amp;gt;\mathbb E[Y]&amp;lt;/math&amp;gt; fixed, &amp;lt;math&amp;gt;\mathrm{Var}(Y )&amp;lt;/math&amp;gt; is a maximized when &amp;lt;math&amp;gt;p_1 = p_2 = \dots = p_n&amp;lt;/math&amp;gt;. That is to say, the variation in the sum is greatest when individuals are most alike.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Each member of a group of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; players rolls a (fair) 6-sided die. For any pair of players who throw the same number, the group scores &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; point. Find the mean and variance of the total score of the group.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        An urn contains &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; balls numbered &amp;lt;math&amp;gt;1, 2, \ldots, n&amp;lt;/math&amp;gt;. We select &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; balls uniformly at random &amp;lt;strong&amp;gt;without replacement&amp;lt;/strong&amp;gt; and add up their numbers. Find the mean and variance of the sum.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (IV)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
      Let &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; be an integer-valued, positive random variable and let &amp;lt;math&amp;gt;\{X_i\}_{i=1}^{\infty}&amp;lt;/math&amp;gt; be indepedently identically distributed random variables that are independent of &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt;, too. &lt;br /&gt;
Precisely, for any finite subset &amp;lt;math&amp;gt;I \subseteq\mathbb{N}_+&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\{X_i\}_{i \in I}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; are mutually independent. Let &amp;lt;math&amp;gt;X = \sum_{i=1}^N X_i&amp;lt;/math&amp;gt;, show that &amp;lt;math&amp;gt;\textbf{Var}[X] = \textbf{Var}[X_1] \mathbb{E}[N] + \mathbb{E}[X_1]^2 \textbf{Var}[N]&amp;lt;/math&amp;gt;.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Moments (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Show that [math]G(t) = \frac{e^t}{4} + \frac{e^{-t}}{2} + \frac{1}{4}[/math] is a moment-generating function of a random variable, and write the probability mass function of this random variable.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Moments (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X\sim \text{Geo}(p)&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt;. Find &amp;lt;math&amp;gt;\mathbb{E}[X^3]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbb{E}[X^4]&amp;lt;/math&amp;gt;. &lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Moments (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X\sim \text{Pois}(\lambda)&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;\lambda &amp;gt;0 &amp;lt;/math&amp;gt;. Find &amp;lt;math&amp;gt;\mathbb{E}[X^3]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbb{E}[X^4]&amp;lt;/math&amp;gt;. &lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be discrete random variables with correlation &amp;lt;math&amp;gt;\rho&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;|\rho|= 1&amp;lt;/math&amp;gt; if and only if &amp;lt;math&amp;gt;X=aY+b&amp;lt;/math&amp;gt; for some real numbers &amp;lt;math&amp;gt;a,b&amp;lt;/math&amp;gt;.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
    Let [math]X[/math] and [math]Y[/math] be discrete random variables with mean &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;, variance &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;, and correlation [math]\rho[/math]. Show that [math]\mathbb{E}(\max\{X^2,Y^2\})\leq 1+\sqrt{1-\rho^2}[/math]. (Hint: use the identity [math]\max\{a,b\} = \frac{1}{2}(a+b+|a-b|)[/math].)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
   Let [math]X[/math] and [math]Y[/math] be independent Bernoulli random variables with parameter [math]1/2[/math]. Show that [math]X+Y[/math] and [math]|X-Y|[/math] are dependent though uncorrelated.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Inequalities) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Reverse Markov&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable with bounded range &amp;lt;math&amp;gt;0 \le X \le U&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;U &amp;gt; 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbf{Pr}(X \le a) \le \frac{U-\mathbf{E}[X]}{U-a}&amp;lt;/math&amp;gt; for any &amp;lt;math&amp;gt;0 &amp;lt; a &amp;lt; U&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Markov&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable. Show that for all &amp;lt;math&amp;gt;\beta \geq 0&amp;lt;/math&amp;gt; and all &amp;lt;math&amp;gt;x &amp;gt; 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}(X\geq x)\leq \mathbb{E}(e^{\beta X})e^{-\beta x}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Cantelli&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable with mean &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; and variance &amp;lt;math&amp;gt;\sigma^2&amp;lt;/math&amp;gt;. Prove that for any &amp;lt;math&amp;gt;\lambda &amp;gt; 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}[X \ge \lambda] \le \frac{\sigma^2}{\lambda^2+\sigma^2}&amp;lt;/math&amp;gt;. (Hint: You may first show that &amp;lt;math&amp;gt;\mathbf{Pr}[X \ge \lambda] \le \frac{\sigma^2 + u^2}{(\lambda + u)^2}&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;u &amp;gt; 0&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; &amp;lt;strong&amp;gt;[Union of events]&amp;lt;/strong&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;A_1,A_2,\ldots,A_n&amp;lt;/math&amp;gt; be events with &amp;lt;math&amp;gt;\mathbf{Pr}[A_i] &amp;gt; 0&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;1 \le i \le n&amp;lt;/math&amp;gt;. Define &amp;lt;math&amp;gt;a = \sum_{i=1}^n \mathbf{Pr}[A_i]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;b = \sum_{1 \le i&amp;lt;j\le n} \mathbf{Pr}[A_i \cap A_j]&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbf{Pr}[A_1 \cup A_2 \cup \ldots \cup A_n] \ge \max\left\{2-\frac{a+2b}{a^2}, \frac{a^2}{a+b}\right\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Probability meets graph theory) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;    &lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;4-clique threshold&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Prove that &amp;lt;math&amp;gt;p = n^{-2/3}&amp;lt;/math&amp;gt; is the threshold probability for the existence of 4-clique. &lt;br /&gt;
        Formally, you are required to show that &lt;br /&gt;
        &amp;lt;ol type=&amp;quot;a&amp;quot;&amp;gt;&lt;br /&gt;
            &amp;lt;li&amp;gt;&lt;br /&gt;
                with a probability approaching to &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; (as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity) the Erdős–Rényi random graph &amp;lt;math&amp;gt;\mathbf{G} = \mathbf{G}(n,p)&amp;lt;/math&amp;gt; contains a 4-clique when &amp;lt;math&amp;gt;p = \omega(n^{-2/3})&amp;lt;/math&amp;gt;; (Hint: use Chebyshev&#039;s inequality.)&lt;br /&gt;
            &amp;lt;/li&amp;gt;&lt;br /&gt;
            &amp;lt;li&amp;gt;&lt;br /&gt;
                with a probability approaching to &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; (as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity) the Erdős–Rényi random graph &amp;lt;math&amp;gt;\mathbf{G} = \mathbf{G}(n,p)&amp;lt;/math&amp;gt; contains a 4-clique when &amp;lt;math&amp;gt;p = o(n^{-2/3})&amp;lt;/math&amp;gt;. (Hint: use Markov&#039;s inequality.)&lt;br /&gt;
            &amp;lt;/li&amp;gt;&lt;br /&gt;
        &amp;lt;/ol&amp;gt;&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_3&amp;diff=13045</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 3</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_3&amp;diff=13045"/>
		<updated>2025-04-06T10:50:13Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: /* Problem 2 (Inequalities) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 3==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Warm-up Problems) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
      Let &amp;lt;math&amp;gt;X_1,X_2,...,X_n&amp;lt;/math&amp;gt; be independent random variables, and suppose that &amp;lt;math&amp;gt;X_k&amp;lt;/math&amp;gt; is Bernoulli with parameter &amp;lt;math&amp;gt;p_k&amp;lt;/math&amp;gt;. Let &amp;lt;math&amp;gt;Y= X_1 + X_2 + \dots + X_n&amp;lt;/math&amp;gt;. Show that, for &amp;lt;math&amp;gt;\mathbb E[Y]&amp;lt;/math&amp;gt; fixed, &amp;lt;math&amp;gt;\mathrm{Var}(Y )&amp;lt;/math&amp;gt; is a maximized when &amp;lt;math&amp;gt;p_1 = p_2 = \dots = p_n&amp;lt;/math&amp;gt;. That is to say, the variation in the sum is greatest when individuals are most alike.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Each member of a group of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; players rolls a (fair) 6-sided die. For any pair of players who throw the same number, the group scores &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; point. Find the mean and variance of the total score of the group.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        An urn contains &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; balls numbered &amp;lt;math&amp;gt;1, 2, \ldots, n&amp;lt;/math&amp;gt;. We select &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; balls uniformly at random &amp;lt;strong&amp;gt;without replacement&amp;lt;/strong&amp;gt; and add up their numbers. Find the mean and variance of the sum.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (IV)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
      Let &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; be an integer-valued, positive random variable and let &amp;lt;math&amp;gt;\{X_i\}_{i=1}^{\infty}&amp;lt;/math&amp;gt; be indepedently identically distributed random variables that are independent of &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt;, too. &lt;br /&gt;
Precisely, for any finite subset &amp;lt;math&amp;gt;I \subseteq\mathbb{N}_+&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\{X_i\}_{i \in I}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; are mutually independent. Let &amp;lt;math&amp;gt;X = \sum_{i=1}^N X_i&amp;lt;/math&amp;gt;, show that &amp;lt;math&amp;gt;\textbf{Var}[X] = \textbf{Var}[X_1] \mathbb{E}[N] + \mathbb{E}[X_1]^2 \textbf{Var}[N]&amp;lt;/math&amp;gt;.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Moments (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Show that [math]G(t) = \frac{e^t}{4} + \frac{e^{-t}}{2} + \frac{1}{4}[/math] is a moment-generating function of a random variable, and write the probability mass function of this random variable.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Moments (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X\sim \text{Geo}(p)&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt;. Find &amp;lt;math&amp;gt;\mathbb{E}[X^3]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbb{E}[X^4]&amp;lt;/math&amp;gt;. &lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Moments (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X\sim \text{Pois}(\lambda)&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;\lambda &amp;gt;0 &amp;lt;/math&amp;gt;. Find &amp;lt;math&amp;gt;\mathbb{E}[X^3]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbb{E}[X^4]&amp;lt;/math&amp;gt;. &lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be discrete random variables with correlation &amp;lt;math&amp;gt;\rho&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;|\rho|= 1&amp;lt;/math&amp;gt; if and only if &amp;lt;math&amp;gt;X=aY+b&amp;lt;/math&amp;gt; for some real numbers &amp;lt;math&amp;gt;a,b&amp;lt;/math&amp;gt;.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
    Let [math]X[/math] and [math]Y[/math] be discrete random variables with mean &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;, variance &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;, and correlation [math]\rho[/math]. Show that [math]\mathbb{E}(\max\{X^2,Y^2\})\leq 1+\sqrt{1-\rho^2}[/math]. (Hint: use the identity [math]\max\{a,b\} = \frac{1}{2}(a+b+|a-b|)[/math].)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
   Let [math]X[/math] and [math]Y[/math] be independent Bernoulli random variables with parameter [math]1/2[/math]. Show that [math]X+Y[/math] and [math]|X-Y|[/math] are dependent though uncorrelated.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Inequalities) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Reverse Markov&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable with bounded range &amp;lt;math&amp;gt;0 \le X \le U&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;U &amp;gt; 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbf{Pr}(X \le a) \le \frac{U-\mathbf{E}[X]}{U-a}&amp;lt;/math&amp;gt; for any &amp;lt;math&amp;gt;0 &amp;lt; a &amp;lt; U&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Markov&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable. Show that for all &amp;lt;math&amp;gt;\beta \geq 0&amp;lt;/math&amp;gt; and all &amp;lt;math&amp;gt;x &amp;gt; 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}(X\geq x)\leq \mathbb{E}(e^{\beta X})e^{-\beta x}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Cantelli&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable with mean &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; and variance &amp;lt;math&amp;gt;\sigma^2&amp;lt;/math&amp;gt;. Prove that for any &amp;lt;math&amp;gt;\lambda &amp;gt; 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}[X \ge \lambda] \le \frac{\sigma^2}{\lambda^2+\sigma^2}&amp;lt;/math&amp;gt;. (Hint: You may first show that &amp;lt;math&amp;gt;\mathbf{Pr}[X \ge \lambda] \le \frac{\sigma^2 + u^2}{(\lambda + u)^2}&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;u &amp;gt; 0&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; &amp;lt;strong&amp;gt;[Union of events]&amp;lt;/strong&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;A_1,A_2,\ldots,A_n&amp;lt;/math&amp;gt; be events. Define &amp;lt;math&amp;gt;a = \sum_{i=1}^n \mathbf{Pr}[A_i]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;b = \sum_{1 \le i&amp;lt;j\le n} \mathbf{Pr}[A_i \cap A_j]&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbf{Pr}[A_1 \cup A_2 \cup \ldots \cup A_n] \ge \max\left\{2-\frac{a+2b}{a^2}, \frac{a^2}{a+b}\right\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Probability meets graph theory) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;    &lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;4-clique threshold&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Prove that &amp;lt;math&amp;gt;p = n^{-2/3}&amp;lt;/math&amp;gt; is the threshold probability for the existence of 4-clique. &lt;br /&gt;
        Formally, you are required to show that &lt;br /&gt;
        &amp;lt;ol type=&amp;quot;a&amp;quot;&amp;gt;&lt;br /&gt;
            &amp;lt;li&amp;gt;&lt;br /&gt;
                with a probability approaching to &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; (as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity) the Erdős–Rényi random graph &amp;lt;math&amp;gt;\mathbf{G} = \mathbf{G}(n,p)&amp;lt;/math&amp;gt; contains a 4-clique when &amp;lt;math&amp;gt;p = \omega(n^{-2/3})&amp;lt;/math&amp;gt;; (Hint: use Chebyshev&#039;s inequality.)&lt;br /&gt;
            &amp;lt;/li&amp;gt;&lt;br /&gt;
            &amp;lt;li&amp;gt;&lt;br /&gt;
                with a probability approaching to &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; (as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity) the Erdős–Rényi random graph &amp;lt;math&amp;gt;\mathbf{G} = \mathbf{G}(n,p)&amp;lt;/math&amp;gt; contains a 4-clique when &amp;lt;math&amp;gt;p = o(n^{-2/3})&amp;lt;/math&amp;gt;. (Hint: use Markov&#039;s inequality.)&lt;br /&gt;
            &amp;lt;/li&amp;gt;&lt;br /&gt;
        &amp;lt;/ol&amp;gt;&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_3&amp;diff=13044</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 3</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_3&amp;diff=13044"/>
		<updated>2025-04-06T10:49:52Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: /* Problem 2 (Inequalities) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 3==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Warm-up Problems) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
      Let &amp;lt;math&amp;gt;X_1,X_2,...,X_n&amp;lt;/math&amp;gt; be independent random variables, and suppose that &amp;lt;math&amp;gt;X_k&amp;lt;/math&amp;gt; is Bernoulli with parameter &amp;lt;math&amp;gt;p_k&amp;lt;/math&amp;gt;. Let &amp;lt;math&amp;gt;Y= X_1 + X_2 + \dots + X_n&amp;lt;/math&amp;gt;. Show that, for &amp;lt;math&amp;gt;\mathbb E[Y]&amp;lt;/math&amp;gt; fixed, &amp;lt;math&amp;gt;\mathrm{Var}(Y )&amp;lt;/math&amp;gt; is a maximized when &amp;lt;math&amp;gt;p_1 = p_2 = \dots = p_n&amp;lt;/math&amp;gt;. That is to say, the variation in the sum is greatest when individuals are most alike.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Each member of a group of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; players rolls a (fair) 6-sided die. For any pair of players who throw the same number, the group scores &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; point. Find the mean and variance of the total score of the group.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        An urn contains &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; balls numbered &amp;lt;math&amp;gt;1, 2, \ldots, n&amp;lt;/math&amp;gt;. We select &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; balls uniformly at random &amp;lt;strong&amp;gt;without replacement&amp;lt;/strong&amp;gt; and add up their numbers. Find the mean and variance of the sum.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (IV)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
      Let &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; be an integer-valued, positive random variable and let &amp;lt;math&amp;gt;\{X_i\}_{i=1}^{\infty}&amp;lt;/math&amp;gt; be indepedently identically distributed random variables that are independent of &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt;, too. &lt;br /&gt;
Precisely, for any finite subset &amp;lt;math&amp;gt;I \subseteq\mathbb{N}_+&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\{X_i\}_{i \in I}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; are mutually independent. Let &amp;lt;math&amp;gt;X = \sum_{i=1}^N X_i&amp;lt;/math&amp;gt;, show that &amp;lt;math&amp;gt;\textbf{Var}[X] = \textbf{Var}[X_1] \mathbb{E}[N] + \mathbb{E}[X_1]^2 \textbf{Var}[N]&amp;lt;/math&amp;gt;.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Moments (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Show that [math]G(t) = \frac{e^t}{4} + \frac{e^{-t}}{2} + \frac{1}{4}[/math] is a moment-generating function of a random variable, and write the probability mass function of this random variable.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Moments (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X\sim \text{Geo}(p)&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt;. Find &amp;lt;math&amp;gt;\mathbb{E}[X^3]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbb{E}[X^4]&amp;lt;/math&amp;gt;. &lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Moments (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X\sim \text{Pois}(\lambda)&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;\lambda &amp;gt;0 &amp;lt;/math&amp;gt;. Find &amp;lt;math&amp;gt;\mathbb{E}[X^3]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbb{E}[X^4]&amp;lt;/math&amp;gt;. &lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be discrete random variables with correlation &amp;lt;math&amp;gt;\rho&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;|\rho|= 1&amp;lt;/math&amp;gt; if and only if &amp;lt;math&amp;gt;X=aY+b&amp;lt;/math&amp;gt; for some real numbers &amp;lt;math&amp;gt;a,b&amp;lt;/math&amp;gt;.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
    Let [math]X[/math] and [math]Y[/math] be discrete random variables with mean &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;, variance &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;, and correlation [math]\rho[/math]. Show that [math]\mathbb{E}(\max\{X^2,Y^2\})\leq 1+\sqrt{1-\rho^2}[/math]. (Hint: use the identity [math]\max\{a,b\} = \frac{1}{2}(a+b+|a-b|)[/math].)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
   Let [math]X[/math] and [math]Y[/math] be independent Bernoulli random variables with parameter [math]1/2[/math]. Show that [math]X+Y[/math] and [math]|X-Y|[/math] are dependent though uncorrelated.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Inequalities) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Reverse Markov&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable with bounded range &amp;lt;math&amp;gt;0 \le X \le U&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;U &amp;gt; 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbf{Pr}(X \le a) \le \frac{U-\mathbf{E}[X]}{U-a}&amp;lt;/math&amp;gt; for any &amp;lt;math&amp;gt;0 &amp;lt; a &amp;lt; U&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Markov&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable. Show that for all &amp;lt;math&amp;gt;\beta \geq 0&amp;lt;/math&amp;gt; and all &amp;lt;math&amp;gt;x &amp;gt; 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}(X\geq x)\leq \mathbb{E}(e^{\beta X})e^{-\beta x}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Cantelli&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable with mean &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; and variance &amp;lt;math&amp;gt;\sigma^2&amp;lt;/math&amp;gt;. Prove that for any &amp;lt;math&amp;gt;\lambda &amp;gt; 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}[X \ge \lambda] \le \frac{\sigma^2}{\lambda^2+\sigma^2}&amp;lt;/math&amp;gt;. (Hint: You may first show that &amp;lt;math&amp;gt;\mathbf{Pr}[X \ge \lambda] \le \frac{\sigma^2 + u^2}{(\lambda + u)^2}&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;u &amp;gt; 0&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; &amp;lt;strong&amp;gt;[Union of events]&amp;lt;/strong&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;A_1,A_2,\ldots,A_n&amp;lt;/math&amp;gt; be events. Define &amp;lt;math&amp;gt;a = \sum_{i=1}^n \mathbf{Pr}[A_i]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;b = \sum_{1 \le i&amp;lt;j\le n} \mathbf{Pr}[A_i \cap A_j]&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbf{Pr}[A_1 \cup A_2 \cup \ldots A_n] \ge \max\left\{2-\frac{a+2b}{a^2}, \frac{a^2}{a+b}\right\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Probability meets graph theory) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;    &lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;4-clique threshold&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Prove that &amp;lt;math&amp;gt;p = n^{-2/3}&amp;lt;/math&amp;gt; is the threshold probability for the existence of 4-clique. &lt;br /&gt;
        Formally, you are required to show that &lt;br /&gt;
        &amp;lt;ol type=&amp;quot;a&amp;quot;&amp;gt;&lt;br /&gt;
            &amp;lt;li&amp;gt;&lt;br /&gt;
                with a probability approaching to &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; (as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity) the Erdős–Rényi random graph &amp;lt;math&amp;gt;\mathbf{G} = \mathbf{G}(n,p)&amp;lt;/math&amp;gt; contains a 4-clique when &amp;lt;math&amp;gt;p = \omega(n^{-2/3})&amp;lt;/math&amp;gt;; (Hint: use Chebyshev&#039;s inequality.)&lt;br /&gt;
            &amp;lt;/li&amp;gt;&lt;br /&gt;
            &amp;lt;li&amp;gt;&lt;br /&gt;
                with a probability approaching to &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; (as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity) the Erdős–Rényi random graph &amp;lt;math&amp;gt;\mathbf{G} = \mathbf{G}(n,p)&amp;lt;/math&amp;gt; contains a 4-clique when &amp;lt;math&amp;gt;p = o(n^{-2/3})&amp;lt;/math&amp;gt;. (Hint: use Markov&#039;s inequality.)&lt;br /&gt;
            &amp;lt;/li&amp;gt;&lt;br /&gt;
        &amp;lt;/ol&amp;gt;&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12956</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 2</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12956"/>
		<updated>2025-03-13T15:31:13Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: /* Problem 5 (Probability meets graph theory) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 2==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Warm-up problems, 12 points) ==&lt;br /&gt;
* [&#039;&#039;&#039;Function of random variable (I)&#039;&#039;&#039;] Express the distribution functions of [math]X^+ = \max\{0,X\}[/math], [math]X^- = -\min\{0,X\}[/math], [math]|X|=X^+ + X^-[/math], [math]-X[/math], in terms of the distribution function [math]F[/math] of the random variable [math]X[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Function of random variable (II)&#039;&#039;&#039;] Let [math]X[/math] be a random variable with distribution function [math]\max(0,\min(1,x))[/math]. Let [math]F[/math] be a distribution function which is continuous and strictly increasing. Show that [math]Y=F^{-1}(X)[/math] be a random variable with distribution function [math]F[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Independence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;1\leq r\leq n&amp;lt;/math&amp;gt; be independent random variables which are symmetric about &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;; that is, &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-X_r&amp;lt;/math&amp;gt; have the same distributions. Show that, for all &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}[S_n \geq x] = \mathbf{Pr}[S_n \leq -x]&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;S_n = \sum_{r=1}^n X_r&amp;lt;/math&amp;gt;. Is the conclusion true without the assumtion of independence?&lt;br /&gt;
* [&#039;&#039;&#039;Dependence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be discrete random variables with joint mass function &amp;lt;math&amp;gt;f(x,y) = \frac{C}{(x+y-1)(x+y)(x+y+1)}&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;x,y \in \mathbb{N}_+&amp;lt;/math&amp;gt; (in other words, &amp;lt;math&amp;gt;x,y = 1,2,3,\cdots&amp;lt;/math&amp;gt;). Find (1) the value of &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;, (2) marginal mass function of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and (3) &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt;.&lt;br /&gt;
* [&#039;&#039;&#039;Expectation&#039;&#039;&#039;] It is required to place in order &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; books &amp;lt;math&amp;gt;B_1, B_2, \cdots, B_n&amp;lt;/math&amp;gt; on a library shelf in such way that readers searching from left to right waste as little time as possible on average. Assuming that a random reader requires book &amp;lt;math&amp;gt;B_i&amp;lt;/math&amp;gt; with probability &amp;lt;math&amp;gt;p_i&amp;lt;/math&amp;gt;, find the ordering of the books which minimizes the number of titles examined by a random reader before discovery of the required book.&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Entropy of discrete random variable&amp;lt;/strong&amp;gt;] Let [math]X[/math] be a discrete random variable with range of values [math][N] = \{1,2,\ldots,N\}[/math] and probability mass function [math]p[/math]. Define [math]H(X) = -\sum_{n \ge 1} p(n) \log p(n)[/math] with convention [math]0\log 0 = 0[/math]. Prove that [math]H(X) \le \log N[/math] using Jensen&#039;s inequality.&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Law of total expectation&amp;lt;/strong&amp;gt;] Let [math]X \sim \mathrm{Geom}(p)[/math] for some parameter [math]p \in (0,1)[/math]. Calculate [math]\mathbf{E}[X][/math] using the law of total expectation.&lt;br /&gt;
&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Random number of random variables&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\{X_n\}_{n \ge 1}&amp;lt;/math&amp;gt; be identically distributed random variable and &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; be a random variable taking values in the non-negative integers and independent of the &amp;lt;math&amp;gt;X_n&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;n \ge 1&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{E}\left[\sum_{i=1}^N X_i\right] = \mathbf{E}[N] \mathbf{E}[X_1]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Distribution of random variable, 8 points) ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Cumulative distribution function (CDF)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a random variable with cumulative distribution function &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Show that &amp;lt;math&amp;gt;Y = aX+b&amp;lt;/math&amp;gt; is a random variable where &amp;lt;math&amp;gt;a&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;b&amp;lt;/math&amp;gt; are real constants, and express the CDF of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; by &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;. (Hint: Try expressing the event [math]Y=aX+b\le y[/math] by countably many set operations on the events defined on [math]X[/math].)&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Let &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; be the CDF of random variable &amp;lt;math&amp;gt;Z:\Omega\rightarrow \mathbb{R}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;0\leq \lambda \leq 1&amp;lt;/math&amp;gt;, show that &lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;math&amp;gt;\lambda F + (1-\lambda)G&amp;lt;/math&amp;gt; is a CDF function. &amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;The product &amp;lt;math&amp;gt;FG&amp;lt;/math&amp;gt; is a CDF function, and if &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; are independent, then &amp;lt;math&amp;gt;FG&amp;lt;/math&amp;gt; is the CDF of &amp;lt;math&amp;gt;\max\{X,Z\}&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Probability mass function (PMF)&amp;lt;/strong&amp;gt;] We toss &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; coins, and each one shows heads with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;, independently of each of the others. Each coin which shows head is tossed again. (If the coin shows tail, it won&#039;t be tossed again.) Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of heads resulting from the &amp;lt;strong&amp;gt;second&amp;lt;/strong&amp;gt; round of tosses, and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the number of heads resulting from &amp;lt;strong&amp;gt;all&amp;lt;/strong&amp;gt; tosses, which includes the first and (possible) second round of each toss.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the PMF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{E}[Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Let &amp;lt;math&amp;gt;p_X&amp;lt;/math&amp;gt; be the PMF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, show that [math]p_X(k-1)p_X(k+1)\leq [p_X(k)]^2[/math] for &amp;lt;math&amp;gt;1\leq k \leq n-1&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Discrete random variable) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (I)&amp;lt;/strong&amp;gt;] Every package of some intrinsically dull commodity includes a small and exciting plastic object. There are [math]c[/math] different types of object, and each package is equally likely to contain any given type. You buy one package each day.&amp;lt;/p&amp;gt;&lt;br /&gt;
    &amp;lt;ol&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse between the acquisitions of the [math]j[/math]-th new type of object and the [math](j + 1)[/math]-th new type.&amp;lt;/li&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse before you have a full set of objects.&amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (II)&amp;lt;/strong&amp;gt;] Prove that geometric distribution is the only discrete memoryless distribution with range values &amp;lt;math&amp;gt;\mathbb{N}_+&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;n_1,n_2 \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;0 \le p \le 1&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Bin}(n_1,p),Y \sim \mathrm{Bin}(n_2,p)&amp;lt;/math&amp;gt; be independent random variables. Prove that &amp;lt;math&amp;gt;X+Y \sim \mathrm{Bin}(n_1+n_2,p)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Negative binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; follows the negative binomial distribution with parameter &amp;lt;math&amp;gt;r \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt;. Calculate &amp;lt;math&amp;gt;\mathbf{Var}[X] = \mathbf{E}[X^2] - \left(\mathbf{E}[X]\right)^2&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Hypergeometric distribution&amp;lt;/strong&amp;gt;] An urn contains [math]N[/math] balls, [math]b[/math] of which are blue and [math]r = N -b[/math] of which are red. A random sample of [math]n[/math] balls is drawn &amp;lt;strong&amp;gt;without replacement&amp;lt;/strong&amp;gt; (无放回) from the urn. Let [math]B[/math] the number of blue balls in this sample. Show that if [math]N, b[/math], and [math]r[/math] approach [math]+\infty[/math] in such a way that [math]b/N \rightarrow  p[/math] and [math]r/N \rightarrow 1 - p[/math], then [math]\mathbf{Pr}(B = k) \rightarrow {n\choose k}p^k(1-p)^{n-k}[/math] for [math]0\leq k \leq n[/math].&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Poisson distribution&amp;lt;/strong&amp;gt;] In your pocket is a random number &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; of coins, where &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; has the Poisson distribution with parameter &amp;lt;math&amp;gt;\lambda&amp;lt;/math&amp;gt;. You toss each coin once, with heads showing with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; each time. Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the (random) number of heads outcomes and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the (also random) number of tails.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the joint mass function of &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find PMF of the marginal distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; in &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;. Are &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; independent?&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Conditional distribution (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be independent &amp;lt;math&amp;gt;\text{Bin}(n, p)&amp;lt;/math&amp;gt; random variables, and let &amp;lt;math&amp;gt;Z = X + Y&amp;lt;/math&amp;gt;. Show that the conditional distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; given &amp;lt;math&amp;gt;Z = N&amp;lt;/math&amp;gt; is the hypergeometric distribution.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Conditional distribution (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;\lambda,\mu &amp;gt; 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;n \in \mathbb{N}&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Pois}(\lambda), Y \sim \mathrm{Pois}(\mu)&amp;lt;/math&amp;gt; be independent random variables. Find out the conditional distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, given &amp;lt;math&amp;gt;X+Y = n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 4 (Linearity of Expectation, 12 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Streak&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Suppose we flip a fair coin &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; times independently to obtain a sequence of flips &amp;lt;math&amp;gt;X_1, X_2, \ldots , X_n&amp;lt;/math&amp;gt;. A streak of flips is a consecutive subsequence of flips that are all the same. For example, if &amp;lt;math&amp;gt;X_3&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;X_4&amp;lt;/math&amp;gt;, and &amp;lt;math&amp;gt;X_5&amp;lt;/math&amp;gt; are all heads, there is a streak of length &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; starting at the third lip. (If &amp;lt;math&amp;gt;X_6&amp;lt;/math&amp;gt; is also heads, then there is also a streak of length &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; starting at the third lip.) Find the expected number of streaks of length &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; for some integer &amp;lt;math&amp;gt;k \ge 1&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Number of cycles&amp;lt;/strong&amp;gt;] &lt;br /&gt;
At a banquet, there are &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people who shake hands according to the following process: In each round, two idle hands are randomly selected and shaken (&amp;lt;strong&amp;gt;these two hands are no longer idle&amp;lt;/strong&amp;gt;). After &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds, there will be no idle hands left, and the &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people will form several cycles. For example, when &amp;lt;math&amp;gt;n=3&amp;lt;/math&amp;gt;, the following situation may occur: the left and right hands of the first person are held together, the left hand of the second person and the right hand of the third person are held together, and the right hand of the second person and the left hand of the third person are held together. In this case, three people form two cycles. How many cycles are expected to be formed after &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds?&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Number of operations&amp;lt;/strong&amp;gt;]&lt;br /&gt;
We have a directed graph &amp;lt;math&amp;gt;G=(V,E)&amp;lt;/math&amp;gt; without self-loops and multi-edges.&lt;br /&gt;
&lt;br /&gt;
Until the graph becomes empty, repeat the following operation:&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
Choose one (unerased) vertex uniformly at random (independently from the previous choices). Then, erase that vertex and all vertices that are reachable from the chosen vertex by traversing some edges. Erasing a vertex will also erase the edges incident to it.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;R_v&amp;lt;/math&amp;gt; denotes the number of vertices from which vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt; is reachable. Prove that the expected value of the number of operations equals to &amp;lt;math&amp;gt;\sum_{v \in V} \frac{1}{R_v}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Expected Mex&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X_1,X_2,\ldots,X_{100} \sim \mathrm{Geo}(1/2)&amp;lt;/math&amp;gt; be independent random variables. Compute &amp;lt;math&amp;gt;\mathbf{E}[\mathrm{mex}(X_1,X_2,\ldots,X_{100})]&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\mathrm{mex}(a_1,a_2,\ldots,a_n)&amp;lt;/math&amp;gt; is the smallest positive integer that does not appear in &amp;lt;math&amp;gt;a_1,a_2,\ldots,a_n&amp;lt;/math&amp;gt;. Your answer is considered correct if the absolute error does not exceed &amp;lt;math&amp;gt;10^{-6}&amp;lt;/math&amp;gt;. (Hint: Although mathematics will help you arrive at elegant and efficient methods, the use of a computer and programming skills will be required.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 5 (Probability meets graph theory) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random social networks&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G = (V, E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph without isolating vertex. &lt;br /&gt;
Let &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; be the degree of vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Let &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be a uniformly chosen vertex, and &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; a uniformly chosen neighbor of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Show that &amp;lt;math&amp;gt;\mathbf{E}[d_Z] \geq \mathbf{E}[d_Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Interpret this inequality in the context of social networks, in which the vertices represent people, and the edges represent friendship.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Turán&#039;s Theorem&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G=(V,E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph, and write &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; for the degree of the vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Use probablistic method to prove that &amp;lt;math&amp;gt;\alpha(G) \ge \sum_{v \in V} \frac{1}{d_v + 1}&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\alpha(G)&amp;lt;/math&amp;gt; is the size of a maximum independent set. (Hint: Consider the following random procedure for generating an independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; from a graph with vertex set &amp;lt;math&amp;gt;V&amp;lt;/math&amp;gt;: First, generate a random permutation of the vertices, denoted as &amp;lt;math&amp;gt;v_1,v_2,\ldots,v_n&amp;lt;/math&amp;gt;. Then, construct the independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; as follows: For each vertex &amp;lt;math&amp;gt;v_i \in V&amp;lt;/math&amp;gt;, add &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; if and only if none of its predecessors in the permutation, i.e., &amp;lt;math&amp;gt;v_1,\ldots,v_{i-1}&amp;lt;/math&amp;gt;, are neighbors of &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Tournament&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Prove that there is a [https://en.wikipedia.org/wiki/Tournament_(graph_theory) tournament] &amp;lt;math&amp;gt;T&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; players and at least &amp;lt;math&amp;gt;n!/2^{n-1}&amp;lt;/math&amp;gt;Hamiltonian paths.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 6 (1D random walk, 8 points) ==&lt;br /&gt;
Let &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt; be a constant, and &amp;lt;math&amp;gt;\{X_n\}_{n \ge 1}&amp;lt;/math&amp;gt; be independent Bernoulli trials with successful probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;.&lt;br /&gt;
Define &amp;lt;math&amp;gt;S_n = 2\sum_{i=1}^n X_i - n&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;S_0 = 0&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Range of random walk&amp;lt;/strong&amp;gt;] The range &amp;lt;math&amp;gt;R_n&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;S_0, S_1, \ldots, S_n&amp;lt;/math&amp;gt; is defined as the number of distinct values taken by the sequence. Show that &amp;lt;math&amp;gt;\mathbf{Pr}\left(R_n = R_{n-1}+1\right) = \mathbf{Pr}\left(\forall 1 \le i \le n, S_i \neq 0\right)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, and deduce that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n]\to \mathbf{Pr}(\forall i \ge 1, S_i \neq 0)&amp;lt;/math&amp;gt;. Hence show that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n] \to |2p-1|&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Symmetric 1D random walk (III)&amp;lt;/strong&amp;gt;] Suppose &amp;lt;math&amp;gt;p = \frac{1}{2}&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{E}[|S_n|] = \Theta(\sqrt{n})&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12953</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 2</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12953"/>
		<updated>2025-03-13T15:15:16Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: /* Problem 4 (Linearity of Expectation) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 2==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Warm-up problems, 12 points) ==&lt;br /&gt;
* [&#039;&#039;&#039;Function of random variable (I)&#039;&#039;&#039;] Express the distribution functions of [math]X^+ = \max\{0,X\}[/math], [math]X^- = -\min\{0,X\}[/math], [math]|X|=X^+ + X^-[/math], [math]-X[/math], in terms of the distribution function [math]F[/math] of the random variable [math]X[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Function of random variable (II)&#039;&#039;&#039;] Let [math]X[/math] be a random variable with distribution function [math]\max(0,\min(1,x))[/math]. Let [math]F[/math] be a distribution function which is continuous and strictly increasing. Show that [math]Y=F^{-1}(X)[/math] be a random variable with distribution function [math]F[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Independence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;1\leq r\leq n&amp;lt;/math&amp;gt; be independent random variables which are symmetric about &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;; that is, &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-X_r&amp;lt;/math&amp;gt; have the same distributions. Show that, for all &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}[S_n \geq x] = \mathbf{Pr}[S_n \leq -x]&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;S_n = \sum_{r=1}^n X_r&amp;lt;/math&amp;gt;. Is the conclusion true without the assumtion of independence?&lt;br /&gt;
* [&#039;&#039;&#039;Dependence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be discrete random variables with joint mass function &amp;lt;math&amp;gt;f(x,y) = \frac{C}{(x+y-1)(x+y)(x+y+1)}&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;x,y \in \mathbb{N}_+&amp;lt;/math&amp;gt; (in other words, &amp;lt;math&amp;gt;x,y = 1,2,3,\cdots&amp;lt;/math&amp;gt;). Find (1) the value of &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;, (2) marginal mass function of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and (3) &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt;.&lt;br /&gt;
* [&#039;&#039;&#039;Expectation&#039;&#039;&#039;] Is it generally true that [math]\mathbf{E}[1/X] = 1/\mathbf{E}[X][/math]? Is it ever true that [math]\mathbf{E}[1/X] = 1/\mathbf{E}[X][/math]?&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Entropy of discrete random variable&amp;lt;/strong&amp;gt;] Let [math]X[/math] be a discrete random variable with range of values [math][N] = \{1,2,\ldots,N\}[/math] and probability mass function [math]p[/math]. Define [math]H(X) = -\sum_{n \ge 1} p(n) \log p(n)[/math] with convention [math]0\log 0 = 0[/math]. Prove that [math]H(X) \le \log N[/math] using Jensen&#039;s inequality.&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Law of total expectation&amp;lt;/strong&amp;gt;] Let [math]X \sim \mathrm{Geom}(p)[/math] for some parameter [math]p \in (0,1)[/math]. Calculate [math]\mathbf{E}[X][/math] using the law of total expectation.&lt;br /&gt;
&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Random number of random variables&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\{X_n\}_{n \ge 1}&amp;lt;/math&amp;gt; be identically distributed random variable and &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; be a random variable taking values in the non-negative integers and independent of the &amp;lt;math&amp;gt;X_n&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;n \ge 1&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{E}\left[\sum_{i=1}^N X_i\right] = \mathbf{E}[N] \mathbf{E}[X_1]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Distribution of random variable, 8 points) ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Cumulative distribution function (CDF)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a random variable with cumulative distribution function &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Show that &amp;lt;math&amp;gt;Y = aX+b&amp;lt;/math&amp;gt; is a random variable where &amp;lt;math&amp;gt;a&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;b&amp;lt;/math&amp;gt; are real constants, and express the CDF of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; by &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;. (Hint: Try expressing the event [math]Y=aX+b\le y[/math] by countably many set operations on the events defined on [math]X[/math].)&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Let &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; be the CDF of random variable &amp;lt;math&amp;gt;Z:\Omega\rightarrow \mathbb{R}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;0\leq \lambda \leq 1&amp;lt;/math&amp;gt;, show that &lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;math&amp;gt;\lambda F + (1-\lambda)G&amp;lt;/math&amp;gt; is a CDF function. &amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;The product &amp;lt;math&amp;gt;FG&amp;lt;/math&amp;gt; is a CDF function, and if &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; are independent, then &amp;lt;math&amp;gt;FG&amp;lt;/math&amp;gt; is the CDF of &amp;lt;math&amp;gt;\max\{X,Z\}&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Probability mass function (PMF)&amp;lt;/strong&amp;gt;] We toss &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; coins, and each one shows heads with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;, independently of each of the others. Each coin which shows head is tossed again. (If the coin shows tail, it won&#039;t be tossed again.) Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of heads resulting from the &amp;lt;strong&amp;gt;second&amp;lt;/strong&amp;gt; round of tosses, and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the number of heads resulting from &amp;lt;strong&amp;gt;all&amp;lt;/strong&amp;gt; tosses, which includes the first and (possible) second round of each toss.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the PMF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{E}[Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Let &amp;lt;math&amp;gt;p_X&amp;lt;/math&amp;gt; be the PMF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, show that [math]p_X(k-1)p_X(k+1)\leq [p_X(k)]^2[/math] for &amp;lt;math&amp;gt;1\leq k \leq n-1&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Discrete random variable) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (I)&amp;lt;/strong&amp;gt;] Every package of some intrinsically dull commodity includes a small and exciting plastic object. There are [math]c[/math] different types of object, and each package is equally likely to contain any given type. You buy one package each day.&amp;lt;/p&amp;gt;&lt;br /&gt;
    &amp;lt;ol&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse between the acquisitions of the [math]j[/math]-th new type of object and the [math](j + 1)[/math]-th new type.&amp;lt;/li&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse before you have a full set of objects.&amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (II)&amp;lt;/strong&amp;gt;] Prove that geometric distribution is the only discrete memoryless distribution with range values &amp;lt;math&amp;gt;\mathbb{N}_+&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;n_1,n_2 \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;0 \le p \le 1&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Bin}(n_1,p),Y \sim \mathrm{Bin}(n_2,p)&amp;lt;/math&amp;gt; be independent random variables. Prove that &amp;lt;math&amp;gt;X+Y \sim \mathrm{Bin}(n_1+n_2,p)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Negative binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; follows the negative binomial distribution with parameter &amp;lt;math&amp;gt;r \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt;. Calculate &amp;lt;math&amp;gt;\mathbf{Var}[X] = \mathbf{E}[X^2] - \left(\mathbf{E}[X]\right)^2&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Hypergeometric distribution&amp;lt;/strong&amp;gt;] An urn contains [math]N[/math] balls, [math]b[/math] of which are blue and [math]r = N -b[/math] of which are red. A random sample of [math]n[/math] balls is drawn &amp;lt;strong&amp;gt;without replacement&amp;lt;/strong&amp;gt; (无放回) from the urn. Let [math]B[/math] the number of blue balls in this sample. Show that if [math]N, b[/math], and [math]r[/math] approach [math]+\infty[/math] in such a way that [math]b/N \rightarrow  p[/math] and [math]r/N \rightarrow 1 - p[/math], then [math]\mathbf{Pr}(B = k) \rightarrow {n\choose k}p^k(1-p)^{n-k}[/math] for [math]0\leq k \leq n[/math].&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Poisson distribution&amp;lt;/strong&amp;gt;] In your pocket is a random number &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; of coins, where &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; has the Poisson distribution with parameter &amp;lt;math&amp;gt;\lambda&amp;lt;/math&amp;gt;. You toss each coin once, with heads showing with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; each time. Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the (random) number of heads outcomes and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the (also random) number of tails.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the joint mass function of &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find PMF of the marginal distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; in &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;. Are &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; independent?&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Conditional distribution (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be independent &amp;lt;math&amp;gt;\text{Bin}(n, p)&amp;lt;/math&amp;gt; random variables, and let &amp;lt;math&amp;gt;Z = X + Y&amp;lt;/math&amp;gt;. Show that the conditional distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; given &amp;lt;math&amp;gt;Z = N&amp;lt;/math&amp;gt; is the hypergeometric distribution.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Conditional distribution (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;\lambda,\mu &amp;gt; 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;n \in \mathbb{N}&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Pois}(\lambda), Y \sim \mathrm{Pois}(\mu)&amp;lt;/math&amp;gt; be independent random variables. Find out the conditional distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, given &amp;lt;math&amp;gt;X+Y = n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 4 (Linearity of Expectation, 12 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Streak&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Suppose we flip a fair coin &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; times independently to obtain a sequence of flips &amp;lt;math&amp;gt;X_1, X_2, \ldots , X_n&amp;lt;/math&amp;gt;. A streak of flips is a consecutive subsequence of flips that are all the same. For example, if &amp;lt;math&amp;gt;X_3&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;X_4&amp;lt;/math&amp;gt;, and &amp;lt;math&amp;gt;X_5&amp;lt;/math&amp;gt; are all heads, there is a streak of length &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; starting at the third lip. (If &amp;lt;math&amp;gt;X_6&amp;lt;/math&amp;gt; is also heads, then there is also a streak of length &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; starting at the third lip.) Find the expected number of streaks of length &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; for some integer &amp;lt;math&amp;gt;k \ge 1&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Number of cycles&amp;lt;/strong&amp;gt;] &lt;br /&gt;
At a banquet, there are &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people who shake hands according to the following process: In each round, two idle hands are randomly selected and shaken (&amp;lt;strong&amp;gt;these two hands are no longer idle&amp;lt;/strong&amp;gt;). After &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds, there will be no idle hands left, and the &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people will form several cycles. For example, when &amp;lt;math&amp;gt;n=3&amp;lt;/math&amp;gt;, the following situation may occur: the left and right hands of the first person are held together, the left hand of the second person and the right hand of the third person are held together, and the right hand of the second person and the left hand of the third person are held together. In this case, three people form two cycles. How many cycles are expected to be formed after &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds?&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Number of operations&amp;lt;/strong&amp;gt;]&lt;br /&gt;
We have a directed graph &amp;lt;math&amp;gt;G=(V,E)&amp;lt;/math&amp;gt; without self-loops and multi-edges.&lt;br /&gt;
&lt;br /&gt;
Until the graph becomes empty, repeat the following operation:&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
Choose one (unerased) vertex uniformly at random (independently from the previous choices). Then, erase that vertex and all vertices that are reachable from the chosen vertex by traversing some edges. Erasing a vertex will also erase the edges incident to it.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;R_v&amp;lt;/math&amp;gt; denotes the number of vertices from which vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt; is reachable. Prove that the expected value of the number of operations equals to &amp;lt;math&amp;gt;\sum_{v \in V} \frac{1}{R_v}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Expected Mex&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X_1,X_2,\ldots,X_{100} \sim \mathrm{Geo}(1/2)&amp;lt;/math&amp;gt; be independent random variables. Compute &amp;lt;math&amp;gt;\mathbf{E}[\mathrm{mex}(X_1,X_2,\ldots,X_{100})]&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\mathrm{mex}(a_1,a_2,\ldots,a_n)&amp;lt;/math&amp;gt; is the smallest positive integer that does not appear in &amp;lt;math&amp;gt;a_1,a_2,\ldots,a_n&amp;lt;/math&amp;gt;. Your answer is considered correct if the absolute error does not exceed &amp;lt;math&amp;gt;10^{-6}&amp;lt;/math&amp;gt;. (Hint: Although mathematics will help you arrive at elegant and efficient methods, the use of a computer and programming skills will be required.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 5 (Probability meets graph theory) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random social networks&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G = (V, E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph without isolating vertex. &lt;br /&gt;
Let &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; be the degree of vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Let &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be a uniformly chosen vertex, and &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; a uniformly chosen neighbor of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Show that &amp;lt;math&amp;gt;\mathbf{E}[d_Z] \geq \mathbf{E}[d_Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Interpret this inequality in the context of social networks, in which the vertices represent people, and the edges represent friendship.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Turán&#039;s Theorem&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G=(V,E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph, and write &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; for the degree of the vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Use probablistic method to prove that &amp;lt;math&amp;gt;\alpha(G) \ge \sum_{v \in V} \frac{1}{d_v + 1}&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\alpha(G)&amp;lt;/math&amp;gt; is the size of a maximum independent set. (Hint: Consider the following random procedure for generating an independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; from a graph with vertex set &amp;lt;math&amp;gt;V&amp;lt;/math&amp;gt;: First, generate a random permutation of the vertices, denoted as &amp;lt;math&amp;gt;v_1,v_2,\ldots,v_n&amp;lt;/math&amp;gt;. Then, construct the independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; as follows: For each vertex &amp;lt;math&amp;gt;v_i \in V&amp;lt;/math&amp;gt;, add &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; if and only if none of its predecessors in the permutation, i.e., &amp;lt;math&amp;gt;v_1,\ldots,v_{i-1}&amp;lt;/math&amp;gt;, are neighbors of &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Dominating set&amp;lt;/strong&amp;gt;]&lt;br /&gt;
A &#039;&#039;dominating set&#039;&#039; of vertices in an undirected graph &amp;lt;math&amp;gt;G = (V, E)&amp;lt;/math&amp;gt; is a set &amp;lt;math&amp;gt;S \subseteq V&amp;lt;/math&amp;gt; such that every vertex of&lt;br /&gt;
&amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; belongs to &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; or has a neighbor in &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;. &lt;br /&gt;
&lt;br /&gt;
Let &amp;lt;math&amp;gt;G = (V, E)&amp;lt;/math&amp;gt; be an &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;-vertex graph with minimum degree &amp;lt;math&amp;gt;d &amp;gt; 1&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; has a dominating set with at most &amp;lt;math&amp;gt;\frac{n\left(1+\log(d+1)\right)}{d+1}&amp;lt;/math&amp;gt; vertices. (Hint: Consider a random vertex subset &amp;lt;math&amp;gt;S \subseteq V&amp;lt;/math&amp;gt; by including each vertex independently with&lt;br /&gt;
probability &amp;lt;math&amp;gt;p := \log(d + 1)/(d + 1)&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 6 (1D random walk, 8 points) ==&lt;br /&gt;
Let &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt; be a constant, and &amp;lt;math&amp;gt;\{X_n\}_{n \ge 1}&amp;lt;/math&amp;gt; be independent Bernoulli trials with successful probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;.&lt;br /&gt;
Define &amp;lt;math&amp;gt;S_n = 2\sum_{i=1}^n X_i - n&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;S_0 = 0&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Range of random walk&amp;lt;/strong&amp;gt;] The range &amp;lt;math&amp;gt;R_n&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;S_0, S_1, \ldots, S_n&amp;lt;/math&amp;gt; is defined as the number of distinct values taken by the sequence. Show that &amp;lt;math&amp;gt;\mathbf{Pr}\left(R_n = R_{n-1}+1\right) = \mathbf{Pr}\left(\forall 1 \le i \le n, S_i \neq 0\right)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, and deduce that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n]\to \mathbf{Pr}(\forall i \ge 1, S_i \neq 0)&amp;lt;/math&amp;gt;. Hence show that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n] \to |2p-1|&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Symmetric 1D random walk (III)&amp;lt;/strong&amp;gt;] Suppose &amp;lt;math&amp;gt;p = \frac{1}{2}&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{E}[|S_n|] = \Theta(\sqrt{n})&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12952</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 2</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12952"/>
		<updated>2025-03-13T15:13:02Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: /* Problem 4 (Linearity of Expectation) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 2==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Warm-up problems, 12 points) ==&lt;br /&gt;
* [&#039;&#039;&#039;Function of random variable (I)&#039;&#039;&#039;] Express the distribution functions of [math]X^+ = \max\{0,X\}[/math], [math]X^- = -\min\{0,X\}[/math], [math]|X|=X^+ + X^-[/math], [math]-X[/math], in terms of the distribution function [math]F[/math] of the random variable [math]X[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Function of random variable (II)&#039;&#039;&#039;] Let [math]X[/math] be a random variable with distribution function [math]\max(0,\min(1,x))[/math]. Let [math]F[/math] be a distribution function which is continuous and strictly increasing. Show that [math]Y=F^{-1}(X)[/math] be a random variable with distribution function [math]F[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Independence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;1\leq r\leq n&amp;lt;/math&amp;gt; be independent random variables which are symmetric about &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;; that is, &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-X_r&amp;lt;/math&amp;gt; have the same distributions. Show that, for all &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}[S_n \geq x] = \mathbf{Pr}[S_n \leq -x]&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;S_n = \sum_{r=1}^n X_r&amp;lt;/math&amp;gt;. Is the conclusion true without the assumtion of independence?&lt;br /&gt;
* [&#039;&#039;&#039;Dependence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be discrete random variables with joint mass function &amp;lt;math&amp;gt;f(x,y) = \frac{C}{(x+y-1)(x+y)(x+y+1)}&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;x,y \in \mathbb{N}_+&amp;lt;/math&amp;gt; (in other words, &amp;lt;math&amp;gt;x,y = 1,2,3,\cdots&amp;lt;/math&amp;gt;). Find (1) the value of &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;, (2) marginal mass function of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and (3) &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt;.&lt;br /&gt;
* [&#039;&#039;&#039;Expectation&#039;&#039;&#039;] Is it generally true that [math]\mathbf{E}[1/X] = 1/\mathbf{E}[X][/math]? Is it ever true that [math]\mathbf{E}[1/X] = 1/\mathbf{E}[X][/math]?&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Entropy of discrete random variable&amp;lt;/strong&amp;gt;] Let [math]X[/math] be a discrete random variable with range of values [math][N] = \{1,2,\ldots,N\}[/math] and probability mass function [math]p[/math]. Define [math]H(X) = -\sum_{n \ge 1} p(n) \log p(n)[/math] with convention [math]0\log 0 = 0[/math]. Prove that [math]H(X) \le \log N[/math] using Jensen&#039;s inequality.&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Law of total expectation&amp;lt;/strong&amp;gt;] Let [math]X \sim \mathrm{Geom}(p)[/math] for some parameter [math]p \in (0,1)[/math]. Calculate [math]\mathbf{E}[X][/math] using the law of total expectation.&lt;br /&gt;
&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Random number of random variables&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\{X_n\}_{n \ge 1}&amp;lt;/math&amp;gt; be identically distributed random variable and &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; be a random variable taking values in the non-negative integers and independent of the &amp;lt;math&amp;gt;X_n&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;n \ge 1&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{E}\left[\sum_{i=1}^N X_i\right] = \mathbf{E}[N] \mathbf{E}[X_1]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Distribution of random variable, 8 points) ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Cumulative distribution function (CDF)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a random variable with cumulative distribution function &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Show that &amp;lt;math&amp;gt;Y = aX+b&amp;lt;/math&amp;gt; is a random variable where &amp;lt;math&amp;gt;a&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;b&amp;lt;/math&amp;gt; are real constants, and express the CDF of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; by &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;. (Hint: Try expressing the event [math]Y=aX+b\le y[/math] by countably many set operations on the events defined on [math]X[/math].)&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Let &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; be the CDF of random variable &amp;lt;math&amp;gt;Z:\Omega\rightarrow \mathbb{R}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;0\leq \lambda \leq 1&amp;lt;/math&amp;gt;, show that &lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;math&amp;gt;\lambda F + (1-\lambda)G&amp;lt;/math&amp;gt; is a CDF function. &amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;The product &amp;lt;math&amp;gt;FG&amp;lt;/math&amp;gt; is a CDF function, and if &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; are independent, then &amp;lt;math&amp;gt;FG&amp;lt;/math&amp;gt; is the CDF of &amp;lt;math&amp;gt;\max\{X,Z\}&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Probability mass function (PMF)&amp;lt;/strong&amp;gt;] We toss &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; coins, and each one shows heads with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;, independently of each of the others. Each coin which shows head is tossed again. (If the coin shows tail, it won&#039;t be tossed again.) Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of heads resulting from the &amp;lt;strong&amp;gt;second&amp;lt;/strong&amp;gt; round of tosses, and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the number of heads resulting from &amp;lt;strong&amp;gt;all&amp;lt;/strong&amp;gt; tosses, which includes the first and (possible) second round of each toss.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the PMF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{E}[Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Let &amp;lt;math&amp;gt;p_X&amp;lt;/math&amp;gt; be the PMF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, show that [math]p_X(k-1)p_X(k+1)\leq [p_X(k)]^2[/math] for &amp;lt;math&amp;gt;1\leq k \leq n-1&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Discrete random variable) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (I)&amp;lt;/strong&amp;gt;] Every package of some intrinsically dull commodity includes a small and exciting plastic object. There are [math]c[/math] different types of object, and each package is equally likely to contain any given type. You buy one package each day.&amp;lt;/p&amp;gt;&lt;br /&gt;
    &amp;lt;ol&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse between the acquisitions of the [math]j[/math]-th new type of object and the [math](j + 1)[/math]-th new type.&amp;lt;/li&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse before you have a full set of objects.&amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (II)&amp;lt;/strong&amp;gt;] Prove that geometric distribution is the only discrete memoryless distribution with range values &amp;lt;math&amp;gt;\mathbb{N}_+&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;n_1,n_2 \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;0 \le p \le 1&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Bin}(n_1,p),Y \sim \mathrm{Bin}(n_2,p)&amp;lt;/math&amp;gt; be independent random variables. Prove that &amp;lt;math&amp;gt;X+Y \sim \mathrm{Bin}(n_1+n_2,p)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Negative binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; follows the negative binomial distribution with parameter &amp;lt;math&amp;gt;r \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt;. Calculate &amp;lt;math&amp;gt;\mathbf{Var}[X] = \mathbf{E}[X^2] - \left(\mathbf{E}[X]\right)^2&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Hypergeometric distribution&amp;lt;/strong&amp;gt;] An urn contains [math]N[/math] balls, [math]b[/math] of which are blue and [math]r = N -b[/math] of which are red. A random sample of [math]n[/math] balls is drawn &amp;lt;strong&amp;gt;without replacement&amp;lt;/strong&amp;gt; (无放回) from the urn. Let [math]B[/math] the number of blue balls in this sample. Show that if [math]N, b[/math], and [math]r[/math] approach [math]+\infty[/math] in such a way that [math]b/N \rightarrow  p[/math] and [math]r/N \rightarrow 1 - p[/math], then [math]\mathbf{Pr}(B = k) \rightarrow {n\choose k}p^k(1-p)^{n-k}[/math] for [math]0\leq k \leq n[/math].&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Poisson distribution&amp;lt;/strong&amp;gt;] In your pocket is a random number &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; of coins, where &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; has the Poisson distribution with parameter &amp;lt;math&amp;gt;\lambda&amp;lt;/math&amp;gt;. You toss each coin once, with heads showing with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; each time. Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the (random) number of heads outcomes and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the (also random) number of tails.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the joint mass function of &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find PMF of the marginal distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; in &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;. Are &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; independent?&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Conditional distribution (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be independent &amp;lt;math&amp;gt;\text{Bin}(n, p)&amp;lt;/math&amp;gt; random variables, and let &amp;lt;math&amp;gt;Z = X + Y&amp;lt;/math&amp;gt;. Show that the conditional distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; given &amp;lt;math&amp;gt;Z = N&amp;lt;/math&amp;gt; is the hypergeometric distribution.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Conditional distribution (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;\lambda,\mu &amp;gt; 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;n \in \mathbb{N}&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Pois}(\lambda), Y \sim \mathrm{Pois}(\mu)&amp;lt;/math&amp;gt; be independent random variables. Find out the conditional distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, given &amp;lt;math&amp;gt;X+Y = n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 4 (Linearity of Expectation) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Streak&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Suppose we flip a fair coin &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; times independently to obtain a sequence of flips &amp;lt;math&amp;gt;X_1, X_2, \ldots , X_n&amp;lt;/math&amp;gt;. A streak of flips is a consecutive subsequence of flips that are all the same. For example, if &amp;lt;math&amp;gt;X_3&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;X_4&amp;lt;/math&amp;gt;, and &amp;lt;math&amp;gt;X_5&amp;lt;/math&amp;gt; are all heads, there is a streak of length &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; starting at the third lip. (If &amp;lt;math&amp;gt;X_6&amp;lt;/math&amp;gt; is also heads, then there is also a streak of length &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; starting at the third lip.) Find the expected number of streaks of length &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; for some integer &amp;lt;math&amp;gt;k \ge 1&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Number of cycles&amp;lt;/strong&amp;gt;] &lt;br /&gt;
At a banquet, there are &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people who shake hands according to the following process: In each round, two idle hands are randomly selected and shaken (&amp;lt;strong&amp;gt;these two hands are no longer idle&amp;lt;/strong&amp;gt;). After &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds, there will be no idle hands left, and the &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people will form several cycles. For example, when &amp;lt;math&amp;gt;n=3&amp;lt;/math&amp;gt;, the following situation may occur: the left and right hands of the first person are held together, the left hand of the second person and the right hand of the third person are held together, and the right hand of the second person and the left hand of the third person are held together. In this case, three people form two cycles. How many cycles are expected to be formed after &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds?&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Number of operations&amp;lt;/strong&amp;gt;]&lt;br /&gt;
We have a directed graph &amp;lt;math&amp;gt;G=(V,E)&amp;lt;/math&amp;gt; without self-loops and multi-edges.&lt;br /&gt;
&lt;br /&gt;
Until the graph becomes empty, repeat the following operation:&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
Choose one (unerased) vertex uniformly at random (independently from the previous choices). Then, erase that vertex and all vertices that are reachable from the chosen vertex by traversing some edges. Erasing a vertex will also erase the edges incident to it.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;R_v&amp;lt;/math&amp;gt; denotes the number of vertices from which vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt; is reachable. Prove that the expected value of the number of operations equals to &amp;lt;math&amp;gt;\sum_{v \in V} \frac{1}{R_v}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Expected Mex&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X_1,X_2,\ldots,X_{100} \sim \mathrm{Geo}(1/2)&amp;lt;/math&amp;gt; be independent random variables. Compute &amp;lt;math&amp;gt;\mathbf{E}[\mathrm{mex}(X_1,X_2,\ldots,X_{100})]&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\mathrm{mex}(a_1,a_2,\ldots,a_n)&amp;lt;/math&amp;gt; is the smallest positive integer that does not appear in &amp;lt;math&amp;gt;a_1,a_2,\ldots,a_n&amp;lt;/math&amp;gt;. Your answer is considered correct if the absolute error does not exceed &amp;lt;math&amp;gt;10^{-6}&amp;lt;/math&amp;gt;. (Hint: Although mathematics will help you arrive at elegant and efficient methods, the use of a computer and programming skills will be required.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 5 (Probability meets graph theory) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random social networks&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G = (V, E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph without isolating vertex. &lt;br /&gt;
Let &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; be the degree of vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Let &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be a uniformly chosen vertex, and &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; a uniformly chosen neighbor of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Show that &amp;lt;math&amp;gt;\mathbf{E}[d_Z] \geq \mathbf{E}[d_Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Interpret this inequality in the context of social networks, in which the vertices represent people, and the edges represent friendship.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Turán&#039;s Theorem&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G=(V,E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph, and write &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; for the degree of the vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Use probablistic method to prove that &amp;lt;math&amp;gt;\alpha(G) \ge \sum_{v \in V} \frac{1}{d_v + 1}&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\alpha(G)&amp;lt;/math&amp;gt; is the size of a maximum independent set. (Hint: Consider the following random procedure for generating an independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; from a graph with vertex set &amp;lt;math&amp;gt;V&amp;lt;/math&amp;gt;: First, generate a random permutation of the vertices, denoted as &amp;lt;math&amp;gt;v_1,v_2,\ldots,v_n&amp;lt;/math&amp;gt;. Then, construct the independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; as follows: For each vertex &amp;lt;math&amp;gt;v_i \in V&amp;lt;/math&amp;gt;, add &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; if and only if none of its predecessors in the permutation, i.e., &amp;lt;math&amp;gt;v_1,\ldots,v_{i-1}&amp;lt;/math&amp;gt;, are neighbors of &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Dominating set&amp;lt;/strong&amp;gt;]&lt;br /&gt;
A &#039;&#039;dominating set&#039;&#039; of vertices in an undirected graph &amp;lt;math&amp;gt;G = (V, E)&amp;lt;/math&amp;gt; is a set &amp;lt;math&amp;gt;S \subseteq V&amp;lt;/math&amp;gt; such that every vertex of&lt;br /&gt;
&amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; belongs to &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; or has a neighbor in &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;. &lt;br /&gt;
&lt;br /&gt;
Let &amp;lt;math&amp;gt;G = (V, E)&amp;lt;/math&amp;gt; be an &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;-vertex graph with minimum degree &amp;lt;math&amp;gt;d &amp;gt; 1&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; has a dominating set with at most &amp;lt;math&amp;gt;\frac{n\left(1+\log(d+1)\right)}{d+1}&amp;lt;/math&amp;gt; vertices. (Hint: Consider a random vertex subset &amp;lt;math&amp;gt;S \subseteq V&amp;lt;/math&amp;gt; by including each vertex independently with&lt;br /&gt;
probability &amp;lt;math&amp;gt;p := \log(d + 1)/(d + 1)&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 6 (1D random walk, 8 points) ==&lt;br /&gt;
Let &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt; be a constant, and &amp;lt;math&amp;gt;\{X_n\}_{n \ge 1}&amp;lt;/math&amp;gt; be independent Bernoulli trials with successful probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;.&lt;br /&gt;
Define &amp;lt;math&amp;gt;S_n = 2\sum_{i=1}^n X_i - n&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;S_0 = 0&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Range of random walk&amp;lt;/strong&amp;gt;] The range &amp;lt;math&amp;gt;R_n&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;S_0, S_1, \ldots, S_n&amp;lt;/math&amp;gt; is defined as the number of distinct values taken by the sequence. Show that &amp;lt;math&amp;gt;\mathbf{Pr}\left(R_n = R_{n-1}+1\right) = \mathbf{Pr}\left(\forall 1 \le i \le n, S_i \neq 0\right)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, and deduce that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n]\to \mathbf{Pr}(\forall i \ge 1, S_i \neq 0)&amp;lt;/math&amp;gt;. Hence show that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n] \to |2p-1|&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Symmetric 1D random walk (III)&amp;lt;/strong&amp;gt;] Suppose &amp;lt;math&amp;gt;p = \frac{1}{2}&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{E}[|S_n|] = \Theta(\sqrt{n})&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12951</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 2</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12951"/>
		<updated>2025-03-13T15:09:49Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: /* Problem 4 (Linearity of Expectation, 12 points) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 2==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Warm-up problems, 12 points) ==&lt;br /&gt;
* [&#039;&#039;&#039;Function of random variable (I)&#039;&#039;&#039;] Express the distribution functions of [math]X^+ = \max\{0,X\}[/math], [math]X^- = -\min\{0,X\}[/math], [math]|X|=X^+ + X^-[/math], [math]-X[/math], in terms of the distribution function [math]F[/math] of the random variable [math]X[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Function of random variable (II)&#039;&#039;&#039;] Let [math]X[/math] be a random variable with distribution function [math]\max(0,\min(1,x))[/math]. Let [math]F[/math] be a distribution function which is continuous and strictly increasing. Show that [math]Y=F^{-1}(X)[/math] be a random variable with distribution function [math]F[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Independence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;1\leq r\leq n&amp;lt;/math&amp;gt; be independent random variables which are symmetric about &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;; that is, &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-X_r&amp;lt;/math&amp;gt; have the same distributions. Show that, for all &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}[S_n \geq x] = \mathbf{Pr}[S_n \leq -x]&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;S_n = \sum_{r=1}^n X_r&amp;lt;/math&amp;gt;. Is the conclusion true without the assumtion of independence?&lt;br /&gt;
* [&#039;&#039;&#039;Dependence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be discrete random variables with joint mass function &amp;lt;math&amp;gt;f(x,y) = \frac{C}{(x+y-1)(x+y)(x+y+1)}&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;x,y \in \mathbb{N}_+&amp;lt;/math&amp;gt; (in other words, &amp;lt;math&amp;gt;x,y = 1,2,3,\cdots&amp;lt;/math&amp;gt;). Find (1) the value of &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;, (2) marginal mass function of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and (3) &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt;.&lt;br /&gt;
* [&#039;&#039;&#039;Expectation&#039;&#039;&#039;] Is it generally true that [math]\mathbf{E}[1/X] = 1/\mathbf{E}[X][/math]? Is it ever true that [math]\mathbf{E}[1/X] = 1/\mathbf{E}[X][/math]?&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Entropy of discrete random variable&amp;lt;/strong&amp;gt;] Let [math]X[/math] be a discrete random variable with range of values [math][N] = \{1,2,\ldots,N\}[/math] and probability mass function [math]p[/math]. Define [math]H(X) = -\sum_{n \ge 1} p(n) \log p(n)[/math] with convention [math]0\log 0 = 0[/math]. Prove that [math]H(X) \le \log N[/math] using Jensen&#039;s inequality.&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Law of total expectation&amp;lt;/strong&amp;gt;] Let [math]X \sim \mathrm{Geom}(p)[/math] for some parameter [math]p \in (0,1)[/math]. Calculate [math]\mathbf{E}[X][/math] using the law of total expectation.&lt;br /&gt;
&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Random number of random variables&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\{X_n\}_{n \ge 1}&amp;lt;/math&amp;gt; be identically distributed random variable and &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; be a random variable taking values in the non-negative integers and independent of the &amp;lt;math&amp;gt;X_n&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;n \ge 1&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{E}\left[\sum_{i=1}^N X_i\right] = \mathbf{E}[N] \mathbf{E}[X_1]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Distribution of random variable, 8 points) ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Cumulative distribution function (CDF)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a random variable with cumulative distribution function &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Show that &amp;lt;math&amp;gt;Y = aX+b&amp;lt;/math&amp;gt; is a random variable where &amp;lt;math&amp;gt;a&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;b&amp;lt;/math&amp;gt; are real constants, and express the CDF of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; by &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;. (Hint: Try expressing the event [math]Y=aX+b\le y[/math] by countably many set operations on the events defined on [math]X[/math].)&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Let &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; be the CDF of random variable &amp;lt;math&amp;gt;Z:\Omega\rightarrow \mathbb{R}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;0\leq \lambda \leq 1&amp;lt;/math&amp;gt;, show that &lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;math&amp;gt;\lambda F + (1-\lambda)G&amp;lt;/math&amp;gt; is a CDF function. &amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;The product &amp;lt;math&amp;gt;FG&amp;lt;/math&amp;gt; is a CDF function, and if &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; are independent, then &amp;lt;math&amp;gt;FG&amp;lt;/math&amp;gt; is the CDF of &amp;lt;math&amp;gt;\max\{X,Z\}&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Probability mass function (PMF)&amp;lt;/strong&amp;gt;] We toss &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; coins, and each one shows heads with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;, independently of each of the others. Each coin which shows head is tossed again. (If the coin shows tail, it won&#039;t be tossed again.) Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of heads resulting from the &amp;lt;strong&amp;gt;second&amp;lt;/strong&amp;gt; round of tosses, and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the number of heads resulting from &amp;lt;strong&amp;gt;all&amp;lt;/strong&amp;gt; tosses, which includes the first and (possible) second round of each toss.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the PMF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{E}[Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Let &amp;lt;math&amp;gt;p_X&amp;lt;/math&amp;gt; be the PMF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, show that [math]p_X(k-1)p_X(k+1)\leq [p_X(k)]^2[/math] for &amp;lt;math&amp;gt;1\leq k \leq n-1&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Discrete random variable) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (I)&amp;lt;/strong&amp;gt;] Every package of some intrinsically dull commodity includes a small and exciting plastic object. There are [math]c[/math] different types of object, and each package is equally likely to contain any given type. You buy one package each day.&amp;lt;/p&amp;gt;&lt;br /&gt;
    &amp;lt;ol&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse between the acquisitions of the [math]j[/math]-th new type of object and the [math](j + 1)[/math]-th new type.&amp;lt;/li&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse before you have a full set of objects.&amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (II)&amp;lt;/strong&amp;gt;] Prove that geometric distribution is the only discrete memoryless distribution with range values &amp;lt;math&amp;gt;\mathbb{N}_+&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;n_1,n_2 \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;0 \le p \le 1&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Bin}(n_1,p),Y \sim \mathrm{Bin}(n_2,p)&amp;lt;/math&amp;gt; be independent random variables. Prove that &amp;lt;math&amp;gt;X+Y \sim \mathrm{Bin}(n_1+n_2,p)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Negative binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; follows the negative binomial distribution with parameter &amp;lt;math&amp;gt;r \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt;. Calculate &amp;lt;math&amp;gt;\mathbf{Var}[X] = \mathbf{E}[X^2] - \left(\mathbf{E}[X]\right)^2&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Hypergeometric distribution&amp;lt;/strong&amp;gt;] An urn contains [math]N[/math] balls, [math]b[/math] of which are blue and [math]r = N -b[/math] of which are red. A random sample of [math]n[/math] balls is drawn &amp;lt;strong&amp;gt;without replacement&amp;lt;/strong&amp;gt; (无放回) from the urn. Let [math]B[/math] the number of blue balls in this sample. Show that if [math]N, b[/math], and [math]r[/math] approach [math]+\infty[/math] in such a way that [math]b/N \rightarrow  p[/math] and [math]r/N \rightarrow 1 - p[/math], then [math]\mathbf{Pr}(B = k) \rightarrow {n\choose k}p^k(1-p)^{n-k}[/math] for [math]0\leq k \leq n[/math].&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Poisson distribution&amp;lt;/strong&amp;gt;] In your pocket is a random number &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; of coins, where &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; has the Poisson distribution with parameter &amp;lt;math&amp;gt;\lambda&amp;lt;/math&amp;gt;. You toss each coin once, with heads showing with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; each time. Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the (random) number of heads outcomes and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the (also random) number of tails.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the joint mass function of &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find PMF of the marginal distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; in &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;. Are &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; independent?&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Conditional distribution (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be independent &amp;lt;math&amp;gt;\text{Bin}(n, p)&amp;lt;/math&amp;gt; random variables, and let &amp;lt;math&amp;gt;Z = X + Y&amp;lt;/math&amp;gt;. Show that the conditional distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; given &amp;lt;math&amp;gt;Z = N&amp;lt;/math&amp;gt; is the hypergeometric distribution.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Conditional distribution (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;\lambda,\mu &amp;gt; 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;n \in \mathbb{N}&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Pois}(\lambda), Y \sim \mathrm{Pois}(\mu)&amp;lt;/math&amp;gt; be independent random variables. Find out the conditional distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, given &amp;lt;math&amp;gt;X+Y = n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 4 (Linearity of Expectation) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Streak&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Suppose we flip a fair coin &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; times independently to obtain a sequence of flips &amp;lt;math&amp;gt;X_1, X_2, \ldots , X_n&amp;lt;/math&amp;gt;. A streak of flips is a consecutive subsequence of flips that are all the same. For example, if &amp;lt;math&amp;gt;X_3&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;X_4&amp;lt;/math&amp;gt;, and &amp;lt;math&amp;gt;X_5&amp;lt;/math&amp;gt; are all heads, there is a streak of length &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; starting at the third lip. (If &amp;lt;math&amp;gt;X_6&amp;lt;/math&amp;gt; is also heads, then there is also a streak of length &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; starting at the third lip.) Find the expected number of streaks of length &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; for some integer &amp;lt;math&amp;gt;k \ge 1&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Number of cycles&amp;lt;/strong&amp;gt;] &lt;br /&gt;
At a banquet, there are &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people who shake hands according to the following process: In each round, two idle hands are randomly selected and shaken (&amp;lt;strong&amp;gt;these two hands are no longer idle&amp;lt;/strong&amp;gt;). After &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds, there will be no idle hands left, and the &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people will form several cycles. For example, when &amp;lt;math&amp;gt;n=3&amp;lt;/math&amp;gt;, the following situation may occur: the left and right hands of the first person are held together, the left hand of the second person and the right hand of the third person are held together, and the right hand of the second person and the left hand of the third person are held together. In this case, three people form two cycles. How many cycles are expected to be formed after &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds?&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Number of operations&amp;lt;/strong&amp;gt;]&lt;br /&gt;
We have a directed graph without self-loops and multi-edges.&lt;br /&gt;
&lt;br /&gt;
Until the graph becomes empty, repeat the following operation:&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
Choose one (unerased) vertex uniformly at random (independently from the previous choices). Then, erase that vertex and all vertices that are reachable from the chosen vertex by traversing some edges. Erasing a vertex will also erase the edges incident to it.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
Find the expected value of the number of times the operation is done.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Expected Mex&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X_1,X_2,\ldots,X_{100} \sim \mathrm{Geo}(1/2)&amp;lt;/math&amp;gt; be independent random variables. Compute &amp;lt;math&amp;gt;\mathbf{E}[\mathrm{mex}(X_1,X_2,\ldots,X_{100})]&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\mathrm{mex}(a_1,a_2,\ldots,a_n)&amp;lt;/math&amp;gt; is the smallest positive integer that does not appear in &amp;lt;math&amp;gt;a_1,a_2,\ldots,a_n&amp;lt;/math&amp;gt;. Your answer is considered correct if the absolute error does not exceed &amp;lt;math&amp;gt;10^{-6}&amp;lt;/math&amp;gt;. (Hint: Although mathematics will help you arrive at elegant and efficient methods, the use of a computer and programming skills will be required.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 5 (Probability meets graph theory) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random social networks&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G = (V, E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph without isolating vertex. &lt;br /&gt;
Let &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; be the degree of vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Let &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be a uniformly chosen vertex, and &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; a uniformly chosen neighbor of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Show that &amp;lt;math&amp;gt;\mathbf{E}[d_Z] \geq \mathbf{E}[d_Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Interpret this inequality in the context of social networks, in which the vertices represent people, and the edges represent friendship.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Turán&#039;s Theorem&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G=(V,E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph, and write &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; for the degree of the vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Use probablistic method to prove that &amp;lt;math&amp;gt;\alpha(G) \ge \sum_{v \in V} \frac{1}{d_v + 1}&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\alpha(G)&amp;lt;/math&amp;gt; is the size of a maximum independent set. (Hint: Consider the following random procedure for generating an independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; from a graph with vertex set &amp;lt;math&amp;gt;V&amp;lt;/math&amp;gt;: First, generate a random permutation of the vertices, denoted as &amp;lt;math&amp;gt;v_1,v_2,\ldots,v_n&amp;lt;/math&amp;gt;. Then, construct the independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; as follows: For each vertex &amp;lt;math&amp;gt;v_i \in V&amp;lt;/math&amp;gt;, add &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; if and only if none of its predecessors in the permutation, i.e., &amp;lt;math&amp;gt;v_1,\ldots,v_{i-1}&amp;lt;/math&amp;gt;, are neighbors of &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Dominating set&amp;lt;/strong&amp;gt;]&lt;br /&gt;
A &#039;&#039;dominating set&#039;&#039; of vertices in an undirected graph &amp;lt;math&amp;gt;G = (V, E)&amp;lt;/math&amp;gt; is a set &amp;lt;math&amp;gt;S \subseteq V&amp;lt;/math&amp;gt; such that every vertex of&lt;br /&gt;
&amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; belongs to &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; or has a neighbor in &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;. &lt;br /&gt;
&lt;br /&gt;
Let &amp;lt;math&amp;gt;G = (V, E)&amp;lt;/math&amp;gt; be an &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;-vertex graph with minimum degree &amp;lt;math&amp;gt;d &amp;gt; 1&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; has a dominating set with at most &amp;lt;math&amp;gt;\frac{n\left(1+\log(d+1)\right)}{d+1}&amp;lt;/math&amp;gt; vertices. (Hint: Consider a random vertex subset &amp;lt;math&amp;gt;S \subseteq V&amp;lt;/math&amp;gt; by including each vertex independently with&lt;br /&gt;
probability &amp;lt;math&amp;gt;p := \log(d + 1)/(d + 1)&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 6 (1D random walk, 8 points) ==&lt;br /&gt;
Let &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt; be a constant, and &amp;lt;math&amp;gt;\{X_n\}_{n \ge 1}&amp;lt;/math&amp;gt; be independent Bernoulli trials with successful probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;.&lt;br /&gt;
Define &amp;lt;math&amp;gt;S_n = 2\sum_{i=1}^n X_i - n&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;S_0 = 0&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Range of random walk&amp;lt;/strong&amp;gt;] The range &amp;lt;math&amp;gt;R_n&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;S_0, S_1, \ldots, S_n&amp;lt;/math&amp;gt; is defined as the number of distinct values taken by the sequence. Show that &amp;lt;math&amp;gt;\mathbf{Pr}\left(R_n = R_{n-1}+1\right) = \mathbf{Pr}\left(\forall 1 \le i \le n, S_i \neq 0\right)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, and deduce that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n]\to \mathbf{Pr}(\forall i \ge 1, S_i \neq 0)&amp;lt;/math&amp;gt;. Hence show that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n] \to |2p-1|&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Symmetric 1D random walk (III)&amp;lt;/strong&amp;gt;] Suppose &amp;lt;math&amp;gt;p = \frac{1}{2}&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{E}[|S_n|] = \Theta(\sqrt{n})&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12947</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 2</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12947"/>
		<updated>2025-03-13T14:58:54Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: /* Problem 4 (Linearity of Expectation, 12 points) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 2==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Warm-up problems, 12 points) ==&lt;br /&gt;
* [&#039;&#039;&#039;Function of random variable (I)&#039;&#039;&#039;] Show that, if &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; are random variables, then so are &amp;lt;math&amp;gt;X+Y&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;XY&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\min\{X,Y\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
* [&#039;&#039;&#039;Function of random variable (II)&#039;&#039;&#039;] Let [math]X[/math] be a random variable with distribution function [math]\max(0,\min(1,x))[/math]. Let [math]F[/math] be a distribution function which is continuous and strictly increasing. Show that [math]Y=F^{-1}(X)[/math] be a random variable with distribution function [math]F[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Independence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;1\leq r\leq n&amp;lt;/math&amp;gt; be independent random variables which are symmetric about &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;; that is, &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-X_r&amp;lt;/math&amp;gt; have the same distributions. Show that, for all &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}[S_n \geq x] = \mathbf{Pr}[S_n \leq -x]&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;S_n = \sum_{r=1}^n X_r&amp;lt;/math&amp;gt;. Is the conclusion true without the assumtion of independence?&lt;br /&gt;
* [&#039;&#039;&#039;Dependence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be discrete random variables with joint mass function &amp;lt;math&amp;gt;f(x,y) = \frac{C}{(x+y-1)(x+y)(x+y+1)}&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;x,y \in \mathbb{N}_+&amp;lt;/math&amp;gt; (in other words, &amp;lt;math&amp;gt;x,y = 1,2,3,\cdots&amp;lt;/math&amp;gt;). Find (1) the value of &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;, (2) marginal mass function of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and (3) &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt;.&lt;br /&gt;
* [&#039;&#039;&#039;Expectation&#039;&#039;&#039;] Is it generally true that [math]\mathbf{E}[1/X] = 1/\mathbf{E}[X][/math]? Is it ever true that [math]\mathbf{E}[1/X] = 1/\mathbf{E}[X][/math]?&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Entropy of discrete random variable&amp;lt;/strong&amp;gt;] Let [math]X[/math] be a discrete random variable with range of values [math][N] = \{1,2,\ldots,N\}[/math] and probability mass function [math]p[/math]. Define [math]H(X) = -\sum_{n \ge 1} p(n) \log p(n)[/math] with convention [math]0\log 0 = 0[/math]. Prove that [math]H(X) \le \log N[/math] using Jensen&#039;s inequality.&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Law of total expectation&amp;lt;/strong&amp;gt;] Let [math]X \sim \mathrm{Geom}(p)[/math] for some parameter [math]p \in (0,1)[/math]. Calculate [math]\mathbf{E}[X][/math] using the law of total expectation.&lt;br /&gt;
&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Random number of random variables&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\{X_n\}_{n \ge 1}&amp;lt;/math&amp;gt; be identically distributed random variable and &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; be a random variable taking values in the non-negative integers and independent of the &amp;lt;math&amp;gt;X_n&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;n \ge 1&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{E}\left[\sum_{i=1}^N X_i\right] = \mathbf{E}[N] \mathbf{E}[X_1]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Distribution of random variable, 8 points) ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Cumulative distribution function (CDF)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a random variable with cumulative distribution function &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Show that &amp;lt;math&amp;gt;Y = aX+b&amp;lt;/math&amp;gt; is a random variable where &amp;lt;math&amp;gt;a&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;b&amp;lt;/math&amp;gt; are real constants, and express the CDF of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; by &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;. (Hint: Try expressing the event [math]Y=aX+b\le y[/math] by countably many set operations on the events defined on [math]X[/math].)&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Let &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; be the CDF of random variable &amp;lt;math&amp;gt;Z:\Omega\rightarrow \mathbb{R}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;0\leq \lambda \leq 1&amp;lt;/math&amp;gt;, show that &lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;math&amp;gt;\lambda F + (1-\lambda)G&amp;lt;/math&amp;gt; is a CDF function. &amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;The product &amp;lt;math&amp;gt;FG&amp;lt;/math&amp;gt; is a CDF function, and if &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; are independent, then &amp;lt;math&amp;gt;FG&amp;lt;/math&amp;gt; is the CDF of &amp;lt;math&amp;gt;\max\{X,Z\}&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Probability mass function (PMF)&amp;lt;/strong&amp;gt;] We toss &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; coins, and each one shows heads with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;, independently of each of the others. Each coin which shows head is tossed again. (If the coin shows tail, it won&#039;t be tossed again.) Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of heads resulting from the &amp;lt;strong&amp;gt;second&amp;lt;/strong&amp;gt; round of tosses, and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the number of heads resulting from &amp;lt;strong&amp;gt;all&amp;lt;/strong&amp;gt; tosses, which includes the first and (possible) second round of each toss.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the PMF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{E}[Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Let &amp;lt;math&amp;gt;p_X&amp;lt;/math&amp;gt; be the PMF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, show that [math]p_X(k-1)p_X(k+1)\leq [p_X(k)]^2[/math] for &amp;lt;math&amp;gt;1\leq k \leq n-1&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Discrete random variable) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (I)&amp;lt;/strong&amp;gt;] Every package of some intrinsically dull commodity includes a small and exciting plastic object. There are [math]c[/math] different types of object, and each package is equally likely to contain any given type. You buy one package each day.&amp;lt;/p&amp;gt;&lt;br /&gt;
    &amp;lt;ol&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse between the acquisitions of the [math]j[/math]-th new type of object and the [math](j + 1)[/math]-th new type.&amp;lt;/li&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse before you have a full set of objects.&amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (II)&amp;lt;/strong&amp;gt;] Prove that geometric distribution is the only discrete memoryless distribution with range values &amp;lt;math&amp;gt;\mathbb{N}_+&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;n_1,n_2 \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;0 \le p \le 1&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Bin}(n_1,p),Y \sim \mathrm{Bin}(n_2,p)&amp;lt;/math&amp;gt; be independent random variables. Prove that &amp;lt;math&amp;gt;X+Y \sim \mathrm{Bin}(n_1+n_2,p)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Negative binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; follows the negative binomial distribution with parameter &amp;lt;math&amp;gt;r \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt;. Calculate &amp;lt;math&amp;gt;\mathbf{Var}[X] = \mathbf{E}[X^2] - \left(\mathbf{E}[X]\right)^2&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Hypergeometric distribution&amp;lt;/strong&amp;gt;] An urn contains [math]N[/math] balls, [math]b[/math] of which are blue and [math]r = N -b[/math] of which are red. A random sample of [math]n[/math] balls is drawn &amp;lt;strong&amp;gt;without replacement&amp;lt;/strong&amp;gt; (无放回) from the urn. Let [math]B[/math] the number of blue balls in this sample. Show that if [math]N, b[/math], and [math]r[/math] approach [math]+\infty[/math] in such a way that [math]b/N \rightarrow  p[/math] and [math]r/N \rightarrow 1 - p[/math], then [math]\mathbf{Pr}(B = k) \rightarrow {n\choose k}p^k(1-p)^{n-k}[/math] for [math]0\leq k \leq n[/math].&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Poisson distribution&amp;lt;/strong&amp;gt;] In your pocket is a random number &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; of coins, where &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; has the Poisson distribution with parameter &amp;lt;math&amp;gt;\lambda&amp;lt;/math&amp;gt;. You toss each coin once, with heads showing with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; each time. Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the (random) number of heads outcomes and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the (also random) number of tails.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the joint mass function of &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find PMF of the marginal distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; in &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;. Are &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; independent?&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Conditional distribution (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be independent &amp;lt;math&amp;gt;\text{Bin}(n, p)&amp;lt;/math&amp;gt; random variables, and let &amp;lt;math&amp;gt;Z = X + Y&amp;lt;/math&amp;gt;. Show that the conditional distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; given &amp;lt;math&amp;gt;Z = N&amp;lt;/math&amp;gt; is the hypergeometric distribution.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Conditional distribution (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;\lambda,\mu &amp;gt; 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;n \in \mathbb{N}&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Pois}(\lambda), Y \sim \mathrm{Pois}(\mu)&amp;lt;/math&amp;gt; be independent random variables. Find out the conditional distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, given &amp;lt;math&amp;gt;X+Y = n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 4 (Linearity of Expectation, 12 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Streak&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Suppose we flip a fair coin &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; times independently to obtain a sequence of flips &amp;lt;math&amp;gt;X_1, X_2, \ldots , X_n&amp;lt;/math&amp;gt;. A streak of flips is a consecutive subsequence of flips that are all the same. For example, if &amp;lt;math&amp;gt;X_3&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;X_4&amp;lt;/math&amp;gt;, and &amp;lt;math&amp;gt;X_5&amp;lt;/math&amp;gt; are all heads, there is a streak of length &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; starting at the third lip. (If &amp;lt;math&amp;gt;X_6&amp;lt;/math&amp;gt; is also heads, then there is also a streak of length &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; starting at the third lip.) Find the expected number of streaks of length &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; for some integer &amp;lt;math&amp;gt;k \ge 1&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Number of cycles&amp;lt;/strong&amp;gt;] &lt;br /&gt;
At a banquet, there are &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people who shake hands according to the following process: In each round, two idle hands are randomly selected and shaken (&amp;lt;strong&amp;gt;these two hands are no longer idle&amp;lt;/strong&amp;gt;). After &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds, there will be no idle hands left, and the &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people will form several cycles. For example, when &amp;lt;math&amp;gt;n=3&amp;lt;/math&amp;gt;, the following situation may occur: the left and right hands of the first person are held together, the left hand of the second person and the right hand of the third person are held together, and the right hand of the second person and the left hand of the third person are held together. In this case, three people form two cycles. How many cycles are expected to be formed after &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds?&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Expected Mex&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X_1,X_2,\ldots,X_{100} \sim \mathrm{Geo}(1/2)&amp;lt;/math&amp;gt; be independent random variables. Compute &amp;lt;math&amp;gt;\mathbf{E}[\mathrm{mex}(X_1,X_2,\ldots,X_{100})]&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\mathrm{mex}(a_1,a_2,\ldots,a_n)&amp;lt;/math&amp;gt; is the smallest positive integer that does not appear in &amp;lt;math&amp;gt;a_1,a_2,\ldots,a_n&amp;lt;/math&amp;gt;. Your answer is considered correct if the absolute error does not exceed &amp;lt;math&amp;gt;10^{-6}&amp;lt;/math&amp;gt;. (Hint: Although mathematics will help you arrive at elegant and efficient methods, the use of a computer and programming skills will be required.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 5 (Probability meets graph theory) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random social networks&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G = (V, E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph without isolating vertex. &lt;br /&gt;
Let &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; be the degree of vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Let &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be a uniformly chosen vertex, and &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; a uniformly chosen neighbor of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Show that &amp;lt;math&amp;gt;\mathbf{E}[d_Z] \geq \mathbf{E}[d_Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Interpret this inequality in the context of social networks, in which the vertices represent people, and the edges represent friendship.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Turán&#039;s Theorem&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G=(V,E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph, and write &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; for the degree of the vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Use probablistic method to prove that &amp;lt;math&amp;gt;\alpha(G) \ge \sum_{v \in V} \frac{1}{d_v + 1}&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\alpha(G)&amp;lt;/math&amp;gt; is the size of a maximum independent set. (Hint: Consider the following random procedure for generating an independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; from a graph with vertex set &amp;lt;math&amp;gt;V&amp;lt;/math&amp;gt;: First, generate a random permutation of the vertices, denoted as &amp;lt;math&amp;gt;v_1,v_2,\ldots,v_n&amp;lt;/math&amp;gt;. Then, construct the independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; as follows: For each vertex &amp;lt;math&amp;gt;v_i \in V&amp;lt;/math&amp;gt;, add &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; if and only if none of its predecessors in the permutation, i.e., &amp;lt;math&amp;gt;v_1,\ldots,v_{i-1}&amp;lt;/math&amp;gt;, are neighbors of &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Dominating set&amp;lt;/strong&amp;gt;]&lt;br /&gt;
A &#039;&#039;dominating set&#039;&#039; of vertices in an undirected graph &amp;lt;math&amp;gt;G = (V, E)&amp;lt;/math&amp;gt; is a set &amp;lt;math&amp;gt;S \subseteq V&amp;lt;/math&amp;gt; such that every vertex of&lt;br /&gt;
&amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; belongs to &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; or has a neighbor in &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;. &lt;br /&gt;
&lt;br /&gt;
Let &amp;lt;math&amp;gt;G = (V, E)&amp;lt;/math&amp;gt; be an &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;-vertex graph with minimum degree &amp;lt;math&amp;gt;d &amp;gt; 1&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; has a dominating set with at most &amp;lt;math&amp;gt;\frac{n\left(1+\log(d+1)\right)}{d+1}&amp;lt;/math&amp;gt; vertices. (Hint: Consider a random vertex subset &amp;lt;math&amp;gt;S \subseteq V&amp;lt;/math&amp;gt; by including each vertex independently with&lt;br /&gt;
probability &amp;lt;math&amp;gt;p := \log(d + 1)/(d + 1)&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 6 (1D random walk, 8 points) ==&lt;br /&gt;
Let &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt; be a constant, and &amp;lt;math&amp;gt;\{X_n\}_{n \ge 1}&amp;lt;/math&amp;gt; be independent Bernoulli trials with successful probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;.&lt;br /&gt;
Define &amp;lt;math&amp;gt;S_n = 2\sum_{i=1}^n X_i - n&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;S_0 = 0&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Range of random walk&amp;lt;/strong&amp;gt;] The range &amp;lt;math&amp;gt;R_n&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;S_0, S_1, \ldots, S_n&amp;lt;/math&amp;gt; is defined as the number of distinct values taken by the sequence. Show that &amp;lt;math&amp;gt;\mathbf{Pr}\left(R_n = R_{n-1}+1\right) = \mathbf{Pr}\left(\forall 1 \le i \le n, S_i \neq 0\right)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, and deduce that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n]\to \mathbf{Pr}(\forall i \ge 1, S_i \neq 0)&amp;lt;/math&amp;gt;. Hence show that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n] \to |2p-1|&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Symmetric 1D random walk (III)&amp;lt;/strong&amp;gt;] Suppose &amp;lt;math&amp;gt;p = \frac{1}{2}&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{E}[|S_n|] = \Theta(\sqrt{n})&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12946</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 2</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12946"/>
		<updated>2025-03-13T14:55:56Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: Created page with &amp;quot;*每道题目的解答都要有完整的解题过程，中英文不限。  *我们推荐大家使用LaTeX, markdown等对作业进行排版。  == Assumption throughout Problem Set 2== &amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;  &amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;  &amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural l...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 2==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Warm-up problems, 12 points) ==&lt;br /&gt;
* [&#039;&#039;&#039;Function of random variable (I)&#039;&#039;&#039;] Show that, if &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; are random variables, then so are &amp;lt;math&amp;gt;X+Y&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;XY&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\min\{X,Y\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
* [&#039;&#039;&#039;Function of random variable (II)&#039;&#039;&#039;] Let [math]X[/math] be a random variable with distribution function [math]\max(0,\min(1,x))[/math]. Let [math]F[/math] be a distribution function which is continuous and strictly increasing. Show that [math]Y=F^{-1}(X)[/math] be a random variable with distribution function [math]F[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Independence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;1\leq r\leq n&amp;lt;/math&amp;gt; be independent random variables which are symmetric about &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;; that is, &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-X_r&amp;lt;/math&amp;gt; have the same distributions. Show that, for all &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}[S_n \geq x] = \mathbf{Pr}[S_n \leq -x]&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;S_n = \sum_{r=1}^n X_r&amp;lt;/math&amp;gt;. Is the conclusion true without the assumtion of independence?&lt;br /&gt;
* [&#039;&#039;&#039;Dependence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be discrete random variables with joint mass function &amp;lt;math&amp;gt;f(x,y) = \frac{C}{(x+y-1)(x+y)(x+y+1)}&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;x,y \in \mathbb{N}_+&amp;lt;/math&amp;gt; (in other words, &amp;lt;math&amp;gt;x,y = 1,2,3,\cdots&amp;lt;/math&amp;gt;). Find (1) the value of &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;, (2) marginal mass function of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and (3) &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt;.&lt;br /&gt;
* [&#039;&#039;&#039;Expectation&#039;&#039;&#039;] Is it generally true that [math]\mathbf{E}[1/X] = 1/\mathbf{E}[X][/math]? Is it ever true that [math]\mathbf{E}[1/X] = 1/\mathbf{E}[X][/math]?&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Entropy of discrete random variable&amp;lt;/strong&amp;gt;] Let [math]X[/math] be a discrete random variable with range of values [math][N] = \{1,2,\ldots,N\}[/math] and probability mass function [math]p[/math]. Define [math]H(X) = -\sum_{n \ge 1} p(n) \log p(n)[/math] with convention [math]0\log 0 = 0[/math]. Prove that [math]H(X) \le \log N[/math] using Jensen&#039;s inequality.&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Law of total expectation&amp;lt;/strong&amp;gt;] Let [math]X \sim \mathrm{Geom}(p)[/math] for some parameter [math]p \in (0,1)[/math]. Calculate [math]\mathbf{E}[X][/math] using the law of total expectation.&lt;br /&gt;
&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Random number of random variables&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\{X_n\}_{n \ge 1}&amp;lt;/math&amp;gt; be identically distributed random variable and &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; be a random variable taking values in the non-negative integers and independent of the &amp;lt;math&amp;gt;X_n&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;n \ge 1&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{E}\left[\sum_{i=1}^N X_i\right] = \mathbf{E}[N] \mathbf{E}[X_1]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Distribution of random variable, 8 points) ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Cumulative distribution function (CDF)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a random variable with cumulative distribution function &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Show that &amp;lt;math&amp;gt;Y = aX+b&amp;lt;/math&amp;gt; is a random variable where &amp;lt;math&amp;gt;a&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;b&amp;lt;/math&amp;gt; are real constants, and express the CDF of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; by &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;. (Hint: Try expressing the event [math]Y=aX+b\le y[/math] by countably many set operations on the events defined on [math]X[/math].)&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Let &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; be the CDF of random variable &amp;lt;math&amp;gt;Z:\Omega\rightarrow \mathbb{R}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;0\leq \lambda \leq 1&amp;lt;/math&amp;gt;, show that &lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;math&amp;gt;\lambda F + (1-\lambda)G&amp;lt;/math&amp;gt; is a CDF function. &amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;The product &amp;lt;math&amp;gt;FG&amp;lt;/math&amp;gt; is a CDF function, and if &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; are independent, then &amp;lt;math&amp;gt;FG&amp;lt;/math&amp;gt; is the CDF of &amp;lt;math&amp;gt;\max\{X,Z\}&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Probability mass function (PMF)&amp;lt;/strong&amp;gt;] We toss &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; coins, and each one shows heads with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;, independently of each of the others. Each coin which shows head is tossed again. (If the coin shows tail, it won&#039;t be tossed again.) Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of heads resulting from the &amp;lt;strong&amp;gt;second&amp;lt;/strong&amp;gt; round of tosses, and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the number of heads resulting from &amp;lt;strong&amp;gt;all&amp;lt;/strong&amp;gt; tosses, which includes the first and (possible) second round of each toss.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the PMF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{E}[Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Let &amp;lt;math&amp;gt;p_X&amp;lt;/math&amp;gt; be the PMF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, show that [math]p_X(k-1)p_X(k+1)\leq [p_X(k)]^2[/math] for &amp;lt;math&amp;gt;1\leq k \leq n-1&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Discrete random variable) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (I)&amp;lt;/strong&amp;gt;] Every package of some intrinsically dull commodity includes a small and exciting plastic object. There are [math]c[/math] different types of object, and each package is equally likely to contain any given type. You buy one package each day.&amp;lt;/p&amp;gt;&lt;br /&gt;
    &amp;lt;ol&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse between the acquisitions of the [math]j[/math]-th new type of object and the [math](j + 1)[/math]-th new type.&amp;lt;/li&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse before you have a full set of objects.&amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (II)&amp;lt;/strong&amp;gt;] Prove that geometric distribution is the only discrete memoryless distribution with range values &amp;lt;math&amp;gt;\mathbb{N}_+&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;n_1,n_2 \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;0 \le p \le 1&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Bin}(n_1,p),Y \sim \mathrm{Bin}(n_2,p)&amp;lt;/math&amp;gt; be independent random variables. Prove that &amp;lt;math&amp;gt;X+Y \sim \mathrm{Bin}(n_1+n_2,p)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Negative binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; follows the negative binomial distribution with parameter &amp;lt;math&amp;gt;r \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt;. Calculate &amp;lt;math&amp;gt;\mathbf{Var}[X] = \mathbf{E}[X^2] - \left(\mathbf{E}[X]\right)^2&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Hypergeometric distribution&amp;lt;/strong&amp;gt;] An urn contains [math]N[/math] balls, [math]b[/math] of which are blue and [math]r = N -b[/math] of which are red. A random sample of [math]n[/math] balls is drawn &amp;lt;strong&amp;gt;without replacement&amp;lt;/strong&amp;gt; (无放回) from the urn. Let [math]B[/math] the number of blue balls in this sample. Show that if [math]N, b[/math], and [math]r[/math] approach [math]+\infty[/math] in such a way that [math]b/N \rightarrow  p[/math] and [math]r/N \rightarrow 1 - p[/math], then [math]\mathbf{Pr}(B = k) \rightarrow {n\choose k}p^k(1-p)^{n-k}[/math] for [math]0\leq k \leq n[/math].&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Poisson distribution&amp;lt;/strong&amp;gt;] In your pocket is a random number &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; of coins, where &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; has the Poisson distribution with parameter &amp;lt;math&amp;gt;\lambda&amp;lt;/math&amp;gt;. You toss each coin once, with heads showing with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; each time. Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the (random) number of heads outcomes and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the (also random) number of tails.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the joint mass function of &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find PMF of the marginal distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; in &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;. Are &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; independent?&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Conditional distribution (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be independent &amp;lt;math&amp;gt;\text{Bin}(n, p)&amp;lt;/math&amp;gt; random variables, and let &amp;lt;math&amp;gt;Z = X + Y&amp;lt;/math&amp;gt;. Show that the conditional distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; given &amp;lt;math&amp;gt;Z = N&amp;lt;/math&amp;gt; is the hypergeometric distribution.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Conditional distribution (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;\lambda,\mu &amp;gt; 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;n \in \mathbb{N}&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Pois}(\lambda), Y \sim \mathrm{Pois}(\mu)&amp;lt;/math&amp;gt; be independent random variables. Find out the conditional distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, given &amp;lt;math&amp;gt;X+Y = n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 4 (Linearity of Expectation, 12 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Streak&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Suppose we flip a fair coin &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; times independently to obtain a sequence of flips &amp;lt;math&amp;gt;X_1, X_2, \ldots , X_n&amp;lt;/math&amp;gt;. A streak of flips is a consecutive subsequence of flips that are all the same. For example, if &amp;lt;math&amp;gt;X_3&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;X_4&amp;lt;/math&amp;gt;, and &amp;lt;math&amp;gt;X_5&amp;lt;/math&amp;gt; are all heads, there is a streak of length &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; starting at the third lip. (If &amp;lt;math&amp;gt;X_6&amp;lt;/math&amp;gt; is also heads, then there is also a streak of length &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; starting at the third lip.) Find the expected number of streaks of length &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; for some integer &amp;lt;math&amp;gt;k \ge 1&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Number of cycles&amp;lt;/strong&amp;gt;] &lt;br /&gt;
At a banquet, there are &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people who shake hands according to the following process: In each round, two idle hands are randomly selected and shaken (&amp;lt;strong&amp;gt;these two hands are no longer idle&amp;lt;/strong&amp;gt;). After &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds, there will be no idle hands left, and the &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people will form several cycles. For example, when &amp;lt;math&amp;gt;n=3&amp;lt;/math&amp;gt;, the following situation may occur: the left and right hands of the first person are held together, the left hand of the second person and the right hand of the third person are held together, and the right hand of the second person and the left hand of the third person are held together. In this case, three people form two cycles. How many cycles are expected to be formed after &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds?&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Paper Cutting&amp;lt;/strong&amp;gt;]&lt;br /&gt;
We have a rectangular piece of paper divided into &amp;lt;math&amp;gt;H \times W&amp;lt;/math&amp;gt; squares, where two of those squares are painted black and the rest are painted white. If we let &amp;lt;math&amp;gt;(i,j)&amp;lt;/math&amp;gt; denote the square at the &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt;-th row and &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt;-th column, the squares painted black are &amp;lt;math&amp;gt;(h_1,w_1)&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;(h_2,w_2)&amp;lt;/math&amp;gt;. Bob will repeat the following operation to cut the piece of paper: &lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
Assume that we have &lt;br /&gt;
&amp;lt;math&amp;gt;h \times w&amp;lt;/math&amp;gt; squares remaining. There are &lt;br /&gt;
&amp;lt;math&amp;gt;(h−1)&amp;lt;/math&amp;gt; horizontal lines and &lt;br /&gt;
&amp;lt;math&amp;gt;(w−1)&amp;lt;/math&amp;gt; vertical lines that are parallel to the edges of the piece and pass the borders of the squares. He chooses one of these lines uniformly at random and cuts the piece into two along that line. Then, if the two black squares are on the same piece, he throws away the other piece and continues the process; otherwise, he ends the process. &lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Find the expected value of the number of times Bob cuts a piece of paper until he ends the process.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Expected Mex&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X_1,X_2,\ldots,X_{100} \sim \mathrm{Geo}(1/2)&amp;lt;/math&amp;gt; be independent random variables. Compute &amp;lt;math&amp;gt;\mathbf{E}[\mathrm{mex}(X_1,X_2,\ldots,X_{100})]&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\mathrm{mex}(a_1,a_2,\ldots,a_n)&amp;lt;/math&amp;gt; is the smallest positive integer that does not appear in &amp;lt;math&amp;gt;a_1,a_2,\ldots,a_n&amp;lt;/math&amp;gt;. Your answer is considered correct if the absolute error does not exceed &amp;lt;math&amp;gt;10^{-6}&amp;lt;/math&amp;gt;. (Hint: Although mathematics will help you arrive at elegant and efficient methods, the use of a computer and programming skills will be required.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 5 (Probability meets graph theory) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random social networks&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G = (V, E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph without isolating vertex. &lt;br /&gt;
Let &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; be the degree of vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Let &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be a uniformly chosen vertex, and &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; a uniformly chosen neighbor of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Show that &amp;lt;math&amp;gt;\mathbf{E}[d_Z] \geq \mathbf{E}[d_Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Interpret this inequality in the context of social networks, in which the vertices represent people, and the edges represent friendship.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Turán&#039;s Theorem&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G=(V,E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph, and write &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; for the degree of the vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Use probablistic method to prove that &amp;lt;math&amp;gt;\alpha(G) \ge \sum_{v \in V} \frac{1}{d_v + 1}&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\alpha(G)&amp;lt;/math&amp;gt; is the size of a maximum independent set. (Hint: Consider the following random procedure for generating an independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; from a graph with vertex set &amp;lt;math&amp;gt;V&amp;lt;/math&amp;gt;: First, generate a random permutation of the vertices, denoted as &amp;lt;math&amp;gt;v_1,v_2,\ldots,v_n&amp;lt;/math&amp;gt;. Then, construct the independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; as follows: For each vertex &amp;lt;math&amp;gt;v_i \in V&amp;lt;/math&amp;gt;, add &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; if and only if none of its predecessors in the permutation, i.e., &amp;lt;math&amp;gt;v_1,\ldots,v_{i-1}&amp;lt;/math&amp;gt;, are neighbors of &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Dominating set&amp;lt;/strong&amp;gt;]&lt;br /&gt;
A &#039;&#039;dominating set&#039;&#039; of vertices in an undirected graph &amp;lt;math&amp;gt;G = (V, E)&amp;lt;/math&amp;gt; is a set &amp;lt;math&amp;gt;S \subseteq V&amp;lt;/math&amp;gt; such that every vertex of&lt;br /&gt;
&amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; belongs to &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; or has a neighbor in &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;. &lt;br /&gt;
&lt;br /&gt;
Let &amp;lt;math&amp;gt;G = (V, E)&amp;lt;/math&amp;gt; be an &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;-vertex graph with minimum degree &amp;lt;math&amp;gt;d &amp;gt; 1&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; has a dominating set with at most &amp;lt;math&amp;gt;\frac{n\left(1+\log(d+1)\right)}{d+1}&amp;lt;/math&amp;gt; vertices. (Hint: Consider a random vertex subset &amp;lt;math&amp;gt;S \subseteq V&amp;lt;/math&amp;gt; by including each vertex independently with&lt;br /&gt;
probability &amp;lt;math&amp;gt;p := \log(d + 1)/(d + 1)&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 6 (1D random walk, 8 points) ==&lt;br /&gt;
Let &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt; be a constant, and &amp;lt;math&amp;gt;\{X_n\}_{n \ge 1}&amp;lt;/math&amp;gt; be independent Bernoulli trials with successful probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;.&lt;br /&gt;
Define &amp;lt;math&amp;gt;S_n = 2\sum_{i=1}^n X_i - n&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;S_0 = 0&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Range of random walk&amp;lt;/strong&amp;gt;] The range &amp;lt;math&amp;gt;R_n&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;S_0, S_1, \ldots, S_n&amp;lt;/math&amp;gt; is defined as the number of distinct values taken by the sequence. Show that &amp;lt;math&amp;gt;\mathbf{Pr}\left(R_n = R_{n-1}+1\right) = \mathbf{Pr}\left(\forall 1 \le i \le n, S_i \neq 0\right)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, and deduce that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n]\to \mathbf{Pr}(\forall i \ge 1, S_i \neq 0)&amp;lt;/math&amp;gt;. Hence show that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n] \to |2p-1|&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Symmetric 1D random walk (III)&amp;lt;/strong&amp;gt;] Suppose &amp;lt;math&amp;gt;p = \frac{1}{2}&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{E}[|S_n|] = \Theta(\sqrt{n})&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)&amp;diff=12945</id>
		<title>概率论与数理统计 (Spring 2025)</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)&amp;diff=12945"/>
		<updated>2025-03-13T14:47:27Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: /* Assignments */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Infobox&lt;br /&gt;
|name         = Infobox&lt;br /&gt;
|bodystyle    = &lt;br /&gt;
|title        = &amp;lt;font size=3&amp;gt;&#039;&#039;&#039;概率论与数理统计&#039;&#039;&#039;&amp;lt;br&amp;gt;&lt;br /&gt;
&#039;&#039;&#039;Probability Theory&#039;&#039;&#039; &amp;lt;br&amp;gt; &amp;amp; &#039;&#039;&#039;Mathematical Statistics&#039;&#039;&#039;&amp;lt;/font&amp;gt;&lt;br /&gt;
|titlestyle   = &lt;br /&gt;
&lt;br /&gt;
|image        = &lt;br /&gt;
|imagestyle   = &lt;br /&gt;
|caption      = &lt;br /&gt;
|captionstyle = &lt;br /&gt;
|headerstyle  = background:#ccf;&lt;br /&gt;
|labelstyle   = background:#ddf;&lt;br /&gt;
|datastyle    = &lt;br /&gt;
&lt;br /&gt;
|header1 =Instructor&lt;br /&gt;
|label1  = &lt;br /&gt;
|data1   = &lt;br /&gt;
|header2 = &lt;br /&gt;
|label2  = &lt;br /&gt;
|data2   = &#039;&#039;&#039;尹一通&#039;&#039;&#039;&lt;br /&gt;
|header3 = &lt;br /&gt;
|label3  = Email&lt;br /&gt;
|data3   = yinyt@nju.edu.cn  &lt;br /&gt;
|header4 =&lt;br /&gt;
|label4  = office&lt;br /&gt;
|data4   = 计算机系 804&lt;br /&gt;
|header5 = &lt;br /&gt;
|label5  = &lt;br /&gt;
|data5   = &#039;&#039;&#039;刘景铖&#039;&#039;&#039;&lt;br /&gt;
|header6 = &lt;br /&gt;
|label6  = Email&lt;br /&gt;
|data6   = liu@nju.edu.cn  &lt;br /&gt;
|header7 =&lt;br /&gt;
|label7  = office&lt;br /&gt;
|data7   = 计算机系 516&lt;br /&gt;
|header8 = Class&lt;br /&gt;
|label8  = &lt;br /&gt;
|data8   = &lt;br /&gt;
|header9 =&lt;br /&gt;
|label9  = Class meeting&lt;br /&gt;
|data9   = Monday, 10am-12pm&amp;lt;br&amp;gt;&lt;br /&gt;
Wednesday (单), 10am-12pm&amp;lt;br&amp;gt;&lt;br /&gt;
仙Ⅰ-204&lt;br /&gt;
|header10=&lt;br /&gt;
|label10 = Office hour&lt;br /&gt;
|data10  = Friday, 2pm-3pm &amp;lt;br&amp;gt;计算机系 804（尹一通）&amp;lt;br&amp;gt;计算机系 516（刘景铖）&lt;br /&gt;
|header11= Textbook&lt;br /&gt;
|label11 = &lt;br /&gt;
|data11  = &lt;br /&gt;
|header12=&lt;br /&gt;
|label12 = &lt;br /&gt;
|data12  = [[File:概率导论.jpeg|border|100px]]&lt;br /&gt;
|header13=&lt;br /&gt;
|label13 = &lt;br /&gt;
|data13  = &#039;&#039;&#039;概率导论&#039;&#039;&#039;（第2版·修订版）&amp;lt;br&amp;gt; Dimitri P. Bertsekas and John N. Tsitsiklis&amp;lt;br&amp;gt; 郑忠国 童行伟 译；人民邮电出版社 (2022)&lt;br /&gt;
|header14=&lt;br /&gt;
|label14 = &lt;br /&gt;
|data14  = [[File:Grimmett_probability.jpg|border|100px]]&lt;br /&gt;
|header15=&lt;br /&gt;
|label15 = &lt;br /&gt;
|data15  = &#039;&#039;&#039;Probability and Random Processes&#039;&#039;&#039; (4E) &amp;lt;br&amp;gt; Geoffrey Grimmett and David Stirzaker &amp;lt;br&amp;gt;  Oxford University Press (2020)&lt;br /&gt;
|header16=&lt;br /&gt;
|label16 = &lt;br /&gt;
|data16  = [[File:Probability_and_Computing_2ed.jpg|border|100px]]&lt;br /&gt;
|header17=&lt;br /&gt;
|label17 = &lt;br /&gt;
|data17  = &#039;&#039;&#039;Probability and Computing&#039;&#039;&#039; (2E) &amp;lt;br&amp;gt; Michael Mitzenmacher and Eli Upfal &amp;lt;br&amp;gt;   Cambridge University Press (2017)&lt;br /&gt;
|belowstyle = background:#ddf;&lt;br /&gt;
|below = &lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
This is the webpage for the &#039;&#039;Probability Theory and Mathematical Statistics&#039;&#039; (概率论与数理统计) class of Spring 2025. Students who take this class should check this page periodically for content updates and new announcements. &lt;br /&gt;
&lt;br /&gt;
= Announcement =&lt;br /&gt;
* TBA&lt;br /&gt;
&lt;br /&gt;
= Course info =&lt;br /&gt;
* &#039;&#039;&#039;Instructor &#039;&#039;&#039;: &lt;br /&gt;
:* [http://tcs.nju.edu.cn/yinyt/ 尹一通]：[mailto:yinyt@nju.edu.cn &amp;lt;yinyt@nju.edu.cn&amp;gt;]，计算机系 804 &lt;br /&gt;
:* [https://liuexp.github.io 刘景铖]：[mailto:liu@nju.edu.cn &amp;lt;liu@nju.edu.cn&amp;gt;]，计算机系 516 &lt;br /&gt;
* &#039;&#039;&#039;Teaching assistant&#039;&#039;&#039;:&lt;br /&gt;
** [https://sites.google.com/view/xinyuanzhang 张昕渊]：[mailto:zhangxy@smail.nju.edu.cn &amp;lt;zhangxy@smail.nju.edu.cn&amp;gt;]，计算机系 410&lt;br /&gt;
** 邹宗瑞：[mailto:zou.zongrui@smail.nju.edu.cn &amp;lt;zou.zongrui@smail.nju.edu.cn&amp;gt;]，计算机系 410&lt;br /&gt;
* &#039;&#039;&#039;Class meeting&#039;&#039;&#039;:&lt;br /&gt;
** 周一：10am-12pm，仙Ⅰ-204&lt;br /&gt;
** 周三（单）：10am-12pm，仙Ⅰ-204&lt;br /&gt;
* &#039;&#039;&#039;Office hour&#039;&#039;&#039;: &lt;br /&gt;
:* 周五 2pm-3pm, 计算机系 804（尹一通）&lt;br /&gt;
:* 周五 2pm-3pm, 计算机系 516（刘景铖）&lt;br /&gt;
:* &#039;&#039;&#039;QQ群&#039;&#039;&#039;: 974742320（申请加入需提供姓名、院系、学号）&lt;br /&gt;
&lt;br /&gt;
= Syllabus =&lt;br /&gt;
课程内容分为三大部分：&lt;br /&gt;
* &#039;&#039;&#039;经典概率论&#039;&#039;&#039;：概率空间、随机变量及其数字特征、多维与连续随机变量、极限定理等内容&lt;br /&gt;
* &#039;&#039;&#039;概率与计算&#039;&#039;&#039;：测度集中现象 (concentration of measure)、概率法 (the probabilistic method)、离散随机过程的相关专题&lt;br /&gt;
* &#039;&#039;&#039;数理统计&#039;&#039;&#039;：参数估计、假设检验、贝叶斯估计、线性回归等统计推断等概念&lt;br /&gt;
&lt;br /&gt;
对于第一和第二部分，要求清楚掌握基本概念，深刻理解关键的现象与规律以及背后的原理，并可以灵活运用所学方法求解相关问题。对于第三部分，要求熟悉数理统计的若干基本概念，以及典型的统计模型和统计推断问题。&lt;br /&gt;
&lt;br /&gt;
经过本课程的训练，力求使学生能够熟悉掌握概率的语言，并会利用概率思维来理解客观世界并对其建模，以及驾驭概率的数学工具来分析和求解专业问题。&lt;br /&gt;
&lt;br /&gt;
=== 教材与参考书 Course Materials ===&lt;br /&gt;
* &#039;&#039;&#039;[BT]&#039;&#039;&#039; 概率导论（第2版·修订版），[美]伯特瑟卡斯（Dimitri P.Bertsekas）[美]齐齐克利斯（John N.Tsitsiklis）著，郑忠国 童行伟 译，人民邮电出版社（2022）。&lt;br /&gt;
* &#039;&#039;&#039;[GS]&#039;&#039;&#039; &#039;&#039;Probability and Random Processes&#039;&#039;, by Geoffrey Grimmett and David Stirzaker; Oxford University Press; 4th edition (2020).&lt;br /&gt;
* &#039;&#039;&#039;[MU]&#039;&#039;&#039; &#039;&#039;Probability and Computing: Randomization and Probabilistic Techniques in Algorithms and Data Analysis&#039;&#039;, by Michael Mitzenmacher, Eli Upfal; Cambridge University Press; 2nd edition (2017).&lt;br /&gt;
&lt;br /&gt;
=== 成绩 Grading Policy ===&lt;br /&gt;
* 课程成绩：本课程将会有若干次作业和一次期末考试。最终成绩将由平时作业成绩和期末考试成绩综合得出。&lt;br /&gt;
* 迟交：如果有特殊的理由，无法按时完成作业，请提前联系授课老师，给出正当理由。否则迟交的作业将不被接受。&lt;br /&gt;
&lt;br /&gt;
=== &amp;lt;font color=red&amp;gt; 学术诚信 Academic Integrity &amp;lt;/font&amp;gt;===&lt;br /&gt;
学术诚信是所有从事学术活动的学生和学者最基本的职业道德底线，本课程将不遗余力的维护学术诚信规范，违反这一底线的行为将不会被容忍。&lt;br /&gt;
&lt;br /&gt;
作业完成的原则：署你名字的工作必须是你个人的贡献。在完成作业的过程中，允许讨论，前提是讨论的所有参与者均处于同等完成度。但关键想法的执行、以及作业文本的写作必须独立完成，并在作业中致谢（acknowledge）所有参与讨论的人。符合规则的讨论与致谢将不会影响得分。不允许其他任何形式的合作——尤其是与已经完成作业的同学“讨论”。&lt;br /&gt;
&lt;br /&gt;
本课程将对剽窃行为采取零容忍的态度。在完成作业过程中，对他人工作（出版物、互联网资料、其他人的作业等）直接的文本抄袭和对关键思想、关键元素的抄袭，按照 [http://www.acm.org/publications/policies/plagiarism_policy ACM Policy on Plagiarism]的解释，都将视为剽窃。剽窃者成绩将被取消。如果发现互相抄袭行为，&amp;lt;font color=red&amp;gt; 抄袭和被抄袭双方的成绩都将被取消&amp;lt;/font&amp;gt;。因此请主动防止自己的作业被他人抄袭。&lt;br /&gt;
&lt;br /&gt;
学术诚信影响学生个人的品行，也关乎整个教育系统的正常运转。为了一点分数而做出学术不端的行为，不仅使自己沦为一个欺骗者，也使他人的诚实努力失去意义。让我们一起努力维护一个诚信的环境。&lt;br /&gt;
&lt;br /&gt;
= Assignments =&lt;br /&gt;
*[[概率论与数理统计 (Spring 2025)/Problem Set 1|Problem Set 1]]  请在 2025/3/10 上课之前(10am UTC+8)提交到 [mailto:pr_nju@163.com pr_nju@163.com] (文件名为&#039;&amp;lt;font color=red &amp;gt;学号_姓名_A1.pdf&amp;lt;/font&amp;gt;&#039;).&lt;br /&gt;
** [[概率论与数理统计 (Spring 2025)/第一次作业提交名单|第一次作业提交名单]]&lt;br /&gt;
&lt;br /&gt;
*[[概率论与数理统计 (Spring 2025)/Problem Set 2|Problem Set 2]]  请在 --- 上课之前(10am UTC+8)提交到 [mailto:pr_nju@163.com pr_nju@163.com] (文件名为&#039;&amp;lt;font color=red &amp;gt;学号_姓名_A2.pdf&amp;lt;/font&amp;gt;&#039;).&lt;br /&gt;
&lt;br /&gt;
= Lectures =&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2025/Intro.pdf 课程简介]&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2025/ProbSpace.pdf 概率空间]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 第1章&#039;&#039;&#039; 或 &#039;&#039;&#039;[GS] Chapter 1&#039;&#039;&#039;&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2025)/Entropy and volume of Hamming balls|Entropy and volume of Hamming balls]]&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2025/RandVar.pdf 随机变量]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 第2章&#039;&#039;&#039; 或 &#039;&#039;&#039;[GS] Chapter 2, Sections 3.1~3.5, 3.7&#039;&#039;&#039;&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[MU] Chapter 2&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= Concepts =&lt;br /&gt;
* [https://plato.stanford.edu/entries/probability-interpret/ Interpretations of probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/History_of_probability History of probability]&lt;br /&gt;
* Example problems:&lt;br /&gt;
** [https://dornsifecms.usc.edu/assets/sites/520/docs/VonNeumann-ams12p36-38.pdf von Neumann&#039;s Bernoulli factory] and other [https://peteroupc.github.io/bernoulli.html Bernoulli factory algorithms]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Boy_or_Girl_paradox Boy or Girl paradox]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Monty_Hall_problem Monty Hall problem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bertrand_paradox_(probability) Bertrand paradox]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hard_spheres Hard spheres model] and [https://en.wikipedia.org/wiki/Ising_model Ising model]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/PageRank &#039;&#039;PageRank&#039;&#039;] and stationary [https://en.wikipedia.org/wiki/Random_walk random walk]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Diffusion_process Diffusion process] and [https://en.wikipedia.org/wiki/Diffusion_model diffusion model]&lt;br /&gt;
*[https://en.wikipedia.org/wiki/Probability_space Probability space]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Sample_space Sample space]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Event_(probability_theory) Event] and [https://en.wikipedia.org/wiki/Σ-algebra &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-algebra]&lt;br /&gt;
** Kolmogorov&#039;s [https://en.wikipedia.org/wiki/Probability_axioms axioms of probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Discrete_uniform_distribution Classical] and [https://en.wikipedia.org/wiki/Geometric_probability goemetric probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Boole%27s_inequality Union bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Inclusion%E2%80%93exclusion_principle Inclusion-Exclusion principle]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Boole%27s_inequality#Bonferroni_inequalities Bonferroni inequalities]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Conditional_probability Conditional probability]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chain_rule_(probability) Chain rule]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_total_probability Law of total probability]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bayes%27_theorem Bayes&#039; law]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Independence_(probability_theory) Independence] &lt;br /&gt;
** [https://en.wikipedia.org/wiki/Pairwise_independence Pairwise independence]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Random_variable Random variable]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cumulative_distribution_function Cumulative distribution function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_mass_function Probability mass function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_density_function Probability density function]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Multivariate_random_variable Random vector]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Joint_probability_distribution Joint probability distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Conditional_probability_distribution Conditional probability distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Marginal_distribution Marginal distribution]&lt;br /&gt;
* Some &#039;&#039;&#039;discrete&#039;&#039;&#039; probability distributions&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bernoulli_trial Bernoulli trial] and [https://en.wikipedia.org/wiki/Bernoulli_distribution Bernoulli distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Discrete_uniform_distribution Discrete uniform distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Binomial_distribution Binomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Geometric_distribution Geometric distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Negative_binomial_distribution Negative binomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hypergeometric_distribution Hypergeometric distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Poisson_distribution Poisson distribution]&lt;br /&gt;
** and [https://en.wikipedia.org/wiki/List_of_probability_distributions#Discrete_distributions others]&lt;br /&gt;
* Balls into bins model&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Multinomial_distribution Multinomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Birthday_problem Birthday problem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Coupon_collector%27s_problem Coupon collector]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Balls_into_bins_problem Occupancy problem]&lt;br /&gt;
* Random graphs&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Erd%C5%91s%E2%80%93R%C3%A9nyi_model Erdős–Rényi random graph model]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Galton%E2%80%93Watson_process Galton–Watson branching process]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Expected_value Expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_the_unconscious_statistician Law of the unconscious statistician, &#039;&#039;LOTUS&#039;&#039;]&lt;br /&gt;
** [https://dlsun.github.io/probability/linearity.html Linearity of expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Conditional_expectation Conditional expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_total_expectation Law of total expectation]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Markov%27s_inequality Markov&#039;s inequality]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Chebyshev%27s_inequality Chebyshev&#039;s inequality]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Moment_(mathematics) Moment] related&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Central_moment Central moment]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Variance Variance]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Standard_deviation Standard deviation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Covariance Covariance]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Correlation Correlation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Correlation_does_not_imply_causation Correlation does not imply causation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Skewness Skewness]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Kurtosis Kurtosis]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Moment_problem Moment problem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Measure_(mathematics) Measure]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Borel_set Borel set]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Lebesgue_integration Lebesgue integration]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Measurable_function Measurable function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cantor_set Cantor set]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Non-measurable_set Non-measurable set]&lt;br /&gt;
*[https://en.wikipedia.org/wiki/Probability_density_function Probability density function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_distribution#Absolutely_continuous_probability_distribution Continuous probability distribution]&lt;br /&gt;
**[https://en.wikipedia.org/wiki/Conditional_probability_distribution#Conditional_continuous_distributions Conditional continuous distributions]&lt;br /&gt;
**[https://en.wikipedia.org/wiki/Convolution_of_probability_distributions Convolution of probability distributions] and [https://en.wikipedia.org/wiki/List_of_convolutions_of_probability_distributions List of convolutions of probability distributions]&lt;br /&gt;
* Some &#039;&#039;&#039;continuous&#039;&#039;&#039; probability distributions&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Continuous_uniform_distribution Continuous uniform distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Exponential_distribution Exponential distribution] and [https://en.wikipedia.org/wiki/Poisson_point_process Poisson point process]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Normal_distribution Normal (Gaussion) distribution]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Gaussian_function Gaussian function] and [https://en.wikipedia.org/wiki/Gaussian_integral Gaussian integral]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Standard_normal_table Standard normal table]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/68%E2%80%9395%E2%80%9399.7_rule 68–95–99.7 rule]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Sum_of_normally_distributed_random_variables Sum of normally distributed random variables]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Multivariate_normal_distribution Multivariate normal distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chi-squared_distribution Chi-squared distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Gamma_distribution Gamma distribution] and [https://en.wikipedia.org/wiki/Gamma_function Gamma function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Beta_distribution Beta distribution] and [https://en.wikipedia.org/wiki/Beta_function Beta function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cauchy_distribution Cauchy distribution]&lt;br /&gt;
** and [https://en.wikipedia.org/wiki/List_of_probability_distributions#Absolutely_continuous_distributions others]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Inverse_transform_sampling Inverse transform sampling]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stochastic_dominance Stochastic dominance] and [https://en.wikipedia.org/wiki/Coupling_(probability) Coupling] &lt;br /&gt;
* [https://en.wikipedia.org/wiki/Moment-generating_function Moment-generating function]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Convergence_of_random_variables Convergence of random variables]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Pointwise_convergence Pointwise convergence]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Convergence_of_measures#Weak_convergence_of_measures Weak convergence of measures]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Convergence_in_measure Convergence in measure]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Skorokhod%27s_representation_theorem Skorokhod&#039;s representation theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Continuous_mapping_theorem Continuous mapping theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Borel%E2%80%93Cantelli_lemma Borel–Cantelli lemma] and [https://en.wikipedia.org/wiki/Zero%E2%80%93one_law zero-one laws]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Law_of_large_numbers Law of large numbers]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Central_limit_theorem Central limit theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/De_Moivre%E2%80%93Laplace_theorem De Moivre–Laplace theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Berry%E2%80%93Esseen_theorem Berry–Esseen theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Characteristic_function_(probability_theory) Characteristic function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Lévy%27s_continuity_theorem Lévy&#039;s continuity theorem]&lt;br /&gt;
* Concentration of measures:&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Moment-generating_function Moment-generating function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chernoff_bound Chernoff bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hoeffding%27s_inequality Hoeffding&#039;s bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Doob_martingale#McDiarmid&#039;s_inequality McDiarmid&#039;s inequality]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Doob_martingale Doob martingale] and [https://en.wikipedia.org/wiki/Azuma%27s_inequality Azuma&#039;s inequality]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Sub-Gaussian_distribution Sub-Gaussian distribution]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stochastic_process Stochastic process]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stopping_time Stopping time]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Martingale_(probability_theory) Martingale]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Optional_stopping_theorem Optional stopping theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Wald%27s_equation Wald&#039;s equation]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Markov_chain Markov chain]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Markov_property Markov property] and [https://en.wikipedia.org/wiki/Memorylessness memorylessness]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Stochastic_matrix Stochastic matrix]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Stationary_distribution Stationary distribution]&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_1&amp;diff=12910</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 1</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_1&amp;diff=12910"/>
		<updated>2025-02-20T06:49:25Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 1==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-2-principle-of-inclusion-and-exclusion&amp;quot;&amp;gt;Problem 1 (Principle of Inclusion and Exclusion, 8 points)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Let &amp;lt;math&amp;gt;n\ge 1&amp;lt;/math&amp;gt; be a positive integer and &amp;lt;math&amp;gt;A_1,A_2,\ldots,A_n&amp;lt;/math&amp;gt; be &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; events.&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Union bound&amp;lt;/strong&amp;gt;] Prove &amp;lt;math&amp;gt;\mathbf{Pr}\left( \bigcup_{i=1}^{n} A_i \right) \le \sum_{i=1}^{n} \mathbf{Pr}\left(A_i\right)&amp;lt;/math&amp;gt;&amp;lt;strong&amp;gt; using the definition of probability space&amp;lt;/strong&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Principle of Inclusion and Exclusion (PIE)&amp;lt;/strong&amp;gt;] Prove that &amp;lt;math&amp;gt;\mathbf{Pr}\left( \bigcup_{i=1}^n A_i\right) = \sum_{\emptyset \neq S \subseteq [n]} (-1)^{|S|-1} \mathbf{Pr}\left( \bigcap_{i \in S} A_i \right)&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;[n]=\{1,2,\ldots,n\}&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Surjection&amp;lt;/strong&amp;gt;] For positive integers &amp;lt;math&amp;gt;m\ge n&amp;lt;/math&amp;gt;, prove that the probability of a uniform random function &amp;lt;math&amp;gt;f:[m]\to[n]&amp;lt;/math&amp;gt; to be surjective (满射) is &amp;lt;math&amp;gt;\sum_{k=1}^n(-1)^{n-k}{n\choose k}\left(\frac{k}{n}\right)^m&amp;lt;/math&amp;gt;. &amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Bonferroni&amp;amp;#39;s inequality and Kounias&amp;amp;#39; inequality&amp;lt;/strong&amp;gt;] Prove that &lt;br /&gt;
&amp;lt;math&amp;gt;&lt;br /&gt;
\sum_{i=1}^n \mathbf{Pr}(A_i) - \sum_{1 \le i&amp;lt; j \le n} \mathbf{Pr}(A_i \cap A_j)\le \mathbf{Pr}\left(\bigcup_{i=1}^n A_i\right) \le \sum_{i=1}^n \mathbf{Pr} \left( A_i\right) - \sum_{i=2}^n \mathbf{Pr}(A_1 \cap A_i).&lt;br /&gt;
&amp;lt;/math&amp;gt; &lt;br /&gt;
(Hint: This is sometimes called Kounias&#039; inequality which is weaker than the Bonferroni&#039;s inequality. You can try using Venn diagram to understand these inequalities.)&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-2-probability-space&amp;quot;&amp;gt;Problem 2 (Probability space, 12 points)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Nonexistence of probability space&amp;lt;/strong&amp;gt;] Prove that it is impossible to define a uniform probability law on natural numbers &amp;lt;math&amp;gt;\mathbb{N}&amp;lt;/math&amp;gt;. More precisely, prove that there does not exist a probability space &amp;lt;math&amp;gt;(\mathbb{N},2^{\mathbb{N}},\mathbf{Pr})&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;\mathbf{Pr}(\{i\}) = \mathbf{Pr}(\{j\})&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;i, j \in \mathbb{N}&amp;lt;/math&amp;gt;. &lt;br /&gt;
Please explain why the same argument fails to prove that there is no uniform probability law on the real interval &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt;, that is, there is no such probability space &amp;lt;math&amp;gt;([0,1],\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt; that for any interval &amp;lt;math&amp;gt;(l,r] \subseteq [0,1]&amp;lt;/math&amp;gt;, it holds that &amp;lt;math&amp;gt;(l,r] \in \mathcal{F}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{Pr}( (l,r] ) = r-l&amp;lt;/math&amp;gt;. (Actually, such probability measure does exist and is called the Lebesgue measure on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt;). &amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Smallest &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field (I)&amp;lt;/strong&amp;gt;] For any subset &amp;lt;math&amp;gt;S \subseteq 2^\Omega&amp;lt;/math&amp;gt;, prove that the smallest &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field containing &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; is given by &amp;lt;math&amp;gt;\sigma(S) := \bigcap_{\substack{S \subseteq \mathcal{F} \subseteq 2^\Omega\\ \mathcal{F} \text{ is a } \sigma\text{-field }} } \mathcal{F}&amp;lt;/math&amp;gt;. (Hint: You should show that it is indeed a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field and also it is the smallest one containing &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;.)&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Smallest &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field (II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;S,T \subseteq 2^{\Omega}&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\sigma(S) = \sigma(T)&amp;lt;/math&amp;gt; if and only if &amp;lt;math&amp;gt;S \subseteq \sigma(T)&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;T \subseteq \sigma(S)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Union of &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\mathcal{F}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathcal{G}&amp;lt;/math&amp;gt; be &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-fields of subsets of &amp;lt;math&amp;gt;\Omega&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathcal{F} \cup \mathcal{G}&amp;lt;/math&amp;gt; is not necessarily a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field. Suppose &amp;lt;math&amp;gt;\mathcal{F}_1 \subseteq \mathcal{F}_2\subseteq \mathcal{F}_3\subseteq\ldots&amp;lt;/math&amp;gt; be a sequence of &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field. Is &amp;lt;math&amp;gt;\bigcup_{i=1}^{+\infty} \mathcal{F}_i&amp;lt;/math&amp;gt; a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field?&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Projection&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\mathcal{F}&amp;lt;/math&amp;gt; be a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field of subsets of &amp;lt;math&amp;gt;\Omega&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;T \subseteq \Omega&amp;lt;/math&amp;gt; be a subset. Show that &amp;lt;math&amp;gt;\{S \cap T \mid S \in \mathcal{F}\}&amp;lt;/math&amp;gt; is a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Probability space?&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\Omega = \mathbb{R}&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathcal{F}&amp;lt;/math&amp;gt; is the set of all subsets &amp;lt;math&amp;gt;A \subseteq \Omega&amp;lt;/math&amp;gt; so that &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;\overline{A}&amp;lt;/math&amp;gt; (complement of &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;) is countable, &amp;lt;math&amp;gt;P(A) = 0&amp;lt;/math&amp;gt; in the first case and &amp;lt;math&amp;gt;P(A) = 1&amp;lt;/math&amp;gt; in the second. Is &amp;lt;math&amp;gt;(\Omega,\mathcal{F},P)&amp;lt;/math&amp;gt; a probability space? Please explain your answer.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-3-probability-space&amp;quot;&amp;gt;Problem 3 (Birthday paradox)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Please design a &amp;lt;strong&amp;gt;randomized algorithm using the birthday paradox&amp;lt;/strong&amp;gt; that solves the following problem in &amp;lt;math&amp;gt;\mathrm{poly}(n) \cdot 2^{n/2}&amp;lt;/math&amp;gt; time with high probability (for example, &amp;lt;math&amp;gt;0.99&amp;lt;/math&amp;gt; when &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; is sufficiently large). Please provide a detailed error analysis as well. (&amp;lt;strong&amp;gt;WARNING:&amp;lt;/strong&amp;gt; You will &amp;lt;strong&amp;gt;NOT&amp;lt;/strong&amp;gt; receive any points if you solve this task using Gaussian elimination)&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Given an integer sequence &amp;lt;math&amp;gt;a_1,a_2,\ldots,a_{100n}&amp;lt;/math&amp;gt; of length &amp;lt;math&amp;gt;100 n&amp;lt;/math&amp;gt; satisfying &amp;lt;math&amp;gt;0 \le a_i &amp;lt; 2^n&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;1 \le i \le 100 n&amp;lt;/math&amp;gt;. Please find out a non-empty subset &amp;lt;math&amp;gt;S \subseteq \{1,2,\ldots,100n\}&amp;lt;/math&amp;gt; satisfying &amp;lt;math&amp;gt;\bigoplus_{i \in S} a_i = 0&amp;lt;/math&amp;gt;, i.e. the exclusive or of the elements whose indices are in &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; equals to &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-4-conditional-probability&amp;quot;&amp;gt;Problem 4 (Conditional probability)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Positive correlation&amp;lt;/strong&amp;gt;] We say that events &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; gives &#039;&#039;positive information&#039;&#039; about event &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; if &amp;lt;math&amp;gt;\mathbf{Pr}(A|B) &amp;gt; \mathbf{Pr}(A)&amp;lt;/math&amp;gt;, that is, the occurrence of &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; makes the occurrence of &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; more likely. Now suppose that &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; gives positive information about &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;.&lt;br /&gt;
# Does &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; give positive information about &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;?&lt;br /&gt;
# Does &amp;lt;math&amp;gt;\overline{B}&amp;lt;/math&amp;gt; give negative information about &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;, that is, is it true that &amp;lt;math&amp;gt;\mathbf{Pr}(A|\overline{B}) &amp;lt; \mathbf{Pr}(A)&amp;lt;/math&amp;gt;?&lt;br /&gt;
# Does &amp;lt;math&amp;gt;\overline{B}&amp;lt;/math&amp;gt; give positive information or negative information about &amp;lt;math&amp;gt;\overline{A}&amp;lt;/math&amp;gt;?&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Balls in urns (I)&amp;lt;/strong&amp;gt;] There are &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; urns of which the &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;-th contains &amp;lt;math&amp;gt;r-1&amp;lt;/math&amp;gt; white balls and &amp;lt;math&amp;gt;n-r&amp;lt;/math&amp;gt; black balls. You pick an urn uniformly at random (here, &amp;quot;uniformly&amp;quot; means that each urn has equal probability of being chosen) and remove two balls from that urn, uniformly at random without replacement (which means that each of the &amp;lt;math&amp;gt;{n-1\choose 2}&amp;lt;/math&amp;gt; pairs of balls are chosen to be removed with equal probability). Find the following probabilities:&lt;br /&gt;
# the second ball is black;&lt;br /&gt;
# the second ball is black, given that the first is black.&lt;br /&gt;
&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Balls in urns (II)&amp;lt;/strong&amp;gt;] Suppose that an urn contains &amp;lt;math&amp;gt;w&amp;lt;/math&amp;gt; white balls and &amp;lt;math&amp;gt;b&amp;lt;/math&amp;gt; black balls. The balls are drawn from the urn one by one, each time uniformly and independently at random, without replacement (which means we do not put the chosen ball back after each drawing). Find the probabilities of the events:&lt;br /&gt;
# the first white ball drawn is the &amp;lt;math&amp;gt;(k+1)&amp;lt;/math&amp;gt;th ball;&lt;br /&gt;
# the last ball drawn is white.&lt;br /&gt;
&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-5-independence&amp;quot;&amp;gt;Problem 5 (Independence)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Let&amp;amp;#39;s consider a series of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; outputs &amp;lt;math&amp;gt;(X_1, X_2, \cdots, X_n) \in \{0,1\}^n&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; independent Bernoulli trials, where each trial succeeds with the same probability  &amp;lt;math&amp;gt;0 &amp;lt; p &amp;lt; 1&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Limited independence&amp;lt;/strong&amp;gt;] Construct three events &amp;lt;math&amp;gt;A,B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; Bernoulli trials such that &amp;lt;math&amp;gt;A, B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; are pairwise independent but are not (mutually) independent. You need to prove that the constructed events &amp;lt;math&amp;gt;A, B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; satisfy this. (Hint: Consider the case where &amp;lt;math&amp;gt;n = 2&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;p = 1/2&amp;lt;/math&amp;gt;.)&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Product distribution&amp;lt;/strong&amp;gt;] Suppose someone has observed the output of the &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; trials, and she told you that precisely &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; trials succeeded for some &amp;lt;math&amp;gt;0&amp;lt; k&amp;lt; n&amp;lt;/math&amp;gt;. Now you want to predict the output of the &amp;lt;math&amp;gt;(n+1)&amp;lt;/math&amp;gt;-th trial while the parameter &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; of the Bernoulli trial is unknown. One way to estimate &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; is to find such &amp;lt;math&amp;gt;\hat{p}&amp;lt;/math&amp;gt; that makes the observed outcomes most probable, namely you need to solve &lt;br /&gt;
  &amp;lt;math&amp;gt;\arg \max_{\hat{p}\in(0,1)} \mathbf{Pr}_{\hat{p}} [k \text{ out of } n\text{ trials succeed}].&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Estimate &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; by solving the above optimization problem.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;If someone tells you exactly which &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; trials succeed (in addition to just telling you the number of successful trials, which is &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;), would it help you to estimate &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; more accurately? Why?&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-6-probabilistic-method&amp;quot;&amp;gt;Problem 6 (Probabilistic method)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Tournaments&amp;lt;/strong&amp;gt;] A [[wikipedia:Tournament_(graph_theory)|tournament]] can be interpreted as the outcome of a round-robin tournament in which every player faces every other player exactly once, and no draws occur. Given two players (vertices) &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;, we draw an&lt;br /&gt;
arrow from &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; if &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; beats &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; (and vice versa). We say a tournament has property &amp;lt;math&amp;gt;S_k&amp;lt;/math&amp;gt; if for every &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; players, there exists another player &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt; who defeats all of them. For example, a triangle &amp;lt;math&amp;gt;x\rightarrow y \rightarrow z \rightarrow x&amp;lt;/math&amp;gt; has property &amp;lt;math&amp;gt;S_1&amp;lt;/math&amp;gt; but not &amp;lt;math&amp;gt;S_2&amp;lt;/math&amp;gt;. Prove that if &amp;lt;math&amp;gt;\binom n k(1-2^{-k})^{n-k}&amp;lt;1&amp;lt;/math&amp;gt;, then there  is a tournament on &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; vertices that has the property &amp;lt;math&amp;gt;S_k&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_1&amp;diff=12909</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 1</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_1&amp;diff=12909"/>
		<updated>2025-02-20T06:47:46Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 1==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-2-principle-of-inclusion-and-exclusion&amp;quot;&amp;gt;Problem 1 (Principle of Inclusion and Exclusion, 8 points)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Let &amp;lt;math&amp;gt;n\ge 1&amp;lt;/math&amp;gt; be a positive integer and &amp;lt;math&amp;gt;A_1,A_2,\ldots,A_n&amp;lt;/math&amp;gt; be &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; events.&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Union bound&amp;lt;/strong&amp;gt;] Prove &amp;lt;math&amp;gt;\mathbf{Pr}\left( \bigcup_{i=1}^{n} A_i \right) \le \sum_{i=1}^{n} \mathbf{Pr}\left(A_i\right)&amp;lt;/math&amp;gt;&amp;lt;strong&amp;gt; using the definition of probability space&amp;lt;/strong&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Principle of Inclusion and Exclusion (PIE)&amp;lt;/strong&amp;gt;] Prove that &amp;lt;math&amp;gt;\mathbf{Pr}\left( \bigcup_{i=1}^n A_i\right) = \sum_{\emptyset \neq S \subseteq [n]} (-1)^{|S|-1} \mathbf{Pr}\left( \bigcap_{i \in S} A_i \right)&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;[n]=\{1,2,\ldots,n\}&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Surjection&amp;lt;/strong&amp;gt;] For positive integers &amp;lt;math&amp;gt;m\ge n&amp;lt;/math&amp;gt;, prove that the probability of a uniform random function &amp;lt;math&amp;gt;f:[m]\to[n]&amp;lt;/math&amp;gt; to be surjective (满射) is &amp;lt;math&amp;gt;\sum_{k=1}^n(-1)^{n-k}{n\choose k}\left(\frac{k}{n}\right)^m&amp;lt;/math&amp;gt;. &amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Bonferroni&amp;amp;#39;s inequality and Kounias&amp;amp;#39; inequality&amp;lt;/strong&amp;gt;] Prove that &lt;br /&gt;
&amp;lt;math&amp;gt;&lt;br /&gt;
\sum_{i=1}^n \mathbf{Pr}(A_i) - \sum_{1 \le i&amp;lt; j \le n} \mathbf{Pr}(A_i \cap A_j)\le \mathbf{Pr}\left(\bigcup_{i=1}^n A_i\right) \le \sum_{i=1}^n \mathbf{Pr} \left( A_i\right) - \sum_{i=2}^n \mathbf{Pr}(A_1 \cap A_i).&lt;br /&gt;
&amp;lt;/math&amp;gt; &lt;br /&gt;
(Hint: This is sometimes called Kounias&#039; inequality which is weaker than the Bonferroni&#039;s inequality. You can try using Venn diagram to understand these inequalities.)&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-2-probability-space&amp;quot;&amp;gt;Problem 2 (Probability space, 12 points)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Nonexistence of probability space&amp;lt;/strong&amp;gt;] Prove that it is impossible to define a uniform probability law on natural numbers &amp;lt;math&amp;gt;\mathbb{N}&amp;lt;/math&amp;gt;. More precisely, prove that there does not exist a probability space &amp;lt;math&amp;gt;(\mathbb{N},2^{\mathbb{N}},\mathbf{Pr})&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;\mathbf{Pr}(\{i\}) = \mathbf{Pr}(\{j\})&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;i, j \in \mathbb{N}&amp;lt;/math&amp;gt;. &lt;br /&gt;
Please explain why the same argument fails to prove that there is no uniform probability law on the real interval &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt;, that is, there is no such probability space &amp;lt;math&amp;gt;([0,1],\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt; that for any interval &amp;lt;math&amp;gt;(l,r] \subseteq [0,1]&amp;lt;/math&amp;gt;, it holds that &amp;lt;math&amp;gt;(l,r] \in \mathcal{F}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{Pr}( (l,r] ) = r-l&amp;lt;/math&amp;gt;. (Actually, such probability measure does exist and is called the Lebesgue measure on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt;). &amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Smallest &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field (I)&amp;lt;/strong&amp;gt;] For any subset &amp;lt;math&amp;gt;S \subseteq 2^\Omega&amp;lt;/math&amp;gt;, prove that the smallest &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field containing &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; is given by &amp;lt;math&amp;gt;\sigma(S) := \bigcap_{\substack{S \subseteq \mathcal{F} \subseteq 2^\Omega\\ \mathcal{F} \text{ is a } \sigma\text{-field }} } \mathcal{F}&amp;lt;/math&amp;gt;. (Hint: You should show that it is indeed a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field and also it is the smallest one containing &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;.)&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Smallest &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field (II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;S,T \subseteq 2^{\Omega}&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\sigma(S) = \sigma(T)&amp;lt;/math&amp;gt; if and only if &amp;lt;math&amp;gt;S \subseteq \sigma(T)&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;T \subseteq \sigma(S)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Union of &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\mathcal{F}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathcal{G}&amp;lt;/math&amp;gt; be &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-fields of subsets of &amp;lt;math&amp;gt;\Omega&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathcal{F} \cup \mathcal{G}&amp;lt;/math&amp;gt; is not necessarily a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field. Suppose &amp;lt;math&amp;gt;\mathcal{F}_1, \mathcal{F}_2,\ldots&amp;lt;/math&amp;gt; be a sequence of &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field satisfying &amp;lt;math&amp;gt;\mathcal{F}_1 \subseteq \mathcal{F}_2 \subseteq \ldots&amp;lt;/math&amp;gt;. Is &amp;lt;math&amp;gt;\bigcup_{i=1}^{+\infty} \mathcal{F}_i&amp;lt;/math&amp;gt; a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field?&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Projection&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\mathcal{F}&amp;lt;/math&amp;gt; be a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field of subsets of &amp;lt;math&amp;gt;\Omega&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;T \subseteq \Omega&amp;lt;/math&amp;gt; be a subset. Show that &amp;lt;math&amp;gt;\{S \cap T \mid S \in \mathcal{F}\}&amp;lt;/math&amp;gt; is a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Probability space?&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\Omega = \mathbb{R}&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathcal{F}&amp;lt;/math&amp;gt; is the set of all subsets &amp;lt;math&amp;gt;A \subseteq \Omega&amp;lt;/math&amp;gt; so that &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;\overline{A}&amp;lt;/math&amp;gt; (complement of &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;) is countable, &amp;lt;math&amp;gt;P(A) = 0&amp;lt;/math&amp;gt; in the first case and &amp;lt;math&amp;gt;P(A) = 1&amp;lt;/math&amp;gt; in the second. Is &amp;lt;math&amp;gt;(\Omega,\mathcal{F},P)&amp;lt;/math&amp;gt; a probability space? Please explain your answer.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-3-probability-space&amp;quot;&amp;gt;Problem 3 (Birthday paradox)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Please design a &amp;lt;strong&amp;gt;randomized algorithm using the birthday paradox&amp;lt;/strong&amp;gt; that solves the following problem in &amp;lt;math&amp;gt;\mathrm{poly}(n) \cdot 2^{n/2}&amp;lt;/math&amp;gt; time with high probability (for example, &amp;lt;math&amp;gt;0.99&amp;lt;/math&amp;gt; when &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; is sufficiently large). Please provide a detailed error analysis as well. (&amp;lt;strong&amp;gt;WARNING:&amp;lt;/strong&amp;gt; You will &amp;lt;strong&amp;gt;NOT&amp;lt;/strong&amp;gt; receive any points if you solve this task using Gaussian elimination)&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Given an integer sequence &amp;lt;math&amp;gt;a_1,a_2,\ldots,a_{100n}&amp;lt;/math&amp;gt; of length &amp;lt;math&amp;gt;100 n&amp;lt;/math&amp;gt; satisfying &amp;lt;math&amp;gt;0 \le a_i &amp;lt; 2^n&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;1 \le i \le 100 n&amp;lt;/math&amp;gt;. Please find out a non-empty subset &amp;lt;math&amp;gt;S \subseteq \{1,2,\ldots,100n\}&amp;lt;/math&amp;gt; satisfying &amp;lt;math&amp;gt;\bigoplus_{i \in S} a_i = 0&amp;lt;/math&amp;gt;, i.e. the exclusive or of the elements whose indices are in &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; equals to &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-4-conditional-probability&amp;quot;&amp;gt;Problem 4 (Conditional probability)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Positive correlation&amp;lt;/strong&amp;gt;] We say that events &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; gives &#039;&#039;positive information&#039;&#039; about event &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; if &amp;lt;math&amp;gt;\mathbf{Pr}(A|B) &amp;gt; \mathbf{Pr}(A)&amp;lt;/math&amp;gt;, that is, the occurrence of &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; makes the occurrence of &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; more likely. Now suppose that &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; gives positive information about &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;.&lt;br /&gt;
# Does &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; give positive information about &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;?&lt;br /&gt;
# Does &amp;lt;math&amp;gt;\overline{B}&amp;lt;/math&amp;gt; give negative information about &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;, that is, is it true that &amp;lt;math&amp;gt;\mathbf{Pr}(A|\overline{B}) &amp;lt; \mathbf{Pr}(A)&amp;lt;/math&amp;gt;?&lt;br /&gt;
# Does &amp;lt;math&amp;gt;\overline{B}&amp;lt;/math&amp;gt; give positive information or negative information about &amp;lt;math&amp;gt;\overline{A}&amp;lt;/math&amp;gt;?&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Balls in urns (I)&amp;lt;/strong&amp;gt;] There are &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; urns of which the &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;-th contains &amp;lt;math&amp;gt;r-1&amp;lt;/math&amp;gt; white balls and &amp;lt;math&amp;gt;n-r&amp;lt;/math&amp;gt; black balls. You pick an urn uniformly at random (here, &amp;quot;uniformly&amp;quot; means that each urn has equal probability of being chosen) and remove two balls from that urn, uniformly at random without replacement (which means that each of the &amp;lt;math&amp;gt;{n-1\choose 2}&amp;lt;/math&amp;gt; pairs of balls are chosen to be removed with equal probability). Find the following probabilities:&lt;br /&gt;
# the second ball is black;&lt;br /&gt;
# the second ball is black, given that the first is black.&lt;br /&gt;
&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Balls in urns (II)&amp;lt;/strong&amp;gt;] Suppose that an urn contains &amp;lt;math&amp;gt;w&amp;lt;/math&amp;gt; white balls and &amp;lt;math&amp;gt;b&amp;lt;/math&amp;gt; black balls. The balls are drawn from the urn one by one, each time uniformly and independently at random, without replacement (which means we do not put the chosen ball back after each drawing). Find the probabilities of the events:&lt;br /&gt;
# the first white ball drawn is the &amp;lt;math&amp;gt;(k+1)&amp;lt;/math&amp;gt;th ball;&lt;br /&gt;
# the last ball drawn is white.&lt;br /&gt;
&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-5-independence&amp;quot;&amp;gt;Problem 5 (Independence)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Let&amp;amp;#39;s consider a series of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; outputs &amp;lt;math&amp;gt;(X_1, X_2, \cdots, X_n) \in \{0,1\}^n&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; independent Bernoulli trials, where each trial succeeds with the same probability  &amp;lt;math&amp;gt;0 &amp;lt; p &amp;lt; 1&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Limited independence&amp;lt;/strong&amp;gt;] Construct three events &amp;lt;math&amp;gt;A,B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; Bernoulli trials such that &amp;lt;math&amp;gt;A, B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; are pairwise independent but are not (mutually) independent. You need to prove that the constructed events &amp;lt;math&amp;gt;A, B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; satisfy this. (Hint: Consider the case where &amp;lt;math&amp;gt;n = 2&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;p = 1/2&amp;lt;/math&amp;gt;.)&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Product distribution&amp;lt;/strong&amp;gt;] Suppose someone has observed the output of the &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; trials, and she told you that precisely &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; trials succeeded for some &amp;lt;math&amp;gt;0&amp;lt; k&amp;lt; n&amp;lt;/math&amp;gt;. Now you want to predict the output of the &amp;lt;math&amp;gt;(n+1)&amp;lt;/math&amp;gt;-th trial while the parameter &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; of the Bernoulli trial is unknown. One way to estimate &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; is to find such &amp;lt;math&amp;gt;\hat{p}&amp;lt;/math&amp;gt; that makes the observed outcomes most probable, namely you need to solve &lt;br /&gt;
  &amp;lt;math&amp;gt;\arg \max_{\hat{p}\in(0,1)} \mathbf{Pr}_{\hat{p}} [k \text{ out of } n\text{ trials succeed}].&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Estimate &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; by solving the above optimization problem.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;If someone tells you exactly which &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; trials succeed (in addition to just telling you the number of successful trials, which is &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;), would it help you to estimate &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; more accurately? Why?&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-6-probabilistic-method&amp;quot;&amp;gt;Problem 6 (Probabilistic method)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Tournaments&amp;lt;/strong&amp;gt;] A [[wikipedia:Tournament_(graph_theory)|tournament]] can be interpreted as the outcome of a round-robin tournament in which every player faces every other player exactly once, and no draws occur. Given two players (vertices) &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;, we draw an&lt;br /&gt;
arrow from &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; if &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; beats &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; (and vice versa). We say a tournament has property &amp;lt;math&amp;gt;S_k&amp;lt;/math&amp;gt; if for every &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; players, there exists another player &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt; who defeats all of them. For example, a triangle &amp;lt;math&amp;gt;x\rightarrow y \rightarrow z \rightarrow x&amp;lt;/math&amp;gt; has property &amp;lt;math&amp;gt;S_1&amp;lt;/math&amp;gt; but not &amp;lt;math&amp;gt;S_2&amp;lt;/math&amp;gt;. Prove that if &amp;lt;math&amp;gt;\binom n k(1-2^{-k})^{n-k}&amp;lt;1&amp;lt;/math&amp;gt;, then there  is a tournament on &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; vertices that has the property &amp;lt;math&amp;gt;S_k&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_1&amp;diff=12908</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 1</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_1&amp;diff=12908"/>
		<updated>2025-02-20T06:46:26Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: /* Assumption throughout Problem Set 1 */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 1==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-2-principle-of-inclusion-and-exclusion&amp;quot;&amp;gt;Problem 1 (Principle of Inclusion and Exclusion, 8 points)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Let &amp;lt;math&amp;gt;n\ge 1&amp;lt;/math&amp;gt; be a positive integer and &amp;lt;math&amp;gt;A_1,A_2,\ldots,A_n&amp;lt;/math&amp;gt; be &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; events.&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Union bound&amp;lt;/strong&amp;gt;] Prove &amp;lt;math&amp;gt;\mathbf{Pr}\left( \bigcup_{i=1}^{n} A_i \right) \le \sum_{i=1}^{n} \mathbf{Pr}\left(A_i\right)&amp;lt;/math&amp;gt;&amp;lt;strong&amp;gt; using the definition of probability space&amp;lt;/strong&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Principle of Inclusion and Exclusion (PIE)&amp;lt;/strong&amp;gt;] Prove that &amp;lt;math&amp;gt;\mathbf{Pr}\left( \bigcup_{i=1}^n A_i\right) = \sum_{\emptyset \neq S \subseteq [n]} (-1)^{|S|-1} \mathbf{Pr}\left( \bigcap_{i \in S} A_i \right)&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;[n]=\{1,2,\ldots,n\}&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Surjection&amp;lt;/strong&amp;gt;] For positive integers &amp;lt;math&amp;gt;m\ge n&amp;lt;/math&amp;gt;, prove that the probability of a uniform random function &amp;lt;math&amp;gt;f:[m]\to[n]&amp;lt;/math&amp;gt; to be surjective (满射) is &amp;lt;math&amp;gt;\sum_{k=1}^n(-1)^{n-k}{n\choose k}\left(\frac{k}{n}\right)^m&amp;lt;/math&amp;gt;. &amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Bonferroni&amp;amp;#39;s inequality and Kounias&amp;amp;#39; inequality&amp;lt;/strong&amp;gt;] Prove that &lt;br /&gt;
&amp;lt;math&amp;gt;&lt;br /&gt;
\sum_{i=1}^n \mathbf{Pr}(A_i) - \sum_{1 \le i&amp;lt; j \le n} \mathbf{Pr}(A_i \cap A_j)\le \mathbf{Pr}\left(\bigcup_{i=1}^n A_i\right) \le \sum_{i=1}^n \mathbf{Pr} \left( A_i\right) - \sum_{i=2}^n \mathbf{Pr}(A_1 \cap A_i).&lt;br /&gt;
&amp;lt;/math&amp;gt; &lt;br /&gt;
(Hint: This is sometimes called Kounias&#039; inequality which is weaker than the Bonferroni&#039;s inequality. You can try using Venn diagram to understand these inequalities.)&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-2-probability-space&amp;quot;&amp;gt;Problem 2 (Probability space, 12 points)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Nonexistence of probability space&amp;lt;/strong&amp;gt;] Prove that it is impossible to define a uniform probability law on natural numbers &amp;lt;math&amp;gt;\mathbb{N}&amp;lt;/math&amp;gt;. More precisely, prove that there does not exist a probability space &amp;lt;math&amp;gt;(\mathbb{N},2^{\mathbb{N}},\mathbf{Pr})&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;\mathbf{Pr}(\{i\}) = \mathbf{Pr}(\{j\})&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;i, j \in \mathbb{N}&amp;lt;/math&amp;gt;. &lt;br /&gt;
Please explain why the same argument fails to prove that there is no uniform probability law on the real interval &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt;, that is, there is no such probability space &amp;lt;math&amp;gt;([0,1],\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt; that for any interval &amp;lt;math&amp;gt;(l,r] \subseteq [0,1]&amp;lt;/math&amp;gt;, it holds that &amp;lt;math&amp;gt;(l,r] \in \mathcal{F}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{Pr}( (l,r] ) = r-l&amp;lt;/math&amp;gt;. (Actually, such probability measure does exist and is called the Lebesgue measure on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt;). &amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Smallest &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field (I)&amp;lt;/strong&amp;gt;] For any subset &amp;lt;math&amp;gt;S \subseteq 2^\Omega&amp;lt;/math&amp;gt;, prove that the smallest &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field containing &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; is given by &amp;lt;math&amp;gt;\sigma(S) := \bigcap_{\substack{S \subseteq \mathcal{F} \subseteq 2^\Omega\\ \mathcal{F} \text{ is a } \sigma\text{-field }} } \mathcal{F}&amp;lt;/math&amp;gt;. (Hint: You should show that it is indeed a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field and also it is the smallest one containing &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;.)&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Smallest &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field (II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;S,T \subseteq 2^{\Omega}&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\sigma(S) = \sigma(T)&amp;lt;/math&amp;gt; if and only if &amp;lt;math&amp;gt;S \subseteq \sigma(T)&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;T \subseteq \sigma(S)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Union of &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\mathcal{F}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathcal{G}&amp;lt;/math&amp;gt; be &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-fields of subsets of &amp;lt;math&amp;gt;\Omega&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathcal{F} \cup \mathcal{G}&amp;lt;/math&amp;gt; is not necessarily a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field. Suppose &amp;lt;math&amp;gt;\mathcal{F}_1, \mathcal{F}_2,\ldots&amp;lt;/math&amp;gt; be a sequence of &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field. Is &amp;lt;math&amp;gt;\bigcup_{i=1}^{+\infty} \mathcal{F}_i&amp;lt;/math&amp;gt; a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field?&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Projection&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\mathcal{F}&amp;lt;/math&amp;gt; be a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field of subsets of &amp;lt;math&amp;gt;\Omega&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;T \subseteq \Omega&amp;lt;/math&amp;gt; be a subset. Show that &amp;lt;math&amp;gt;\{S \cap T \mid S \in \mathcal{F}\}&amp;lt;/math&amp;gt; is a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Probability space?&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\Omega = \mathbb{R}&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathcal{F}&amp;lt;/math&amp;gt; is the set of all subsets &amp;lt;math&amp;gt;A \subseteq \Omega&amp;lt;/math&amp;gt; so that &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;\overline{A}&amp;lt;/math&amp;gt; (complement of &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;) is countable, &amp;lt;math&amp;gt;P(A) = 0&amp;lt;/math&amp;gt; in the first case and &amp;lt;math&amp;gt;P(A) = 1&amp;lt;/math&amp;gt; in the second. Is &amp;lt;math&amp;gt;(\Omega,\mathcal{F},P)&amp;lt;/math&amp;gt; a probability space? Please explain your answer.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-3-probability-space&amp;quot;&amp;gt;Problem 3 (Birthday paradox)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Please design a &amp;lt;strong&amp;gt;randomized algorithm using the birthday paradox&amp;lt;/strong&amp;gt; that solves the following problem in &amp;lt;math&amp;gt;\mathrm{poly}(n) \cdot 2^{n/2}&amp;lt;/math&amp;gt; time with high probability (for example, &amp;lt;math&amp;gt;0.99&amp;lt;/math&amp;gt; when &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; is sufficiently large). Please provide a detailed error analysis as well. (&amp;lt;strong&amp;gt;WARNING:&amp;lt;/strong&amp;gt; You will &amp;lt;strong&amp;gt;NOT&amp;lt;/strong&amp;gt; receive any points if you solve this task using Gaussian elimination)&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Given an integer sequence &amp;lt;math&amp;gt;a_1,a_2,\ldots,a_{100n}&amp;lt;/math&amp;gt; of length &amp;lt;math&amp;gt;100 n&amp;lt;/math&amp;gt; satisfying &amp;lt;math&amp;gt;0 \le a_i &amp;lt; 2^n&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;1 \le i \le 100 n&amp;lt;/math&amp;gt;. Please find out a non-empty subset &amp;lt;math&amp;gt;S \subseteq \{1,2,\ldots,100n\}&amp;lt;/math&amp;gt; satisfying &amp;lt;math&amp;gt;\bigoplus_{i \in S} a_i = 0&amp;lt;/math&amp;gt;, i.e. the exclusive or of the elements whose indices are in &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; equals to &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-4-conditional-probability&amp;quot;&amp;gt;Problem 4 (Conditional probability)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Positive correlation&amp;lt;/strong&amp;gt;] We say that events &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; gives &#039;&#039;positive information&#039;&#039; about event &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; if &amp;lt;math&amp;gt;\mathbf{Pr}(A|B) &amp;gt; \mathbf{Pr}(A)&amp;lt;/math&amp;gt;, that is, the occurrence of &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; makes the occurrence of &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; more likely. Now suppose that &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; gives positive information about &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;.&lt;br /&gt;
# Does &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; give positive information about &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;?&lt;br /&gt;
# Does &amp;lt;math&amp;gt;\overline{B}&amp;lt;/math&amp;gt; give negative information about &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;, that is, is it true that &amp;lt;math&amp;gt;\mathbf{Pr}(A|\overline{B}) &amp;lt; \mathbf{Pr}(A)&amp;lt;/math&amp;gt;?&lt;br /&gt;
# Does &amp;lt;math&amp;gt;\overline{B}&amp;lt;/math&amp;gt; give positive information or negative information about &amp;lt;math&amp;gt;\overline{A}&amp;lt;/math&amp;gt;?&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Balls in urns (I)&amp;lt;/strong&amp;gt;] There are &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; urns of which the &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;-th contains &amp;lt;math&amp;gt;r-1&amp;lt;/math&amp;gt; white balls and &amp;lt;math&amp;gt;n-r&amp;lt;/math&amp;gt; black balls. You pick an urn uniformly at random (here, &amp;quot;uniformly&amp;quot; means that each urn has equal probability of being chosen) and remove two balls from that urn, uniformly at random without replacement (which means that each of the &amp;lt;math&amp;gt;{n-1\choose 2}&amp;lt;/math&amp;gt; pairs of balls are chosen to be removed with equal probability). Find the following probabilities:&lt;br /&gt;
# the second ball is black;&lt;br /&gt;
# the second ball is black, given that the first is black.&lt;br /&gt;
&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Balls in urns (II)&amp;lt;/strong&amp;gt;] Suppose that an urn contains &amp;lt;math&amp;gt;w&amp;lt;/math&amp;gt; white balls and &amp;lt;math&amp;gt;b&amp;lt;/math&amp;gt; black balls. The balls are drawn from the urn one by one, each time uniformly and independently at random, without replacement (which means we do not put the chosen ball back after each drawing). Find the probabilities of the events:&lt;br /&gt;
# the first white ball drawn is the &amp;lt;math&amp;gt;(k+1)&amp;lt;/math&amp;gt;th ball;&lt;br /&gt;
# the last ball drawn is white.&lt;br /&gt;
&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-5-independence&amp;quot;&amp;gt;Problem 5 (Independence)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Let&amp;amp;#39;s consider a series of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; outputs &amp;lt;math&amp;gt;(X_1, X_2, \cdots, X_n) \in \{0,1\}^n&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; independent Bernoulli trials, where each trial succeeds with the same probability  &amp;lt;math&amp;gt;0 &amp;lt; p &amp;lt; 1&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Limited independence&amp;lt;/strong&amp;gt;] Construct three events &amp;lt;math&amp;gt;A,B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; Bernoulli trials such that &amp;lt;math&amp;gt;A, B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; are pairwise independent but are not (mutually) independent. You need to prove that the constructed events &amp;lt;math&amp;gt;A, B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; satisfy this. (Hint: Consider the case where &amp;lt;math&amp;gt;n = 2&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;p = 1/2&amp;lt;/math&amp;gt;.)&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Product distribution&amp;lt;/strong&amp;gt;] Suppose someone has observed the output of the &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; trials, and she told you that precisely &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; trials succeeded for some &amp;lt;math&amp;gt;0&amp;lt; k&amp;lt; n&amp;lt;/math&amp;gt;. Now you want to predict the output of the &amp;lt;math&amp;gt;(n+1)&amp;lt;/math&amp;gt;-th trial while the parameter &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; of the Bernoulli trial is unknown. One way to estimate &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; is to find such &amp;lt;math&amp;gt;\hat{p}&amp;lt;/math&amp;gt; that makes the observed outcomes most probable, namely you need to solve &lt;br /&gt;
  &amp;lt;math&amp;gt;\arg \max_{\hat{p}\in(0,1)} \mathbf{Pr}_{\hat{p}} [k \text{ out of } n\text{ trials succeed}].&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Estimate &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; by solving the above optimization problem.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;If someone tells you exactly which &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; trials succeed (in addition to just telling you the number of successful trials, which is &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;), would it help you to estimate &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; more accurately? Why?&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-6-probabilistic-method&amp;quot;&amp;gt;Problem 6 (Probabilistic method)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Tournaments&amp;lt;/strong&amp;gt;] A [[wikipedia:Tournament_(graph_theory)|tournament]] can be interpreted as the outcome of a round-robin tournament in which every player faces every other player exactly once, and no draws occur. Given two players (vertices) &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;, we draw an&lt;br /&gt;
arrow from &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; if &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; beats &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; (and vice versa). We say a tournament has property &amp;lt;math&amp;gt;S_k&amp;lt;/math&amp;gt; if for every &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; players, there exists another player &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt; who defeats all of them. For example, a triangle &amp;lt;math&amp;gt;x\rightarrow y \rightarrow z \rightarrow x&amp;lt;/math&amp;gt; has property &amp;lt;math&amp;gt;S_1&amp;lt;/math&amp;gt; but not &amp;lt;math&amp;gt;S_2&amp;lt;/math&amp;gt;. Prove that if &amp;lt;math&amp;gt;\binom n k(1-2^{-k})^{n-k}&amp;lt;1&amp;lt;/math&amp;gt;, then there  is a tournament on &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; vertices that has the property &amp;lt;math&amp;gt;S_k&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_1&amp;diff=12905</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 1</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_1&amp;diff=12905"/>
		<updated>2025-02-19T13:24:12Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 1==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-2-principle-of-inclusion-and-exclusion&amp;quot;&amp;gt;Problem 1 (Principle of Inclusion and Exclusion)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Let &amp;lt;math&amp;gt;n\ge 1&amp;lt;/math&amp;gt; be a positive integer and &amp;lt;math&amp;gt;A_1,A_2,\ldots,A_n&amp;lt;/math&amp;gt; be &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; events.&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Union bound&amp;lt;/strong&amp;gt;] Prove &amp;lt;math&amp;gt;\mathbf{Pr}\left( \bigcup_{i=1}^{n} A_i \right) \le \sum_{i=1}^{n} \mathbf{Pr}\left(A_i\right)&amp;lt;/math&amp;gt;&amp;lt;strong&amp;gt; using the definition of probability space&amp;lt;/strong&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Principle of Inclusion and Exclusion (PIE)&amp;lt;/strong&amp;gt;] Prove that &amp;lt;math&amp;gt;\mathbf{Pr}\left( \bigcup_{i=1}^n A_i\right) = \sum_{\emptyset \neq S \subseteq [n]} (-1)^{|S|-1} \mathbf{Pr}\left( \bigcap_{i \in S} A_i \right)&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;[n]=\{1,2,\ldots,n\}&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Surjection&amp;lt;/strong&amp;gt;] For positive integers &amp;lt;math&amp;gt;m\ge n&amp;lt;/math&amp;gt;, prove that the probability of a uniform random function &amp;lt;math&amp;gt;f:[m]\to[n]&amp;lt;/math&amp;gt; to be surjective (满射) is &amp;lt;math&amp;gt;\sum_{k=1}^n(-1)^{n-k}{n\choose k}\left(\frac{k}{n}\right)^m&amp;lt;/math&amp;gt;. &amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Bonferroni&amp;amp;#39;s inequality and Kounias&amp;amp;#39; inequality&amp;lt;/strong&amp;gt;] Prove that &lt;br /&gt;
&amp;lt;math&amp;gt;&lt;br /&gt;
\sum_{i=1}^n \mathbf{Pr}(A_i) - \sum_{1 \le i&amp;lt; j \le n} \mathbf{Pr}(A_i \cap A_j)\le \mathbf{Pr}\left(\bigcup_{i=1}^n A_i\right) \le \sum_{i=1}^n \mathbf{Pr} \left( A_i\right) - \sum_{i=2}^n \mathbf{Pr}(A_1 \cap A_i).&lt;br /&gt;
&amp;lt;/math&amp;gt; &lt;br /&gt;
(Hint: This is sometimes called Kounias&#039; inequality which is weaker than the Bonferroni&#039;s inequality. You can try using Venn diagram to understand these inequalities.)&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-2-probability-space&amp;quot;&amp;gt;Problem 2 (Probability space)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Nonexistence of probability space&amp;lt;/strong&amp;gt;] Prove that it is impossible to define a uniform probability law on natural numbers &amp;lt;math&amp;gt;\mathbb{N}&amp;lt;/math&amp;gt;. More precisely, prove that there does not exist a probability space &amp;lt;math&amp;gt;(\mathbb{N},2^{\mathbb{N}},\mathbf{Pr})&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;\mathbf{Pr}(\{i\}) = \mathbf{Pr}(\{j\})&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;i, j \in \mathbb{N}&amp;lt;/math&amp;gt;. &lt;br /&gt;
Please explain why the same argument fails to prove that there is no uniform probability law on the real interval &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt;, that is, there is no such probability space &amp;lt;math&amp;gt;([0,1],\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt; that for any interval &amp;lt;math&amp;gt;(l,r] \subseteq [0,1]&amp;lt;/math&amp;gt;, it holds that &amp;lt;math&amp;gt;(l,r] \in \mathcal{F}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{Pr}( (l,r] ) = r-l&amp;lt;/math&amp;gt;. (Actually, such probability measure does exist and is called the Lebesgue measure on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt;). &amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Smallest &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-algebra (I)&amp;lt;/strong&amp;gt;] For any subset &amp;lt;math&amp;gt;S \subseteq 2^\Omega&amp;lt;/math&amp;gt;, prove that the smallest &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field containing &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; is given by &amp;lt;math&amp;gt;\sigma(S) := \bigcap_{\substack{S \subseteq \mathcal{F} \subseteq 2^\Omega\\ \mathcal{F} \text{ is a } \sigma\text{-field }} } \mathcal{F}&amp;lt;/math&amp;gt;. (Hint: You should show that it is indeed a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field and also it is the smallest one containing &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;.)&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Smallest &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field (II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;S,T \subseteq 2^{\Omega}&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\sigma(S) = \sigma(T)&amp;lt;/math&amp;gt; if and only if &amp;lt;math&amp;gt;S \subseteq \sigma(T)&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;T \subseteq \sigma(S)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Union of &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\mathcal{F}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathcal{G}&amp;lt;/math&amp;gt; be &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-fields of subsets of &amp;lt;math&amp;gt;\Omega&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathcal{F} \cup \mathcal{G}&amp;lt;/math&amp;gt; is not necessarily a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Projection&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\mathcal{F}&amp;lt;/math&amp;gt; be a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field of subsets of &amp;lt;math&amp;gt;\Omega&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;T \subseteq \Omega&amp;lt;/math&amp;gt; be a subset. Show that &amp;lt;math&amp;gt;\{S \cap T \mid S \in \mathcal{F}\}&amp;lt;/math&amp;gt; is a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Probability space?&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\Omega = \mathbb{R}&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathcal{F}&amp;lt;/math&amp;gt; is the set of all subsets &amp;lt;math&amp;gt;A \subseteq \Omega&amp;lt;/math&amp;gt; so that &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;\overline{A}&amp;lt;/math&amp;gt; (complement of &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;) is countable, &amp;lt;math&amp;gt;P(A) = 0&amp;lt;/math&amp;gt; in the first case and &amp;lt;math&amp;gt;P(A) = 1&amp;lt;/math&amp;gt; in the second. Is &amp;lt;math&amp;gt;(\Omega,\mathcal{F},P)&amp;lt;/math&amp;gt; a probability space? Please explain your answer.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-3-probability-space&amp;quot;&amp;gt;Problem 3 (Birthday paradox)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Please design a &amp;lt;strong&amp;gt;randomized algorithm using the birthday paradox&amp;lt;/strong&amp;gt; that solves the following problem in &amp;lt;math&amp;gt;\mathrm{poly}(n) \cdot 2^{n/2}&amp;lt;/math&amp;gt; time with high probability (for example, &amp;lt;math&amp;gt;0.99&amp;lt;/math&amp;gt; when &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; is sufficiently large). Please provide a detailed error analysis as well. (&amp;lt;strong&amp;gt;WARNING:&amp;lt;/strong&amp;gt; You will &amp;lt;strong&amp;gt;NOT&amp;lt;/strong&amp;gt; receive any points if you solve this task using Gaussian elimination)&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Given an integer sequence &amp;lt;math&amp;gt;a_1,a_2,\ldots,a_{100n}&amp;lt;/math&amp;gt; of length &amp;lt;math&amp;gt;100 n&amp;lt;/math&amp;gt; satisfying &amp;lt;math&amp;gt;0 \le a_i &amp;lt; 2^n&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;1 \le i \le 100 n&amp;lt;/math&amp;gt;. Please find out a non-empty subset &amp;lt;math&amp;gt;S \subseteq \{1,2,\ldots,100n\}&amp;lt;/math&amp;gt; satisfying &amp;lt;math&amp;gt;\bigoplus_{i \in S} a_i = 0&amp;lt;/math&amp;gt;, i.e. the exclusive or of the elements whose indices are in &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; equals to &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-4-conditional-probability&amp;quot;&amp;gt;Problem 4 (Conditional probability)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Positive correlation&amp;lt;/strong&amp;gt;] We say that events &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; gives &#039;&#039;positive information&#039;&#039; about event &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; if &amp;lt;math&amp;gt;\mathbf{Pr}(A|B) &amp;gt; \mathbf{Pr}(A)&amp;lt;/math&amp;gt;, that is, the occurrence of &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; makes the occurrence of &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; more likely. Now suppose that &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; gives positive information about &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;.&lt;br /&gt;
# Does &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; give positive information about &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;?&lt;br /&gt;
# Does &amp;lt;math&amp;gt;\overline{B}&amp;lt;/math&amp;gt; give negative information about &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;, that is, is it true that &amp;lt;math&amp;gt;\mathbf{Pr}(A|\overline{B}) &amp;lt; \mathbf{Pr}(A)&amp;lt;/math&amp;gt;?&lt;br /&gt;
# Does &amp;lt;math&amp;gt;\overline{B}&amp;lt;/math&amp;gt; give positive information or negative information about &amp;lt;math&amp;gt;\overline{A}&amp;lt;/math&amp;gt;?&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Balls in urns (I)&amp;lt;/strong&amp;gt;] There are &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; urns of which the &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;-th contains &amp;lt;math&amp;gt;r-1&amp;lt;/math&amp;gt; white balls and &amp;lt;math&amp;gt;n-r&amp;lt;/math&amp;gt; black balls. You pick an urn uniformly at random (here, &amp;quot;uniformly&amp;quot; means that each urn has equal probability of being chosen) and remove two balls from that urn, uniformly at random without replacement (which means that each of the &amp;lt;math&amp;gt;{n-1\choose 2}&amp;lt;/math&amp;gt; pairs of balls are chosen to be removed with equal probability). Find the following probabilities:&lt;br /&gt;
# the second ball is black;&lt;br /&gt;
# the second ball is black, given that the first is black.&lt;br /&gt;
&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Balls in urns (II)&amp;lt;/strong&amp;gt;] Suppose that an urn contains &amp;lt;math&amp;gt;w&amp;lt;/math&amp;gt; white balls and &amp;lt;math&amp;gt;b&amp;lt;/math&amp;gt; black balls. The balls are drawn from the urn one by one, each time uniformly and independently at random, without replacement (which means we do not put the chosen ball back after each drawing). Find the probabilities of the events:&lt;br /&gt;
# the first white ball drawn is the &amp;lt;math&amp;gt;(k+1)&amp;lt;/math&amp;gt;th ball;&lt;br /&gt;
# the last ball drawn is white.&lt;br /&gt;
&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-5-independence&amp;quot;&amp;gt;Problem 5 (Independence)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Let&amp;amp;#39;s consider a series of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; outputs &amp;lt;math&amp;gt;(X_1, X_2, \cdots, X_n) \in \{0,1\}^n&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; independent Bernoulli trials, where each trial succeeds with the same probability  &amp;lt;math&amp;gt;0 &amp;lt; p &amp;lt; 1&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Limited independence&amp;lt;/strong&amp;gt;] Construct three events &amp;lt;math&amp;gt;A,B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; Bernoulli trials such that &amp;lt;math&amp;gt;A, B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; are pairwise independent but are not (mutually) independent. You need to prove that the constructed events &amp;lt;math&amp;gt;A, B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; satisfy this. (Hint: Consider the case where &amp;lt;math&amp;gt;n = 2&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;p = 1/2&amp;lt;/math&amp;gt;.)&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Product distribution&amp;lt;/strong&amp;gt;] Suppose someone has observed the output of the &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; trials, and she told you that precisely &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; trials succeeded for some &amp;lt;math&amp;gt;0&amp;lt; k&amp;lt; n&amp;lt;/math&amp;gt;. Now you want to predict the output of the &amp;lt;math&amp;gt;(n+1)&amp;lt;/math&amp;gt;-th trial while the parameter &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; of the Bernoulli trial is unknown. One way to estimate &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; is to find such &amp;lt;math&amp;gt;\hat{p}&amp;lt;/math&amp;gt; that makes the observed outcomes most probable, namely you need to solve &lt;br /&gt;
  &amp;lt;math&amp;gt;\arg \max_{\hat{p}\in(0,1)} \mathbf{Pr}_{\hat{p}} [k \text{ out of } n\text{ trials succeed}].&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Estimate &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; by solving the above optimization problem.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;If someone tells you exactly which &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; trials succeed (in addition to just telling you the number of successful trials, which is &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;), would it help you to estimate &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; more accurately? Why?&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-6-probabilistic-method&amp;quot;&amp;gt;Problem 6 (Probabilistic method)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Tournaments&amp;lt;/strong&amp;gt;] A [[wikipedia:Tournament_(graph_theory)|tournament]] can be interpreted as the outcome of a round-robin tournament in which every player faces every other player exactly once, and no draws occur. Given two players (vertices) &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;, we draw an&lt;br /&gt;
arrow from &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; if &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; beats &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; (and vice versa). We say a tournament has property &amp;lt;math&amp;gt;S_k&amp;lt;/math&amp;gt; if for every &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; players, there exists another player &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt; who defeats all of them. For example, a triangle &amp;lt;math&amp;gt;x\rightarrow y \rightarrow z \rightarrow x&amp;lt;/math&amp;gt; has property &amp;lt;math&amp;gt;S_1&amp;lt;/math&amp;gt; but not &amp;lt;math&amp;gt;S_2&amp;lt;/math&amp;gt;. Prove that if &amp;lt;math&amp;gt;\binom n k(1-2^{-k})^{n-k}&amp;lt;1&amp;lt;/math&amp;gt;, then there  is a tournament on &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; vertices that has the property &amp;lt;math&amp;gt;S_k&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)&amp;diff=12904</id>
		<title>概率论与数理统计 (Spring 2025)</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)&amp;diff=12904"/>
		<updated>2025-02-19T13:20:27Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: /* Assignments */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Infobox&lt;br /&gt;
|name         = Infobox&lt;br /&gt;
|bodystyle    = &lt;br /&gt;
|title        = &amp;lt;font size=3&amp;gt;&#039;&#039;&#039;概率论与数理统计&#039;&#039;&#039;&amp;lt;br&amp;gt;&lt;br /&gt;
&#039;&#039;&#039;Probability Theory&#039;&#039;&#039; &amp;lt;br&amp;gt; &amp;amp; &#039;&#039;&#039;Mathematical Statistics&#039;&#039;&#039;&amp;lt;/font&amp;gt;&lt;br /&gt;
|titlestyle   = &lt;br /&gt;
&lt;br /&gt;
|image        = &lt;br /&gt;
|imagestyle   = &lt;br /&gt;
|caption      = &lt;br /&gt;
|captionstyle = &lt;br /&gt;
|headerstyle  = background:#ccf;&lt;br /&gt;
|labelstyle   = background:#ddf;&lt;br /&gt;
|datastyle    = &lt;br /&gt;
&lt;br /&gt;
|header1 =Instructor&lt;br /&gt;
|label1  = &lt;br /&gt;
|data1   = &lt;br /&gt;
|header2 = &lt;br /&gt;
|label2  = &lt;br /&gt;
|data2   = &#039;&#039;&#039;尹一通&#039;&#039;&#039;&lt;br /&gt;
|header3 = &lt;br /&gt;
|label3  = Email&lt;br /&gt;
|data3   = yinyt@nju.edu.cn  &lt;br /&gt;
|header4 =&lt;br /&gt;
|label4  = office&lt;br /&gt;
|data4   = 计算机系 804&lt;br /&gt;
|header5 = &lt;br /&gt;
|label5  = &lt;br /&gt;
|data5   = &#039;&#039;&#039;刘景铖&#039;&#039;&#039;&lt;br /&gt;
|header6 = &lt;br /&gt;
|label6  = Email&lt;br /&gt;
|data6   = liu@nju.edu.cn  &lt;br /&gt;
|header7 =&lt;br /&gt;
|label7  = office&lt;br /&gt;
|data7   = 计算机系 516&lt;br /&gt;
|header8 = Class&lt;br /&gt;
|label8  = &lt;br /&gt;
|data8   = &lt;br /&gt;
|header9 =&lt;br /&gt;
|label9  = Class meeting&lt;br /&gt;
|data9   = Monday, 10am-12pm&amp;lt;br&amp;gt;&lt;br /&gt;
Wednesday (单), 10am-12pm&amp;lt;br&amp;gt;&lt;br /&gt;
仙Ⅰ-204&lt;br /&gt;
|header10=&lt;br /&gt;
|label10 = Office hour&lt;br /&gt;
|data10  = Friday, 2pm-3pm &amp;lt;br&amp;gt;计算机系 804（尹一通）&amp;lt;br&amp;gt;计算机系 516（刘景铖）&lt;br /&gt;
|header11= Textbook&lt;br /&gt;
|label11 = &lt;br /&gt;
|data11  = &lt;br /&gt;
|header12=&lt;br /&gt;
|label12 = &lt;br /&gt;
|data12  = [[File:概率导论.jpeg|border|100px]]&lt;br /&gt;
|header13=&lt;br /&gt;
|label13 = &lt;br /&gt;
|data13  = &#039;&#039;&#039;概率导论&#039;&#039;&#039;（第2版·修订版）&amp;lt;br&amp;gt; Dimitri P. Bertsekas and John N. Tsitsiklis&amp;lt;br&amp;gt; 郑忠国 童行伟 译；人民邮电出版社 (2022)&lt;br /&gt;
|header14=&lt;br /&gt;
|label14 = &lt;br /&gt;
|data14  = [[File:Grimmett_probability.jpg|border|100px]]&lt;br /&gt;
|header15=&lt;br /&gt;
|label15 = &lt;br /&gt;
|data15  = &#039;&#039;&#039;Probability and Random Processes&#039;&#039;&#039; (4E) &amp;lt;br&amp;gt; Geoffrey Grimmett and David Stirzaker &amp;lt;br&amp;gt;  Oxford University Press (2020)&lt;br /&gt;
|header16=&lt;br /&gt;
|label16 = &lt;br /&gt;
|data16  = [[File:Probability_and_Computing_2ed.jpg|border|100px]]&lt;br /&gt;
|header17=&lt;br /&gt;
|label17 = &lt;br /&gt;
|data17  = &#039;&#039;&#039;Probability and Computing&#039;&#039;&#039; (2E) &amp;lt;br&amp;gt; Michael Mitzenmacher and Eli Upfal &amp;lt;br&amp;gt;   Cambridge University Press (2017)&lt;br /&gt;
|belowstyle = background:#ddf;&lt;br /&gt;
|below = &lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
This is the webpage for the &#039;&#039;Probability Theory and Mathematical Statistics&#039;&#039; (概率论与数理统计) class of Spring 2025. Students who take this class should check this page periodically for content updates and new announcements. &lt;br /&gt;
&lt;br /&gt;
= Announcement =&lt;br /&gt;
* TBA&lt;br /&gt;
&lt;br /&gt;
= Course info =&lt;br /&gt;
* &#039;&#039;&#039;Instructor &#039;&#039;&#039;: &lt;br /&gt;
:* [http://tcs.nju.edu.cn/yinyt/ 尹一通]：[mailto:yinyt@nju.edu.cn &amp;lt;yinyt@nju.edu.cn&amp;gt;]，计算机系 804 &lt;br /&gt;
:* [https://liuexp.github.io 刘景铖]：[mailto:liu@nju.edu.cn &amp;lt;liu@nju.edu.cn&amp;gt;]，计算机系 516 &lt;br /&gt;
* &#039;&#039;&#039;Teaching assistant&#039;&#039;&#039;:&lt;br /&gt;
** [https://sites.google.com/view/xinyuanzhang 张昕渊]：[mailto:zhangxy@smail.nju.edu.cn &amp;lt;zhangxy@smail.nju.edu.cn&amp;gt;]，计算机系 410&lt;br /&gt;
** 邹宗瑞：[mailto:zou.zongrui@smail.nju.edu.cn &amp;lt;zou.zongrui@smail.nju.edu.cn&amp;gt;]，计算机系 410&lt;br /&gt;
* &#039;&#039;&#039;Class meeting&#039;&#039;&#039;:&lt;br /&gt;
** 周一：10am-12pm，仙Ⅰ-204&lt;br /&gt;
** 周三（单）：10am-12pm，仙Ⅰ-204&lt;br /&gt;
* &#039;&#039;&#039;Office hour&#039;&#039;&#039;: &lt;br /&gt;
:* 周五 2pm-3pm, 计算机系 804（尹一通）&lt;br /&gt;
:* 周五 2pm-3pm, 计算机系 516（刘景铖）&lt;br /&gt;
:* &#039;&#039;&#039;QQ群&#039;&#039;&#039;: 974742320（申请加入需提供姓名、院系、学号）&lt;br /&gt;
&lt;br /&gt;
= Syllabus =&lt;br /&gt;
课程内容分为三大部分：&lt;br /&gt;
* &#039;&#039;&#039;经典概率论&#039;&#039;&#039;：概率空间、随机变量及其数字特征、多维与连续随机变量、极限定理等内容&lt;br /&gt;
* &#039;&#039;&#039;概率与计算&#039;&#039;&#039;：测度集中现象 (concentration of measure)、概率法 (the probabilistic method)、离散随机过程的相关专题&lt;br /&gt;
* &#039;&#039;&#039;数理统计&#039;&#039;&#039;：参数估计、假设检验、贝叶斯估计、线性回归等统计推断等概念&lt;br /&gt;
&lt;br /&gt;
对于第一和第二部分，要求清楚掌握基本概念，深刻理解关键的现象与规律以及背后的原理，并可以灵活运用所学方法求解相关问题。对于第三部分，要求熟悉数理统计的若干基本概念，以及典型的统计模型和统计推断问题。&lt;br /&gt;
&lt;br /&gt;
经过本课程的训练，力求使学生能够熟悉掌握概率的语言，并会利用概率思维来理解客观世界并对其建模，以及驾驭概率的数学工具来分析和求解专业问题。&lt;br /&gt;
&lt;br /&gt;
=== 教材与参考书 Course Materials ===&lt;br /&gt;
* &#039;&#039;&#039;[BT]&#039;&#039;&#039; 概率导论（第2版·修订版），[美]伯特瑟卡斯（Dimitri P.Bertsekas）[美]齐齐克利斯（John N.Tsitsiklis）著，郑忠国 童行伟 译，人民邮电出版社（2022）。&lt;br /&gt;
* &#039;&#039;&#039;[GS]&#039;&#039;&#039; &#039;&#039;Probability and Random Processes&#039;&#039;, by Geoffrey Grimmett and David Stirzaker; Oxford University Press; 4th edition (2020).&lt;br /&gt;
* &#039;&#039;&#039;[MU]&#039;&#039;&#039; &#039;&#039;Probability and Computing: Randomization and Probabilistic Techniques in Algorithms and Data Analysis&#039;&#039;, by Michael Mitzenmacher, Eli Upfal; Cambridge University Press; 2nd edition (2017).&lt;br /&gt;
&lt;br /&gt;
=== 成绩 Grading Policy ===&lt;br /&gt;
* 课程成绩：本课程将会有若干次作业和一次期末考试。最终成绩将由平时作业成绩和期末考试成绩综合得出。&lt;br /&gt;
* 迟交：如果有特殊的理由，无法按时完成作业，请提前联系授课老师，给出正当理由。否则迟交的作业将不被接受。&lt;br /&gt;
&lt;br /&gt;
=== &amp;lt;font color=red&amp;gt; 学术诚信 Academic Integrity &amp;lt;/font&amp;gt;===&lt;br /&gt;
学术诚信是所有从事学术活动的学生和学者最基本的职业道德底线，本课程将不遗余力的维护学术诚信规范，违反这一底线的行为将不会被容忍。&lt;br /&gt;
&lt;br /&gt;
作业完成的原则：署你名字的工作必须是你个人的贡献。在完成作业的过程中，允许讨论，前提是讨论的所有参与者均处于同等完成度。但关键想法的执行、以及作业文本的写作必须独立完成，并在作业中致谢（acknowledge）所有参与讨论的人。符合规则的讨论与致谢将不会影响得分。不允许其他任何形式的合作——尤其是与已经完成作业的同学“讨论”。&lt;br /&gt;
&lt;br /&gt;
本课程将对剽窃行为采取零容忍的态度。在完成作业过程中，对他人工作（出版物、互联网资料、其他人的作业等）直接的文本抄袭和对关键思想、关键元素的抄袭，按照 [http://www.acm.org/publications/policies/plagiarism_policy ACM Policy on Plagiarism]的解释，都将视为剽窃。剽窃者成绩将被取消。如果发现互相抄袭行为，&amp;lt;font color=red&amp;gt; 抄袭和被抄袭双方的成绩都将被取消&amp;lt;/font&amp;gt;。因此请主动防止自己的作业被他人抄袭。&lt;br /&gt;
&lt;br /&gt;
学术诚信影响学生个人的品行，也关乎整个教育系统的正常运转。为了一点分数而做出学术不端的行为，不仅使自己沦为一个欺骗者，也使他人的诚实努力失去意义。让我们一起努力维护一个诚信的环境。&lt;br /&gt;
&lt;br /&gt;
= Assignments =&lt;br /&gt;
*[[概率论与数理统计 (Spring 2025)/Problem Set 1|Problem Set 1]]  请在 2025/3/10 上课之前(10am UTC+8)提交到 [mailto:pr2024_nju@163.com pr2024_nju@163.com] (文件名为&#039;&amp;lt;font color=red &amp;gt;学号_姓名_A1.pdf&amp;lt;/font&amp;gt;&#039;).&lt;br /&gt;
&lt;br /&gt;
= Lectures =&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2025/Intro.pdf 课程简介]&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2025/ProbSpace.pdf 概率空间]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 第1章&#039;&#039;&#039; 或 &#039;&#039;&#039;[GS] Chapter 1&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= Concepts =&lt;br /&gt;
* [https://plato.stanford.edu/entries/probability-interpret/ Interpretations of probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/History_of_probability History of probability]&lt;br /&gt;
* Example problems:&lt;br /&gt;
** [https://dornsifecms.usc.edu/assets/sites/520/docs/VonNeumann-ams12p36-38.pdf von Neumann&#039;s Bernoulli factory] and other [https://peteroupc.github.io/bernoulli.html Bernoulli factory algorithms]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Boy_or_Girl_paradox Boy or Girl paradox]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Monty_Hall_problem Monty Hall problem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bertrand_paradox_(probability) Bertrand paradox]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hard_spheres Hard spheres model] and [https://en.wikipedia.org/wiki/Ising_model Ising model]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/PageRank &#039;&#039;PageRank&#039;&#039;] and stationary [https://en.wikipedia.org/wiki/Random_walk random walk]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Diffusion_process Diffusion process] and [https://en.wikipedia.org/wiki/Diffusion_model diffusion model]&lt;br /&gt;
*[https://en.wikipedia.org/wiki/Probability_space Probability space]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Sample_space Sample space]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Event_(probability_theory) Event] and [https://en.wikipedia.org/wiki/Σ-algebra &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-algebra]&lt;br /&gt;
** Kolmogorov&#039;s [https://en.wikipedia.org/wiki/Probability_axioms axioms of probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Discrete_uniform_distribution Classical] and [https://en.wikipedia.org/wiki/Geometric_probability goemetric probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Boole%27s_inequality Union bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Inclusion%E2%80%93exclusion_principle Inclusion-Exclusion principle]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Boole%27s_inequality#Bonferroni_inequalities Bonferroni inequalities]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Conditional_probability Conditional probability]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chain_rule_(probability) Chain rule]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_total_probability Law of total probability]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bayes%27_theorem Bayes&#039; law]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Independence_(probability_theory) Independence] &lt;br /&gt;
** [https://en.wikipedia.org/wiki/Pairwise_independence Pairwise independence]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Random_variable Random variable]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cumulative_distribution_function Cumulative distribution function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_mass_function Probability mass function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_density_function Probability density function]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Multivariate_random_variable Random vector]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Joint_probability_distribution Joint probability distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Conditional_probability_distribution Conditional probability distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Marginal_distribution Marginal distribution]&lt;br /&gt;
* Some &#039;&#039;&#039;discrete&#039;&#039;&#039; probability distributions&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bernoulli_trial Bernoulli trial] and [https://en.wikipedia.org/wiki/Bernoulli_distribution Bernoulli distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Discrete_uniform_distribution Discrete uniform distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Binomial_distribution Binomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Geometric_distribution Geometric distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Negative_binomial_distribution Negative binomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hypergeometric_distribution Hypergeometric distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Poisson_distribution Poisson distribution]&lt;br /&gt;
** and [https://en.wikipedia.org/wiki/List_of_probability_distributions#Discrete_distributions others]&lt;br /&gt;
* Balls into bins model&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Multinomial_distribution Multinomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Birthday_problem Birthday problem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Coupon_collector%27s_problem Coupon collector]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Balls_into_bins_problem Occupancy problem]&lt;br /&gt;
* Random graphs&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Erd%C5%91s%E2%80%93R%C3%A9nyi_model Erdős–Rényi random graph model]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Galton%E2%80%93Watson_process Galton–Watson branching process]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Expected_value Expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_the_unconscious_statistician Law of the unconscious statistician, &#039;&#039;LOTUS&#039;&#039;]&lt;br /&gt;
** [https://dlsun.github.io/probability/linearity.html Linearity of expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Conditional_expectation Conditional expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_total_expectation Law of total expectation]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Markov%27s_inequality Markov&#039;s inequality]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Chebyshev%27s_inequality Chebyshev&#039;s inequality]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Moment_(mathematics) Moment] related&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Central_moment Central moment]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Variance Variance]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Standard_deviation Standard deviation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Covariance Covariance]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Correlation Correlation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Correlation_does_not_imply_causation Correlation does not imply causation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Skewness Skewness]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Kurtosis Kurtosis]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Moment_problem Moment problem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Measure_(mathematics) Measure]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Borel_set Borel set]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Lebesgue_integration Lebesgue integration]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Measurable_function Measurable function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cantor_set Cantor set]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Non-measurable_set Non-measurable set]&lt;br /&gt;
*[https://en.wikipedia.org/wiki/Probability_density_function Probability density function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_distribution#Absolutely_continuous_probability_distribution Continuous probability distribution]&lt;br /&gt;
**[https://en.wikipedia.org/wiki/Conditional_probability_distribution#Conditional_continuous_distributions Conditional continuous distributions]&lt;br /&gt;
**[https://en.wikipedia.org/wiki/Convolution_of_probability_distributions Convolution of probability distributions] and [https://en.wikipedia.org/wiki/List_of_convolutions_of_probability_distributions List of convolutions of probability distributions]&lt;br /&gt;
* Some &#039;&#039;&#039;continuous&#039;&#039;&#039; probability distributions&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Continuous_uniform_distribution Continuous uniform distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Exponential_distribution Exponential distribution] and [https://en.wikipedia.org/wiki/Poisson_point_process Poisson point process]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Normal_distribution Normal (Gaussion) distribution]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Gaussian_function Gaussian function] and [https://en.wikipedia.org/wiki/Gaussian_integral Gaussian integral]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Standard_normal_table Standard normal table]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/68%E2%80%9395%E2%80%9399.7_rule 68–95–99.7 rule]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Sum_of_normally_distributed_random_variables Sum of normally distributed random variables]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Multivariate_normal_distribution Multivariate normal distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chi-squared_distribution Chi-squared distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Gamma_distribution Gamma distribution] and [https://en.wikipedia.org/wiki/Gamma_function Gamma function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Beta_distribution Beta distribution] and [https://en.wikipedia.org/wiki/Beta_function Beta function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cauchy_distribution Cauchy distribution]&lt;br /&gt;
** and [https://en.wikipedia.org/wiki/List_of_probability_distributions#Absolutely_continuous_distributions others]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Inverse_transform_sampling Inverse transform sampling]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stochastic_dominance Stochastic dominance] and [https://en.wikipedia.org/wiki/Coupling_(probability) Coupling] &lt;br /&gt;
* [https://en.wikipedia.org/wiki/Moment-generating_function Moment-generating function]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Convergence_of_random_variables Convergence of random variables]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Pointwise_convergence Pointwise convergence]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Convergence_of_measures#Weak_convergence_of_measures Weak convergence of measures]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Convergence_in_measure Convergence in measure]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Skorokhod%27s_representation_theorem Skorokhod&#039;s representation theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Continuous_mapping_theorem Continuous mapping theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Borel%E2%80%93Cantelli_lemma Borel–Cantelli lemma] and [https://en.wikipedia.org/wiki/Zero%E2%80%93one_law zero-one laws]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Law_of_large_numbers Law of large numbers]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Central_limit_theorem Central limit theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/De_Moivre%E2%80%93Laplace_theorem De Moivre–Laplace theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Berry%E2%80%93Esseen_theorem Berry–Esseen theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Characteristic_function_(probability_theory) Characteristic function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Lévy%27s_continuity_theorem Lévy&#039;s continuity theorem]&lt;br /&gt;
* Concentration of measures:&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Moment-generating_function Moment-generating function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chernoff_bound Chernoff bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hoeffding%27s_inequality Hoeffding&#039;s bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Doob_martingale#McDiarmid&#039;s_inequality McDiarmid&#039;s inequality]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Doob_martingale Doob martingale] and [https://en.wikipedia.org/wiki/Azuma%27s_inequality Azuma&#039;s inequality]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Sub-Gaussian_distribution Sub-Gaussian distribution]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stochastic_process Stochastic process]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stopping_time Stopping time]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Martingale_(probability_theory) Martingale]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Optional_stopping_theorem Optional stopping theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Wald%27s_equation Wald&#039;s equation]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Markov_chain Markov chain]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Markov_property Markov property] and [https://en.wikipedia.org/wiki/Memorylessness memorylessness]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Stochastic_matrix Stochastic matrix]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Stationary_distribution Stationary distribution]&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_1&amp;diff=12903</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 1</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_1&amp;diff=12903"/>
		<updated>2025-02-19T13:16:56Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 1==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-2-principle-of-inclusion-and-exclusion&amp;quot;&amp;gt;Problem 1 (Principle of Inclusion and Exclusion)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Let &amp;lt;math&amp;gt;n\ge 1&amp;lt;/math&amp;gt; be a positive integer and &amp;lt;math&amp;gt;A_1,A_2,\ldots,A_n&amp;lt;/math&amp;gt; be &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; events.&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Union bound&amp;lt;/strong&amp;gt;] Prove &amp;lt;math&amp;gt;\mathbf{Pr}\left( \bigcup_{i=1}^{n} A_i \right) \le \sum_{i=1}^{n} \mathbf{Pr}\left(A_i\right)&amp;lt;/math&amp;gt;&amp;lt;strong&amp;gt; using the definition of probability space&amp;lt;/strong&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Principle of Inclusion and Exclusion (PIE)&amp;lt;/strong&amp;gt;] Prove that &amp;lt;math&amp;gt;\mathbf{Pr}\left( \bigcup_{i=1}^n A_i\right) = \sum_{\emptyset \neq S \subseteq [n]} (-1)^{|S|-1} \mathbf{Pr}\left( \bigcap_{i \in S} A_i \right)&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;[n]=\{1,2,\ldots,n\}&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Surjection&amp;lt;/strong&amp;gt;] For positive integers &amp;lt;math&amp;gt;m\ge n&amp;lt;/math&amp;gt;, prove that the probability of a uniform random function &amp;lt;math&amp;gt;f:[m]\to[n]&amp;lt;/math&amp;gt; to be surjective (满射) is &amp;lt;math&amp;gt;\sum_{k=1}^n(-1)^{n-k}{n\choose k}\left(\frac{k}{n}\right)^m&amp;lt;/math&amp;gt;. &amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Bonferroni&amp;amp;#39;s inequality and Kounias&amp;amp;#39; inequality&amp;lt;/strong&amp;gt;] Prove that &lt;br /&gt;
&amp;lt;math&amp;gt;&lt;br /&gt;
\sum_{i=1}^n \mathbf{Pr}(A_i) - \sum_{1 \le i&amp;lt; j \le n} \mathbf{Pr}(A_i \cap A_j)\le \mathbf{Pr}\left(\bigcup_{i=1}^n A_i\right) \le \sum_{i=1}^n \mathbf{Pr} \left( A_i\right) - \sum_{i=2}^n \mathbf{Pr}(A_1 \cap A_i).&lt;br /&gt;
&amp;lt;/math&amp;gt; &lt;br /&gt;
(Hint: This is sometimes called Kounias&#039; inequality which is weaker than the Bonferroni&#039;s inequality. You can try using Venn diagram to understand these inequalities.)&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-2-probability-space&amp;quot;&amp;gt;Problem 2 (Probability space)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Nonexistence of probability space&amp;lt;/strong&amp;gt;] Prove that it is impossible to define a uniform probability law on natural numbers &amp;lt;math&amp;gt;\mathbb{N}&amp;lt;/math&amp;gt;. More precisely, prove that there does not exist a probability space &amp;lt;math&amp;gt;(\mathbb{N},2^{\mathbb{N}},\mathbf{Pr})&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;\mathbf{Pr}(\{i\}) = \mathbf{Pr}(\{j\})&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;i, j \in \mathbb{N}&amp;lt;/math&amp;gt;. &lt;br /&gt;
Please explain why the same argument fails to prove that there is no uniform probability law on the real interval &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt;, that is, there is no such probability space &amp;lt;math&amp;gt;([0,1],\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt; that for any interval &amp;lt;math&amp;gt;(l,r] \subseteq [0,1]&amp;lt;/math&amp;gt;, it holds that &amp;lt;math&amp;gt;(l,r] \in \mathcal{F}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{Pr}( (l,r] ) = r-l&amp;lt;/math&amp;gt;. (Actually, such probability measure does exist and is called the Lebesgue measure on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt;). &amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Smallest &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-algebra (I)&amp;lt;/strong&amp;gt;] For any subset &amp;lt;math&amp;gt;S \subseteq 2^\Omega&amp;lt;/math&amp;gt;, prove that the smallest &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field containing &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; is given by &amp;lt;math&amp;gt;\sigma(S) := \bigcap_{\substack{S \subseteq \mathcal{F} \subseteq 2^\Omega\\ \mathcal{F} \text{ is a } \sigma\text{-field }} } \mathcal{F}&amp;lt;/math&amp;gt;. (Hint: You should show that it is indeed a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field and also it is the smallest one containing &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;.)&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Smallest &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field (II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;S,T \subseteq 2^{\Omega}&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\sigma(S) = \sigma(T)&amp;lt;/math&amp;gt; if and only if &amp;lt;math&amp;gt;S \subseteq \sigma(T)&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;T \subseteq \sigma(S)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Union of &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\mathcal{F}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathcal{G}&amp;lt;/math&amp;gt; be &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-fields of subsets of &amp;lt;math&amp;gt;\Omega&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathcal{F} \cup \mathcal{G}&amp;lt;/math&amp;gt; is not necessarily a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Projection&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\mathcal{F}&amp;lt;/math&amp;gt; be a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field of subsets of &amp;lt;math&amp;gt;\Omega&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;T \subseteq \Omega&amp;lt;/math&amp;gt; be a subset. Show that &amp;lt;math&amp;gt;\{S \cap T \mid S \in \mathcal{F}\}&amp;lt;/math&amp;gt; is a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Probability space?&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\Omega = \mathbb{R}&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathcal{F}&amp;lt;/math&amp;gt; is the set of all subsets &amp;lt;math&amp;gt;A \subseteq \Omega&amp;lt;/math&amp;gt; so that &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;\overline{A}&amp;lt;/math&amp;gt; (complement of &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;) is countable, &amp;lt;math&amp;gt;P(A) = 0&amp;lt;/math&amp;gt; in the first case and &amp;lt;math&amp;gt;P(A) = 1&amp;lt;/math&amp;gt; in the second. Is &amp;lt;math&amp;gt;(\Omega,\mathcal{F},P)&amp;lt;/math&amp;gt; a probability space? Please explain your answer.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-3-probability-space&amp;quot;&amp;gt;Problem 3 (Birthday paradox)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Please design a &amp;lt;strong&amp;gt;randomized algorithm using the birthday paradox&amp;lt;/strong&amp;gt; that solves the following problem in &amp;lt;math&amp;gt;\mathrm{poly}(n) \cdot 2^{n/2}&amp;lt;/math&amp;gt; time with high probability (for example, &amp;lt;math&amp;gt;0.99&amp;lt;/math&amp;gt; when &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; is sufficiently large). Please provide a detailed error analysis as well. (WARNING: You will &amp;lt;strong&amp;gt;NOT&amp;lt;/strong&amp;gt; receive any points if you solve this task using Gaussian elimination)&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Given an integer sequence &amp;lt;math&amp;gt;a_1,a_2,\ldots,a_{100n}&amp;lt;/math&amp;gt; of length &amp;lt;math&amp;gt;100 n&amp;lt;/math&amp;gt; satisfying &amp;lt;math&amp;gt;0 \le a_i &amp;lt; 2^n&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;1 \le i \le 100 n&amp;lt;/math&amp;gt;. Please find out a non-empty subset &amp;lt;math&amp;gt;S \subseteq \{1,2,\ldots,100n\}&amp;lt;/math&amp;gt; satisfying &amp;lt;math&amp;gt;\bigoplus_{i \in S} a_i = 0&amp;lt;/math&amp;gt;, i.e. the exclusive or of the elements whose indices are in &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; equals to &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-4-conditional-probability&amp;quot;&amp;gt;Problem 4 (Conditional probability)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Positive correlation&amp;lt;/strong&amp;gt;] We say that events &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; gives &#039;&#039;positive information&#039;&#039; about event &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; if &amp;lt;math&amp;gt;\mathbf{Pr}(A|B) &amp;gt; \mathbf{Pr}(A)&amp;lt;/math&amp;gt;, that is, the occurrence of &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; makes the occurrence of &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; more likely. Now suppose that &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; gives positive information about &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;.&lt;br /&gt;
# Does &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; give positive information about &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;?&lt;br /&gt;
# Does &amp;lt;math&amp;gt;\overline{B}&amp;lt;/math&amp;gt; give negative information about &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;, that is, is it true that &amp;lt;math&amp;gt;\mathbf{Pr}(A|\overline{B}) &amp;lt; \mathbf{Pr}(A)&amp;lt;/math&amp;gt;?&lt;br /&gt;
# Does &amp;lt;math&amp;gt;\overline{B}&amp;lt;/math&amp;gt; give positive information or negative information about &amp;lt;math&amp;gt;\overline{A}&amp;lt;/math&amp;gt;?&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Balls in urns (I)&amp;lt;/strong&amp;gt;] There are &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; urns of which the &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;-th contains &amp;lt;math&amp;gt;r-1&amp;lt;/math&amp;gt; white balls and &amp;lt;math&amp;gt;n-r&amp;lt;/math&amp;gt; black balls. You pick an urn uniformly at random (here, &amp;quot;uniformly&amp;quot; means that each urn has equal probability of being chosen) and remove two balls from that urn, uniformly at random without replacement (which means that each of the &amp;lt;math&amp;gt;{n-1\choose 2}&amp;lt;/math&amp;gt; pairs of balls are chosen to be removed with equal probability). Find the following probabilities:&lt;br /&gt;
# the second ball is black;&lt;br /&gt;
# the second ball is black, given that the first is black.&lt;br /&gt;
&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Balls in urns (II)&amp;lt;/strong&amp;gt;] Suppose that an urn contains &amp;lt;math&amp;gt;w&amp;lt;/math&amp;gt; white balls and &amp;lt;math&amp;gt;b&amp;lt;/math&amp;gt; black balls. The balls are drawn from the urn one by one, each time uniformly and independently at random, without replacement (which means we do not put the chosen ball back after each drawing). Find the probabilities of the events:&lt;br /&gt;
# the first white ball drawn is the &amp;lt;math&amp;gt;(k+1)&amp;lt;/math&amp;gt;th ball;&lt;br /&gt;
# the last ball drawn is white.&lt;br /&gt;
&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-5-independence&amp;quot;&amp;gt;Problem 5 (Independence)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Let&amp;amp;#39;s consider a series of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; outputs &amp;lt;math&amp;gt;(X_1, X_2, \cdots, X_n) \in \{0,1\}^n&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; independent Bernoulli trials, where each trial succeeds with the same probability  &amp;lt;math&amp;gt;0 &amp;lt; p &amp;lt; 1&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Limited independence&amp;lt;/strong&amp;gt;] Construct three events &amp;lt;math&amp;gt;A,B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; Bernoulli trials such that &amp;lt;math&amp;gt;A, B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; are pairwise independent but are not (mutually) independent. You need to prove that the constructed events &amp;lt;math&amp;gt;A, B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; satisfy this. (Hint: Consider the case where &amp;lt;math&amp;gt;n = 2&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;p = 1/2&amp;lt;/math&amp;gt;.)&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Product distribution&amp;lt;/strong&amp;gt;] Suppose someone has observed the output of the &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; trials, and she told you that precisely &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; trials succeeded for some &amp;lt;math&amp;gt;0&amp;lt; k&amp;lt; n&amp;lt;/math&amp;gt;. Now you want to predict the output of the &amp;lt;math&amp;gt;(n+1)&amp;lt;/math&amp;gt;-th trial while the parameter &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; of the Bernoulli trial is unknown. One way to estimate &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; is to find such &amp;lt;math&amp;gt;\hat{p}&amp;lt;/math&amp;gt; that makes the observed outcomes most probable, namely you need to solve &lt;br /&gt;
  &amp;lt;math&amp;gt;\arg \max_{\hat{p}\in(0,1)} \mathbf{Pr}_{\hat{p}} [k \text{ out of } n\text{ trials succeed}].&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Estimate &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; by solving the above optimization problem.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;If someone tells you exactly which &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; trials succeed (in addition to just telling you the number of successful trials, which is &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;), would it help you to estimate &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; more accurately? Why?&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-6-probabilistic-method&amp;quot;&amp;gt;Problem 6 (Probabilistic method)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Tournaments&amp;lt;/strong&amp;gt;] A [[wikipedia:Tournament_(graph_theory)|tournament]] can be interpreted as the outcome of a round-robin tournament in which every player faces every other player exactly once, and no draws occur. Given two players (vertices) &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;, we draw an&lt;br /&gt;
arrow from &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; if &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; beats &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; (and vice versa). We say a tournament has property &amp;lt;math&amp;gt;S_k&amp;lt;/math&amp;gt; if for every &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; players, there exists another player &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt; who defeats all of them. For example, a triangle &amp;lt;math&amp;gt;x\rightarrow y \rightarrow z \rightarrow x&amp;lt;/math&amp;gt; has property &amp;lt;math&amp;gt;S_1&amp;lt;/math&amp;gt; but not &amp;lt;math&amp;gt;S_2&amp;lt;/math&amp;gt;. Prove that if &amp;lt;math&amp;gt;\binom n k(1-2^{-k})^{n-k}&amp;lt;1&amp;lt;/math&amp;gt;, then there  is a tournament on &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; vertices that has the property &amp;lt;math&amp;gt;S_k&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_1&amp;diff=12902</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 1</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_1&amp;diff=12902"/>
		<updated>2025-02-19T13:09:02Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 1==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-2-principle-of-inclusion-and-exclusion&amp;quot;&amp;gt;Problem 1 (Principle of Inclusion and Exclusion)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Let &amp;lt;math&amp;gt;n\ge 1&amp;lt;/math&amp;gt; be a positive integer and &amp;lt;math&amp;gt;A_1,A_2,\ldots,A_n&amp;lt;/math&amp;gt; be &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; events.&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Union bound&amp;lt;/strong&amp;gt;] Prove &amp;lt;math&amp;gt;\mathbf{Pr}\left( \bigcup_{i=1}^{n} A_i \right) \le \sum_{i=1}^{n} \mathbf{Pr}\left(A_i\right)&amp;lt;/math&amp;gt;&amp;lt;strong&amp;gt; using the definition of probability space&amp;lt;/strong&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Principle of Inclusion and Exclusion (PIE)&amp;lt;/strong&amp;gt;] Prove that &amp;lt;math&amp;gt;\mathbf{Pr}\left( \bigcup_{i=1}^n A_i\right) = \sum_{\emptyset \neq S \subseteq [n]} (-1)^{|S|-1} \mathbf{Pr}\left( \bigcap_{i \in S} A_i \right)&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;[n]=\{1,2,\ldots,n\}&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Surjection&amp;lt;/strong&amp;gt;] For positive integers &amp;lt;math&amp;gt;m\ge n&amp;lt;/math&amp;gt;, prove that the probability of a uniform random function &amp;lt;math&amp;gt;f:[m]\to[n]&amp;lt;/math&amp;gt; to be surjective (满射) is &amp;lt;math&amp;gt;\sum_{k=1}^n(-1)^{n-k}{n\choose k}\left(\frac{k}{n}\right)^m&amp;lt;/math&amp;gt;. &amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Bonferroni&amp;amp;#39;s inequality and Kounias&amp;amp;#39; inequality&amp;lt;/strong&amp;gt;] Prove that &lt;br /&gt;
&amp;lt;math&amp;gt;&lt;br /&gt;
\sum_{i=1}^n \mathbf{Pr}(A_i) - \sum_{1 \le i&amp;lt; j \le n} \mathbf{Pr}(A_i \cap A_j)\le \mathbf{Pr}\left(\bigcup_{i=1}^n A_i\right) \le \sum_{i=1}^n \mathbf{Pr} \left( A_i\right) - \sum_{i=2}^n \mathbf{Pr}(A_1 \cap A_i).&lt;br /&gt;
&amp;lt;/math&amp;gt; &lt;br /&gt;
(Hint: This is sometimes called Kounias&#039; inequality which is weaker than the Bonferroni&#039;s inequality. You can try using Venn diagram to understand these inequalities.)&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-2-probability-space&amp;quot;&amp;gt;Problem 2 (Probability space)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Nonexistence of probability space&amp;lt;/strong&amp;gt;] Prove that it is impossible to define a uniform probability law on natural numbers &amp;lt;math&amp;gt;\mathbb{N}&amp;lt;/math&amp;gt;. More precisely, prove that there does not exist a probability space &amp;lt;math&amp;gt;(\mathbb{N},2^{\mathbb{N}},\mathbf{Pr})&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;\mathbf{Pr}(\{i\}) = \mathbf{Pr}(\{j\})&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;i, j \in \mathbb{N}&amp;lt;/math&amp;gt;. &lt;br /&gt;
Please explain why the same argument fails to prove that there is no uniform probability law on the real interval &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt;, that is, there is no such probability space &amp;lt;math&amp;gt;([0,1],\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt; that for any interval &amp;lt;math&amp;gt;(l,r] \subseteq [0,1]&amp;lt;/math&amp;gt;, it holds that &amp;lt;math&amp;gt;(l,r] \in \mathcal{F}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{Pr}( (l,r] ) = r-l&amp;lt;/math&amp;gt;. (Actually, such probability measure does exist and is called the Lebesgue measure on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt;). &amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Smallest &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-algebra (I)&amp;lt;/strong&amp;gt;] For any subset &amp;lt;math&amp;gt;S \subseteq 2^\Omega&amp;lt;/math&amp;gt;, prove that the smallest &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field containing &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; is given by &amp;lt;math&amp;gt;\sigma(S) := \bigcap_{\substack{S \subseteq \mathcal{F} \subseteq 2^\Omega\\ \mathcal{F} \text{ is a } \sigma\text{-field }} } \mathcal{F}&amp;lt;/math&amp;gt;. (Hint: You should show that it is indeed a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field and also it is the smallest one containing &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;.)&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Smallest &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field (II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;S,T \subseteq 2^{\Omega}&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\sigma(S) = \sigma(T)&amp;lt;/math&amp;gt; if and only if &amp;lt;math&amp;gt;S \subseteq \sigma(T)&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;T \subseteq \sigma(S)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Union of &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\mathcal{F}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathcal{G}&amp;lt;/math&amp;gt; be &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-fields of subsets of &amp;lt;math&amp;gt;\Omega&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathcal{F} \cup \mathcal{G}&amp;lt;/math&amp;gt; is not necessarily a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Projection&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\mathcal{F}&amp;lt;/math&amp;gt; be a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field of subsets of &amp;lt;math&amp;gt;\Omega&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;T \subseteq \Omega&amp;lt;/math&amp;gt; be a subset. Show that &amp;lt;math&amp;gt;\{S \cap T \mid S \in \mathcal{F}\}&amp;lt;/math&amp;gt; is a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Probability space?&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\Omega = \mathbb{R}&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathcal{F}&amp;lt;/math&amp;gt; is the set of all subsets &amp;lt;math&amp;gt;A \subseteq \Omega&amp;lt;/math&amp;gt; so that &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;\overline{A}&amp;lt;/math&amp;gt; (complement of &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;) is countable, &amp;lt;math&amp;gt;P(A) = 0&amp;lt;/math&amp;gt; in the first case and &amp;lt;math&amp;gt;P(A) = 1&amp;lt;/math&amp;gt; in the second. Is &amp;lt;math&amp;gt;(\Omega,\mathcal{F},P)&amp;lt;/math&amp;gt; a probability space? Please explain your answer.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-3-probability-space&amp;quot;&amp;gt;Problem 3 (Birthday paradox)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Please design a randomized algorithm using the birthday paradox that solves the following problem in &amp;lt;math&amp;gt;\mathrm{poly}(n) \cdot 2^{n/2}&amp;lt;/math&amp;gt; time with high probability (for example, &amp;lt;math&amp;gt;0.99&amp;lt;/math&amp;gt; when &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; is sufficiently large). Please provide a detailed error analysis as well.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Given an integer sequence &amp;lt;math&amp;gt;a_1,a_2,\ldots,a_{100n}&amp;lt;/math&amp;gt; of length &amp;lt;math&amp;gt;100 n&amp;lt;/math&amp;gt; satisfying &amp;lt;math&amp;gt;0 \le a_i &amp;lt; 2^n&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;1 \le i \le 100 n&amp;lt;/math&amp;gt;. Please find out a non-empty subset &amp;lt;math&amp;gt;S \subseteq \{1,2,\ldots,100n\}&amp;lt;/math&amp;gt; satisfying &amp;lt;math&amp;gt;\bigoplus_{i \in S} a_i = 0&amp;lt;/math&amp;gt;, i.e. the exclusive or of the elements whose indices are in &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; equals to &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-4-conditional-probability&amp;quot;&amp;gt;Problem 4 (Conditional probability)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Positive correlation&amp;lt;/strong&amp;gt;] We say that events &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; gives &#039;&#039;positive information&#039;&#039; about event &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; if &amp;lt;math&amp;gt;\mathbf{Pr}(A|B) &amp;gt; \mathbf{Pr}(A)&amp;lt;/math&amp;gt;, that is, the occurrence of &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; makes the occurrence of &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; more likely. Now suppose that &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; gives positive information about &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;.&lt;br /&gt;
# Does &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; give positive information about &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;?&lt;br /&gt;
# Does &amp;lt;math&amp;gt;\overline{B}&amp;lt;/math&amp;gt; give negative information about &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;, that is, is it true that &amp;lt;math&amp;gt;\mathbf{Pr}(A|\overline{B}) &amp;lt; \mathbf{Pr}(A)&amp;lt;/math&amp;gt;?&lt;br /&gt;
# Does &amp;lt;math&amp;gt;\overline{B}&amp;lt;/math&amp;gt; give positive information or negative information about &amp;lt;math&amp;gt;\overline{A}&amp;lt;/math&amp;gt;?&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Balls in urns (I)&amp;lt;/strong&amp;gt;] There are &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; urns of which the &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;-th contains &amp;lt;math&amp;gt;r-1&amp;lt;/math&amp;gt; white balls and &amp;lt;math&amp;gt;n-r&amp;lt;/math&amp;gt; black balls. You pick an urn uniformly at random (here, &amp;quot;uniformly&amp;quot; means that each urn has equal probability of being chosen) and remove two balls from that urn, uniformly at random without replacement (which means that each of the &amp;lt;math&amp;gt;{n-1\choose 2}&amp;lt;/math&amp;gt; pairs of balls are chosen to be removed with equal probability). Find the following probabilities:&lt;br /&gt;
# the second ball is black;&lt;br /&gt;
# the second ball is black, given that the first is black.&lt;br /&gt;
&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Balls in urns (II)&amp;lt;/strong&amp;gt;] Suppose that an urn contains &amp;lt;math&amp;gt;w&amp;lt;/math&amp;gt; white balls and &amp;lt;math&amp;gt;b&amp;lt;/math&amp;gt; black balls. The balls are drawn from the urn one by one, each time uniformly and independently at random, without replacement (which means we do not put the chosen ball back after each drawing). Find the probabilities of the events:&lt;br /&gt;
# the first white ball drawn is the &amp;lt;math&amp;gt;(k+1)&amp;lt;/math&amp;gt;th ball;&lt;br /&gt;
# the last ball drawn is white.&lt;br /&gt;
&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-5-independence&amp;quot;&amp;gt;Problem 5 (Independence)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Let&amp;amp;#39;s consider a series of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; outputs &amp;lt;math&amp;gt;(X_1, X_2, \cdots, X_n) \in \{0,1\}^n&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; independent Bernoulli trials, where each trial succeeds with the same probability  &amp;lt;math&amp;gt;0 &amp;lt; p &amp;lt; 1&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Limited independence&amp;lt;/strong&amp;gt;] Construct three events &amp;lt;math&amp;gt;A,B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; Bernoulli trials such that &amp;lt;math&amp;gt;A, B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; are pairwise independent but are not (mutually) independent. You need to prove that the constructed events &amp;lt;math&amp;gt;A, B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; satisfy this. (Hint: Consider the case where &amp;lt;math&amp;gt;n = 2&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;p = 1/2&amp;lt;/math&amp;gt;.)&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Product distribution&amp;lt;/strong&amp;gt;] Suppose someone has observed the output of the &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; trials, and she told you that precisely &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; trials succeeded for some &amp;lt;math&amp;gt;0&amp;lt; k&amp;lt; n&amp;lt;/math&amp;gt;. Now you want to predict the output of the &amp;lt;math&amp;gt;(n+1)&amp;lt;/math&amp;gt;-th trial while the parameter &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; of the Bernoulli trial is unknown. One way to estimate &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; is to find such &amp;lt;math&amp;gt;\hat{p}&amp;lt;/math&amp;gt; that makes the observed outcomes most probable, namely you need to solve &lt;br /&gt;
  &amp;lt;math&amp;gt;\arg \max_{\hat{p}\in(0,1)} \mathbf{Pr}_{\hat{p}} [k \text{ out of } n\text{ trials succeed}].&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Estimate &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; by solving the above optimization problem.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;If someone tells you exactly which &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; trials succeed (in addition to just telling you the number of successful trials, which is &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;), would it help you to estimate &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; more accurately? Why?&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-6-probabilistic-method&amp;quot;&amp;gt;Problem 6 (Probabilistic method)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Tournaments&amp;lt;/strong&amp;gt;] A [[wikipedia:Tournament_(graph_theory)|tournament]] can be interpreted as the outcome of a round-robin tournament in which every player faces every other player exactly once, and no draws occur. Given two players (vertices) &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;, we draw an&lt;br /&gt;
arrow from &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; if &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; beats &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; (and vice versa). We say a tournament has property &amp;lt;math&amp;gt;S_k&amp;lt;/math&amp;gt; if for every &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; players, there exists another player &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt; who defeats all of them. For example, a triangle &amp;lt;math&amp;gt;x\rightarrow y \rightarrow z \rightarrow x&amp;lt;/math&amp;gt; has property &amp;lt;math&amp;gt;S_1&amp;lt;/math&amp;gt; but not &amp;lt;math&amp;gt;S_2&amp;lt;/math&amp;gt;. Prove that if &amp;lt;math&amp;gt;\binom n k(1-2^{-k})^{n-k}&amp;lt;1&amp;lt;/math&amp;gt;, then there  is a tournament on &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; vertices that has the property &amp;lt;math&amp;gt;S_k&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_1&amp;diff=12890</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 1</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_1&amp;diff=12890"/>
		<updated>2025-02-19T12:04:53Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: Created page with &amp;quot;*每道题目的解答都要有完整的解题过程，中英文不限。  *我们推荐大家使用LaTeX, markdown等对作业进行排版。  == Assumption throughout Problem Set 1== &amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;  &amp;lt;h2 id=&amp;quot;problem-2-principle-of-inclusion-and-exclusion&amp;quot;&amp;gt;Problem 1 (Principle of Inclusion and Exclusion)&amp;lt;/h2&amp;gt; &amp;lt;p&amp;gt;Let &amp;lt;math&amp;gt;n\ge 1&amp;lt;/math&amp;gt; be a positive integer and &amp;lt;math&amp;gt;A_1,A_...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 1==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-2-principle-of-inclusion-and-exclusion&amp;quot;&amp;gt;Problem 1 (Principle of Inclusion and Exclusion)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Let &amp;lt;math&amp;gt;n\ge 1&amp;lt;/math&amp;gt; be a positive integer and &amp;lt;math&amp;gt;A_1,A_2,\ldots,A_n&amp;lt;/math&amp;gt; be &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; events.&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Union bound&amp;lt;/strong&amp;gt;] Prove &amp;lt;math&amp;gt;\mathbf{Pr}\left( \bigcup_{i=1}^{n} A_i \right) \le \sum_{i=1}^{n} \mathbf{Pr}\left(A_i\right)&amp;lt;/math&amp;gt;&amp;lt;strong&amp;gt; using the definition of probability space&amp;lt;/strong&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Principle of Inclusion and Exclusion (PIE)&amp;lt;/strong&amp;gt;] Prove that &amp;lt;math&amp;gt;\mathbf{Pr}\left( \bigcup_{i=1}^n A_i\right) = \sum_{\emptyset \neq S \subseteq [n]} (-1)^{|S|-1} \mathbf{Pr}\left( \bigcap_{i \in S} A_i \right)&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;[n]=\{1,2,\ldots,n\}&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Surjection&amp;lt;/strong&amp;gt;] For positive integers &amp;lt;math&amp;gt;m\ge n&amp;lt;/math&amp;gt;, prove that the probability of a uniform random function &amp;lt;math&amp;gt;f:[m]\to[n]&amp;lt;/math&amp;gt; to be surjective (满射) is &amp;lt;math&amp;gt;\sum_{k=1}^n(-1)^{n-k}{n\choose k}\left(\frac{k}{n}\right)^m&amp;lt;/math&amp;gt;. &amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Coprime integers&amp;lt;/strong&amp;gt;] Given positive integers &amp;lt;math&amp;gt;n \ge 2&amp;lt;/math&amp;gt;, calculate the number of integer pairs &amp;lt;math&amp;gt;(x,y)&amp;lt;/math&amp;gt; satisfying &amp;lt;math&amp;gt;1 \le x &amp;lt; y \le n&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathrm{gcd}(x,y) = 1&amp;lt;/math&amp;gt;. &lt;br /&gt;
(Remark: It suggests the probability that two randomly chosen integers from &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; are coprime tends to &amp;lt;math&amp;gt;\frac{6}{\pi^2}&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Bonferroni&amp;amp;#39;s inequality and Kounias&amp;amp;#39; inequality&amp;lt;/strong&amp;gt;] Prove that &lt;br /&gt;
&amp;lt;math&amp;gt;&lt;br /&gt;
\sum_{i=1}^n \mathbf{Pr}(A_i) - \sum_{1 \le i&amp;lt; j \le n} \mathbf{Pr}(A_i \cap A_j)\le \mathbf{Pr}\left(\bigcup_{i=1}^n A_i\right) \le \sum_{i=1}^n \mathbf{Pr} \left( A_i\right) - \sum_{i=2}^n \mathbf{Pr}(A_1 \cap A_i).&lt;br /&gt;
&amp;lt;/math&amp;gt; &lt;br /&gt;
(Hint: This is sometimes called Kounias&#039; inequality which is weaker than the Bonferroni&#039;s inequality. You can try using Venn diagram to understand these inequalities.)&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-1-classic-examples&amp;quot;&amp;gt;Problem 2 (Symmetric 1D random walk)&amp;lt;/h2&amp;gt;&lt;br /&gt;
A gambler plays a fair gambling game: At each round, he flips a fair coin, earns &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; point if it&#039;s HEADs, and loses &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; point if otherwise.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Symmetric 1D random walk (I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;A_i&amp;lt;/math&amp;gt; be the event that the gambler earns &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; points after playing &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; rounds of the game, that is, the number of times the coin lands on heads is equal to the number of times it lands on tails. Compute &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Pr}(A_i)&amp;lt;/math&amp;gt;. (Hint: You may use Stirling&#039;s approximation to estimate [math]\mathbf{Pr}(A_i)[/math] and derive that &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Pr}(A_i) = +\infty&amp;lt;/math&amp;gt;.)&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Symmetric 1D random walk (II)&amp;lt;/strong&amp;gt;] Suppose that the game ends upon that the gambler loses all his &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt; points. Let &amp;lt;math&amp;gt;B_i&amp;lt;/math&amp;gt; be the event that the game ends within &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; rounds. Compute &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Pr}(\overline{B_i})&amp;lt;/math&amp;gt;. &lt;br /&gt;
(Hint: You may first consider [math] \displaystyle{m = 1}[/math] case. Let [math]\displaystyle{C_i}[/math] be the event that the game ends at the [math]\displaystyle{i}[/math]-th round. (i) Prove that &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Pr}(\overline{B_i}) = \sum_{i=1}^{+\infty} (i-1) \mathbf{Pr}(C_i)&amp;lt;/math&amp;gt;. (ii) Compute &amp;lt;math&amp;gt;\mathbf{Pr}(C_i)&amp;lt;/math&amp;gt;, which is a special case of the ballot problem.  (iii) Finally, use Stirling&#039;s approximation to derive that &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Pr}(\overline{B_i}) = +\infty&amp;lt;/math&amp;gt;.)&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-3-probability-space&amp;quot;&amp;gt;Problem 3 (Probability space)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Nonexistence of probability space&amp;lt;/strong&amp;gt;] Prove that it is impossible to define a uniform probability law on natural numbers &amp;lt;math&amp;gt;\mathbb{N}&amp;lt;/math&amp;gt;. More precisely, prove that there does not exist a probability space &amp;lt;math&amp;gt;(\mathbb{N},2^{\mathbb{N}},\mathbf{Pr})&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;\mathbf{Pr}(\{i\}) = \mathbf{Pr}(\{j\})&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;i, j \in \mathbb{N}&amp;lt;/math&amp;gt;. &lt;br /&gt;
Please explain why the same argument fails to prove that there is no uniform probability law on the real interval &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt;, that is, there is no such probability space &amp;lt;math&amp;gt;([0,1],\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt; that for any interval &amp;lt;math&amp;gt;(l,r] \subseteq [0,1]&amp;lt;/math&amp;gt;, it holds that &amp;lt;math&amp;gt;(l,r] \in \mathcal{F}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{Pr}( (l,r] ) = r-l&amp;lt;/math&amp;gt;. (Actually, such probability measure does exist and is called the Lebesgue measure on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt;). &amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Smallest &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field&amp;lt;/strong&amp;gt;] For any subset &amp;lt;math&amp;gt;S \subseteq 2^\Omega&amp;lt;/math&amp;gt;, prove that the smallest &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field containing &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; is given by &amp;lt;math&amp;gt;\bigcap_{\substack{S \subseteq \mathcal{F} \subseteq 2^\Omega\\ \mathcal{F} \text{ is a } \sigma\text{-field }} } \mathcal{F}&amp;lt;/math&amp;gt;. (Hint: You should show that it is indeed a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field and also it is the smallest one containing &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;.)&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Union of &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\mathcal{F}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathcal{G}&amp;lt;/math&amp;gt; be &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-fields of subsets of &amp;lt;math&amp;gt;\Omega&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathcal{F} \cup \mathcal{G}&amp;lt;/math&amp;gt; is not necessarily a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Probability space?&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\Omega = \mathbb{R}&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathcal{F}&amp;lt;/math&amp;gt; is the set of all subsets &amp;lt;math&amp;gt;A \subseteq \Omega&amp;lt;/math&amp;gt; so that &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;\overline{A}&amp;lt;/math&amp;gt; (complement of &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;) is countable, &amp;lt;math&amp;gt;P(A) = 0&amp;lt;/math&amp;gt; in the first case and &amp;lt;math&amp;gt;P(A) = 1&amp;lt;/math&amp;gt; in the second. Is &amp;lt;math&amp;gt;(\Omega,\mathcal{F},P)&amp;lt;/math&amp;gt; a probability space? Please explain your answer.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;&amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field?&amp;lt;/strong&amp;gt;] A set &amp;lt;math&amp;gt;A \subseteq \mathbb{N}&amp;lt;/math&amp;gt; is said to have asymptotic density &amp;lt;math&amp;gt;\theta&amp;lt;/math&amp;gt; if &amp;lt;math&amp;gt;\lim_{n \to \infty} |A \cap \{1,2,\ldots,n\}| / n = \theta&amp;lt;/math&amp;gt;. Let &amp;lt;math&amp;gt;\mathcal{A}&amp;lt;/math&amp;gt; be the collection of sets for which the asymptotic density exists. Is &amp;lt;math&amp;gt;\mathcal{A}&amp;lt;/math&amp;gt; a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-algebra? Please explain your answer.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-4-conditional-probability&amp;quot;&amp;gt;Problem 4 (Conditional probability)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Positive correlation&amp;lt;/strong&amp;gt;] We say that events &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; gives &#039;&#039;positive information&#039;&#039; about event &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; if &amp;lt;math&amp;gt;\mathbf{Pr}(A|B) &amp;gt; \mathbf{Pr}(A)&amp;lt;/math&amp;gt;, that is, the occurrence of &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; makes the occurrence of &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; more likely. Now suppose that &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; gives positive information about &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;.&lt;br /&gt;
# Does &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; give positive information about &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;?&lt;br /&gt;
# Does &amp;lt;math&amp;gt;\overline{B}&amp;lt;/math&amp;gt; give negative information about &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;, that is, is it true that &amp;lt;math&amp;gt;\mathbf{Pr}(A|\overline{B}) &amp;lt; \mathbf{Pr}(A)&amp;lt;/math&amp;gt;?&lt;br /&gt;
# Does &amp;lt;math&amp;gt;\overline{B}&amp;lt;/math&amp;gt; give positive information or negative information about &amp;lt;math&amp;gt;\overline{A}&amp;lt;/math&amp;gt;?&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Balls in urns (I)&amp;lt;/strong&amp;gt;] There are &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; urns of which the &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;-th contains &amp;lt;math&amp;gt;r-1&amp;lt;/math&amp;gt; white balls and &amp;lt;math&amp;gt;n-r&amp;lt;/math&amp;gt; black balls. You pick an urn uniformly at random (here, &amp;quot;uniformly&amp;quot; means that each urn has equal probability of being chosen) and remove two balls from that urn, uniformly at random without replacement (which means that each of the &amp;lt;math&amp;gt;{n-1\choose 2}&amp;lt;/math&amp;gt; pairs of balls are chosen to be removed with equal probability). Find the following probabilities:&lt;br /&gt;
# the second ball is black;&lt;br /&gt;
# the second ball is black, given that the first is black.&lt;br /&gt;
&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Balls in urns (II)&amp;lt;/strong&amp;gt;] Suppose that an urn contains &amp;lt;math&amp;gt;w&amp;lt;/math&amp;gt; white balls and &amp;lt;math&amp;gt;b&amp;lt;/math&amp;gt; black balls. The balls are drawn from the urn one by one, each time uniformly and independently at random, without replacement (which means we do not put the chosen ball back after each drawing). Find the probabilities of the events:&lt;br /&gt;
# the first white ball drawn is the &amp;lt;math&amp;gt;(k+1)&amp;lt;/math&amp;gt;th ball;&lt;br /&gt;
# the last ball drawn is white.&lt;br /&gt;
&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&#039;&#039;&#039;Noisy channel&#039;&#039;&#039;] When coded messages are sent, there are sometimes errors in transmission. In particular, Morse code uses &amp;quot;dots&amp;quot; and &amp;quot;dashes&amp;quot;, which are known to occur in the proportion of 3:4. In other words, the probability of a &amp;quot;dot&amp;quot; being sent is &amp;lt;math&amp;gt;3/7&amp;lt;/math&amp;gt;, and the probability of a &amp;quot;dash&amp;quot; being sent is &amp;lt;math&amp;gt;4/7&amp;lt;/math&amp;gt;. Suppose due to interference on the transmission channel, with probability &amp;lt;math&amp;gt;1/8&amp;lt;/math&amp;gt;, a dot will be mistakenly received as a dash, and vice versa. If we receive a dot, what is the probability that the transmitted symbol is indeed a dot? &lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-5-independence&amp;quot;&amp;gt;Problem 5 (Independence)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Let&amp;amp;#39;s consider a series of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; outputs &amp;lt;math&amp;gt;(X_1, X_2, \cdots, X_n) \in \{0,1\}^n&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; independent Bernoulli trials, where each trial succeeds with the same probability  &amp;lt;math&amp;gt;0 &amp;lt; p &amp;lt; 1&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Limited independence&amp;lt;/strong&amp;gt;] Construct three events &amp;lt;math&amp;gt;A,B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; Bernoulli trials such that &amp;lt;math&amp;gt;A, B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; are pairwise independent but are not (mutually) independent. You need to prove that the constructed events &amp;lt;math&amp;gt;A, B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; satisfy this. (Hint: Consider the case where &amp;lt;math&amp;gt;n = 2&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;p = 1/2&amp;lt;/math&amp;gt;.)&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Product distribution&amp;lt;/strong&amp;gt;] Suppose someone has observed the output of the &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; trials, and she told you that precisely &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; trials succeeded for some &amp;lt;math&amp;gt;0&amp;lt; k&amp;lt; n&amp;lt;/math&amp;gt;. Now you want to predict the output of the &amp;lt;math&amp;gt;(n+1)&amp;lt;/math&amp;gt;-th trial while the parameter &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; of the Bernoulli trial is unknown. One way to estimate &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; is to find such &amp;lt;math&amp;gt;\hat{p}&amp;lt;/math&amp;gt; that makes the observed outcomes most probable, namely you need to solve &lt;br /&gt;
  &amp;lt;math&amp;gt;\arg \max_{\hat{p}\in(0,1)} \mathbf{Pr}_{\hat{p}} [k \text{ out of } n\text{ trials succeed}].&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Estimate &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; by solving the above optimization problem.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;If someone tells you exactly which &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; trials succeed (in addition to just telling you the number of successful trials, which is &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;), would it help you to estimate &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; more accurately? Why?&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-6-probabilistic-method&amp;quot;&amp;gt;Problem 6 (Probabilistic method)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;A &#039;&#039;&#039;CNF formula&#039;&#039;&#039; （&#039;&#039;&#039;合取范式&#039;&#039;&#039;） &amp;lt;math&amp;gt;\Phi&amp;lt;/math&amp;gt; over &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; Boolean variables &amp;lt;math&amp;gt;x_1,\cdots, x_n&amp;lt;/math&amp;gt; is a &#039;&#039;&#039;conjunction&#039;&#039;&#039; (&amp;lt;math&amp;gt;\land&amp;lt;/math&amp;gt;) of &#039;&#039;&#039;clauses&#039;&#039;&#039; (&#039;&#039;&#039;子句&#039;&#039;&#039;) &amp;lt;math&amp;gt;\Phi=C_1\land C_2\land\cdots\land C_m&amp;lt;/math&amp;gt;, where each clause &amp;lt;math&amp;gt;C_j=\ell_{j_1}\lor\ell_{j_2}\cdots\lor\ell_{j_k}&amp;lt;/math&amp;gt; is a &#039;&#039;&#039;disjunction&#039;&#039;&#039; (&amp;lt;math&amp;gt;\lor&amp;lt;/math&amp;gt;) of &#039;&#039;&#039;literals&#039;&#039;&#039; (&#039;&#039;&#039;文字&#039;&#039;&#039;), where a literal &amp;lt;math&amp;gt;\ell_r&amp;lt;/math&amp;gt; is either a variable &amp;lt;math&amp;gt;x_i&amp;lt;/math&amp;gt; or the negation &amp;lt;math&amp;gt;\bar{x}_i&amp;lt;/math&amp;gt; of a variable. A CNF formula is &#039;&#039;&#039;satisfiable&#039;&#039;&#039; (&#039;&#039;&#039;可满足&#039;&#039;&#039;) if there is a truth assignment &amp;lt;math&amp;gt;x=(x_1,\cdots, x_n)\in \{\mathtt{true},\mathtt{false}\}^n&amp;lt;/math&amp;gt; to the variables such that &amp;lt;math&amp;gt;\Phi(x)=\mathtt{true}&amp;lt;/math&amp;gt;. A &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;-CNF formula is a CNF formula in which each clause contains exactly &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; literals (without repetition).&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Satisfiability (I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\Phi&amp;lt;/math&amp;gt; be a &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;-CNF with less than &amp;lt;math&amp;gt;2^k&amp;lt;/math&amp;gt; clauses. Use the probabilistic method to show that &amp;lt;math&amp;gt;\Phi&amp;lt;/math&amp;gt; must be satisfiable. You should be explicit about the probability space that is used. &amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Satisfiability (II)&amp;lt;/strong&amp;gt;] Give a constructive proof of the same problem above. That is, prove that &amp;lt;math&amp;gt;\Phi&amp;lt;/math&amp;gt; is satisfiable by showing how to construct a truth assignment  &amp;lt;math&amp;gt;x=(x_1,\cdots, x_n)\in \{\mathtt{true},\mathtt{false}\}^n&amp;lt;/math&amp;gt; such that  &amp;lt;math&amp;gt;\Phi(x)=\mathtt{true}&amp;lt;/math&amp;gt;. Your construction does NOT have to be efficient. Please explain the difference between this constructive proof and the proof by the probabilistic method.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Satisfiability (III)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\Phi&amp;lt;/math&amp;gt; be a &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;-CNF with &amp;lt;math&amp;gt;m\geq 2^k&amp;lt;/math&amp;gt; clauses. Use the probabilistic method to show that there exists a truth assignment &amp;lt;math&amp;gt;x=(x_1,\cdots, x_n)\in \{\mathtt{true},\mathtt{false}\}^n&amp;lt;/math&amp;gt; satisfying at least &amp;lt;math&amp;gt;\lfloor m(1-1/2^k) \rfloor&amp;lt;/math&amp;gt; clauses in &amp;lt;math&amp;gt;\Phi&amp;lt;/math&amp;gt;. (Hint: Consider overlaps of events in Venn diagram.) You should be explicit about the probability space that is used.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)&amp;diff=12865</id>
		<title>概率论与数理统计 (Spring 2025)</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)&amp;diff=12865"/>
		<updated>2025-02-17T06:26:53Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Infobox&lt;br /&gt;
|name         = Infobox&lt;br /&gt;
|bodystyle    = &lt;br /&gt;
|title        = &amp;lt;font size=3&amp;gt;&#039;&#039;&#039;概率论与数理统计&#039;&#039;&#039;&amp;lt;br&amp;gt;&lt;br /&gt;
&#039;&#039;&#039;Probability Theory&#039;&#039;&#039; &amp;lt;br&amp;gt; &amp;amp; &#039;&#039;&#039;Mathematical Statistics&#039;&#039;&#039;&amp;lt;/font&amp;gt;&lt;br /&gt;
|titlestyle   = &lt;br /&gt;
&lt;br /&gt;
|image        = &lt;br /&gt;
|imagestyle   = &lt;br /&gt;
|caption      = &lt;br /&gt;
|captionstyle = &lt;br /&gt;
|headerstyle  = background:#ccf;&lt;br /&gt;
|labelstyle   = background:#ddf;&lt;br /&gt;
|datastyle    = &lt;br /&gt;
&lt;br /&gt;
|header1 =Instructor&lt;br /&gt;
|label1  = &lt;br /&gt;
|data1   = &lt;br /&gt;
|header2 = &lt;br /&gt;
|label2  = &lt;br /&gt;
|data2   = &#039;&#039;&#039;尹一通&#039;&#039;&#039;&lt;br /&gt;
|header3 = &lt;br /&gt;
|label3  = Email&lt;br /&gt;
|data3   = yinyt@nju.edu.cn  &lt;br /&gt;
|header4 =&lt;br /&gt;
|label4  = office&lt;br /&gt;
|data4   = 计算机系 804&lt;br /&gt;
|header5 = &lt;br /&gt;
|label5  = &lt;br /&gt;
|data5   = &#039;&#039;&#039;刘景铖&#039;&#039;&#039;&lt;br /&gt;
|header6 = &lt;br /&gt;
|label6  = Email&lt;br /&gt;
|data6   = liu@nju.edu.cn  &lt;br /&gt;
|header7 =&lt;br /&gt;
|label7  = office&lt;br /&gt;
|data7   = 计算机系 516&lt;br /&gt;
|header8 = Class&lt;br /&gt;
|label8  = &lt;br /&gt;
|data8   = &lt;br /&gt;
|header9 =&lt;br /&gt;
|label9  = Class meeting&lt;br /&gt;
|data9   = Monday, 10am-12pm&amp;lt;br&amp;gt;&lt;br /&gt;
Wednesday (单), 10am-12pm&amp;lt;br&amp;gt;&lt;br /&gt;
仙Ⅰ-204&lt;br /&gt;
|header10=&lt;br /&gt;
|label10 = Office hour&lt;br /&gt;
|data10  = Friday, 2pm-3pm &amp;lt;br&amp;gt;计算机系 804（尹一通）&amp;lt;br&amp;gt;计算机系 516（刘景铖）&lt;br /&gt;
|header11= Textbook&lt;br /&gt;
|label11 = &lt;br /&gt;
|data11  = &lt;br /&gt;
|header12=&lt;br /&gt;
|label12 = &lt;br /&gt;
|data12  = [[File:概率导论.jpeg|border|100px]]&lt;br /&gt;
|header13=&lt;br /&gt;
|label13 = &lt;br /&gt;
|data13  = &#039;&#039;&#039;概率导论&#039;&#039;&#039;（第2版·修订版）&amp;lt;br&amp;gt; Dimitri P. Bertsekas and John N. Tsitsiklis&amp;lt;br&amp;gt; 郑忠国 童行伟 译；人民邮电出版社 (2022)&lt;br /&gt;
|header14=&lt;br /&gt;
|label14 = &lt;br /&gt;
|data14  = [[File:Grimmett_probability.jpg|border|100px]]&lt;br /&gt;
|header15=&lt;br /&gt;
|label15 = &lt;br /&gt;
|data15  = &#039;&#039;&#039;Probability and Random Processes&#039;&#039;&#039; (4E) &amp;lt;br&amp;gt; Geoffrey Grimmett and David Stirzaker &amp;lt;br&amp;gt;  Oxford University Press (2020)&lt;br /&gt;
|header16=&lt;br /&gt;
|label16 = &lt;br /&gt;
|data16  = [[File:Probability_and_Computing_2ed.jpg|border|100px]]&lt;br /&gt;
|header17=&lt;br /&gt;
|label17 = &lt;br /&gt;
|data17  = &#039;&#039;&#039;Probability and Computing&#039;&#039;&#039; (2E) &amp;lt;br&amp;gt; Michael Mitzenmacher and Eli Upfal &amp;lt;br&amp;gt;   Cambridge University Press (2017)&lt;br /&gt;
|belowstyle = background:#ddf;&lt;br /&gt;
|below = &lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
This is the webpage for the &#039;&#039;Probability Theory and Mathematical Statistics&#039;&#039; (概率论与数理统计) class of Spring 2025. Students who take this class should check this page periodically for content updates and new announcements. &lt;br /&gt;
&lt;br /&gt;
= Announcement =&lt;br /&gt;
* TBA&lt;br /&gt;
&lt;br /&gt;
= Course info =&lt;br /&gt;
* &#039;&#039;&#039;Instructor &#039;&#039;&#039;: &lt;br /&gt;
:* [http://tcs.nju.edu.cn/yinyt/ 尹一通]：[mailto:yinyt@nju.edu.cn &amp;lt;yinyt@nju.edu.cn&amp;gt;]，计算机系 804 &lt;br /&gt;
:* [https://liuexp.github.io 刘景铖]：[mailto:liu@nju.edu.cn &amp;lt;liu@nju.edu.cn&amp;gt;]，计算机系 516 &lt;br /&gt;
* &#039;&#039;&#039;Teaching assistant&#039;&#039;&#039;:&lt;br /&gt;
** [https://sites.google.com/view/xinyuanzhang 张昕渊]：[mailto:zhangxy@smail.nju.edu.cn &amp;lt;zhangxy@smail.nju.edu.cn&amp;gt;]，计算机系 410&lt;br /&gt;
** 邹宗瑞：[mailto:zou.zongrui@smail.nju.edu.cn &amp;lt;zou.zongrui@smail.nju.edu.cn&amp;gt;]，计算机系 410&lt;br /&gt;
* &#039;&#039;&#039;Class meeting&#039;&#039;&#039;:&lt;br /&gt;
** 周一：10am-12pm，仙Ⅰ-204&lt;br /&gt;
** 周三（单）：10am-12pm，仙Ⅰ-204&lt;br /&gt;
* &#039;&#039;&#039;Office hour&#039;&#039;&#039;: &lt;br /&gt;
:* 周五 2pm-3pm, 计算机系 804（尹一通）&lt;br /&gt;
:* 周五 2pm-3pm, 计算机系 516（刘景铖）&lt;br /&gt;
:* &#039;&#039;&#039;QQ群&#039;&#039;&#039;: 974742320（申请加入需提供姓名、院系、学号）&lt;br /&gt;
&lt;br /&gt;
= Syllabus =&lt;br /&gt;
课程内容分为三大部分：&lt;br /&gt;
* &#039;&#039;&#039;经典概率论&#039;&#039;&#039;：概率空间、随机变量及其数字特征、多维与连续随机变量、极限定理等内容&lt;br /&gt;
* &#039;&#039;&#039;概率与计算&#039;&#039;&#039;：测度集中现象 (concentration of measure)、概率法 (the probabilistic method)、离散随机过程的相关专题&lt;br /&gt;
* &#039;&#039;&#039;数理统计&#039;&#039;&#039;：参数估计、假设检验、贝叶斯估计、线性回归等统计推断等概念&lt;br /&gt;
&lt;br /&gt;
对于第一和第二部分，要求清楚掌握基本概念，深刻理解关键的现象与规律以及背后的原理，并可以灵活运用所学方法求解相关问题。对于第三部分，要求熟悉数理统计的若干基本概念，以及典型的统计模型和统计推断问题。&lt;br /&gt;
&lt;br /&gt;
经过本课程的训练，力求使学生能够熟悉掌握概率的语言，并会利用概率思维来理解客观世界并对其建模，以及驾驭概率的数学工具来分析和求解专业问题。&lt;br /&gt;
&lt;br /&gt;
=== 教材与参考书 Course Materials ===&lt;br /&gt;
* &#039;&#039;&#039;[BT]&#039;&#039;&#039; 概率导论（第2版·修订版），[美]伯特瑟卡斯（Dimitri P.Bertsekas）[美]齐齐克利斯（John N.Tsitsiklis）著，郑忠国 童行伟 译，人民邮电出版社（2022）。&lt;br /&gt;
* &#039;&#039;&#039;[GS]&#039;&#039;&#039; &#039;&#039;Probability and Random Processes&#039;&#039;, by Geoffrey Grimmett and David Stirzaker; Oxford University Press; 4th edition (2020).&lt;br /&gt;
* &#039;&#039;&#039;[MU]&#039;&#039;&#039; &#039;&#039;Probability and Computing: Randomization and Probabilistic Techniques in Algorithms and Data Analysis&#039;&#039;, by Michael Mitzenmacher, Eli Upfal; Cambridge University Press; 2nd edition (2017).&lt;br /&gt;
&lt;br /&gt;
=== 成绩 Grading Policy ===&lt;br /&gt;
* 课程成绩：本课程将会有若干次作业和一次期末考试。最终成绩将由平时作业成绩和期末考试成绩综合得出。&lt;br /&gt;
* 迟交：如果有特殊的理由，无法按时完成作业，请提前联系授课老师，给出正当理由。否则迟交的作业将不被接受。&lt;br /&gt;
&lt;br /&gt;
=== &amp;lt;font color=red&amp;gt; 学术诚信 Academic Integrity &amp;lt;/font&amp;gt;===&lt;br /&gt;
学术诚信是所有从事学术活动的学生和学者最基本的职业道德底线，本课程将不遗余力的维护学术诚信规范，违反这一底线的行为将不会被容忍。&lt;br /&gt;
&lt;br /&gt;
作业完成的原则：署你名字的工作必须是你个人的贡献。在完成作业的过程中，允许讨论，前提是讨论的所有参与者均处于同等完成度。但关键想法的执行、以及作业文本的写作必须独立完成，并在作业中致谢（acknowledge）所有参与讨论的人。符合规则的讨论与致谢将不会影响得分。不允许其他任何形式的合作——尤其是与已经完成作业的同学“讨论”。&lt;br /&gt;
&lt;br /&gt;
本课程将对剽窃行为采取零容忍的态度。在完成作业过程中，对他人工作（出版物、互联网资料、其他人的作业等）直接的文本抄袭和对关键思想、关键元素的抄袭，按照 [http://www.acm.org/publications/policies/plagiarism_policy ACM Policy on Plagiarism]的解释，都将视为剽窃。剽窃者成绩将被取消。如果发现互相抄袭行为，&amp;lt;font color=red&amp;gt; 抄袭和被抄袭双方的成绩都将被取消&amp;lt;/font&amp;gt;。因此请主动防止自己的作业被他人抄袭。&lt;br /&gt;
&lt;br /&gt;
学术诚信影响学生个人的品行，也关乎整个教育系统的正常运转。为了一点分数而做出学术不端的行为，不仅使自己沦为一个欺骗者，也使他人的诚实努力失去意义。让我们一起努力维护一个诚信的环境。&lt;br /&gt;
&lt;br /&gt;
= Assignments =&lt;br /&gt;
TBA&lt;br /&gt;
&lt;br /&gt;
= Lectures =&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2025/Intro.pdf 课程简介]&lt;br /&gt;
&lt;br /&gt;
= Concepts =&lt;br /&gt;
* [https://plato.stanford.edu/entries/probability-interpret/ Interpretations of probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/History_of_probability History of probability]&lt;br /&gt;
* Example problems:&lt;br /&gt;
** [https://dornsifecms.usc.edu/assets/sites/520/docs/VonNeumann-ams12p36-38.pdf von Neumann&#039;s Bernoulli factory] and other [https://peteroupc.github.io/bernoulli.html Bernoulli factory algorithms]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Boy_or_Girl_paradox Boy or Girl paradox]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Monty_Hall_problem Monty Hall problem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bertrand_paradox_(probability) Bertrand paradox]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hard_spheres Hard spheres model] and [https://en.wikipedia.org/wiki/Ising_model Ising model]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/PageRank &#039;&#039;PageRank&#039;&#039;] and stationary [https://en.wikipedia.org/wiki/Random_walk random walk]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Diffusion_process Diffusion process] and [https://en.wikipedia.org/wiki/Diffusion_model diffusion model]&lt;br /&gt;
*[https://en.wikipedia.org/wiki/Probability_space Probability space]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Sample_space Sample space]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Event_(probability_theory) Event] and [https://en.wikipedia.org/wiki/Σ-algebra &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-algebra]&lt;br /&gt;
** Kolmogorov&#039;s [https://en.wikipedia.org/wiki/Probability_axioms axioms of probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Discrete_uniform_distribution Classical] and [https://en.wikipedia.org/wiki/Geometric_probability goemetric probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Boole%27s_inequality Union bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Inclusion%E2%80%93exclusion_principle Inclusion-Exclusion principle]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Boole%27s_inequality#Bonferroni_inequalities Bonferroni inequalities]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Conditional_probability Conditional probability]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chain_rule_(probability) Chain rule]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_total_probability Law of total probability]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bayes%27_theorem Bayes&#039; law]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Independence_(probability_theory) Independence] &lt;br /&gt;
** [https://en.wikipedia.org/wiki/Pairwise_independence Pairwise independence]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Random_variable Random variable]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cumulative_distribution_function Cumulative distribution function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_mass_function Probability mass function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_density_function Probability density function]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Multivariate_random_variable Random vector]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Joint_probability_distribution Joint probability distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Conditional_probability_distribution Conditional probability distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Marginal_distribution Marginal distribution]&lt;br /&gt;
* Some &#039;&#039;&#039;discrete&#039;&#039;&#039; probability distributions&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bernoulli_trial Bernoulli trial] and [https://en.wikipedia.org/wiki/Bernoulli_distribution Bernoulli distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Discrete_uniform_distribution Discrete uniform distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Binomial_distribution Binomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Geometric_distribution Geometric distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Negative_binomial_distribution Negative binomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hypergeometric_distribution Hypergeometric distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Poisson_distribution Poisson distribution]&lt;br /&gt;
** and [https://en.wikipedia.org/wiki/List_of_probability_distributions#Discrete_distributions others]&lt;br /&gt;
* Balls into bins model&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Multinomial_distribution Multinomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Birthday_problem Birthday problem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Coupon_collector%27s_problem Coupon collector]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Balls_into_bins_problem Occupancy problem]&lt;br /&gt;
* Random graphs&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Erd%C5%91s%E2%80%93R%C3%A9nyi_model Erdős–Rényi random graph model]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Galton%E2%80%93Watson_process Galton–Watson branching process]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Expected_value Expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_the_unconscious_statistician Law of the unconscious statistician, &#039;&#039;LOTUS&#039;&#039;]&lt;br /&gt;
** [https://dlsun.github.io/probability/linearity.html Linearity of expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Conditional_expectation Conditional expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_total_expectation Law of total expectation]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Markov%27s_inequality Markov&#039;s inequality]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Chebyshev%27s_inequality Chebyshev&#039;s inequality]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Moment_(mathematics) Moment] related&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Central_moment Central moment]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Variance Variance]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Standard_deviation Standard deviation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Covariance Covariance]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Correlation Correlation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Correlation_does_not_imply_causation Correlation does not imply causation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Skewness Skewness]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Kurtosis Kurtosis]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Moment_problem Moment problem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Measure_(mathematics) Measure]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Borel_set Borel set]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Lebesgue_integration Lebesgue integration]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Measurable_function Measurable function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cantor_set Cantor set]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Non-measurable_set Non-measurable set]&lt;br /&gt;
*[https://en.wikipedia.org/wiki/Probability_density_function Probability density function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_distribution#Absolutely_continuous_probability_distribution Continuous probability distribution]&lt;br /&gt;
**[https://en.wikipedia.org/wiki/Conditional_probability_distribution#Conditional_continuous_distributions Conditional continuous distributions]&lt;br /&gt;
**[https://en.wikipedia.org/wiki/Convolution_of_probability_distributions Convolution of probability distributions] and [https://en.wikipedia.org/wiki/List_of_convolutions_of_probability_distributions List of convolutions of probability distributions]&lt;br /&gt;
* Some &#039;&#039;&#039;continuous&#039;&#039;&#039; probability distributions&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Continuous_uniform_distribution Continuous uniform distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Exponential_distribution Exponential distribution] and [https://en.wikipedia.org/wiki/Poisson_point_process Poisson point process]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Normal_distribution Normal (Gaussion) distribution]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Gaussian_function Gaussian function] and [https://en.wikipedia.org/wiki/Gaussian_integral Gaussian integral]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Standard_normal_table Standard normal table]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/68%E2%80%9395%E2%80%9399.7_rule 68–95–99.7 rule]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Sum_of_normally_distributed_random_variables Sum of normally distributed random variables]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Multivariate_normal_distribution Multivariate normal distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chi-squared_distribution Chi-squared distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Gamma_distribution Gamma distribution] and [https://en.wikipedia.org/wiki/Gamma_function Gamma function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Beta_distribution Beta distribution] and [https://en.wikipedia.org/wiki/Beta_function Beta function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cauchy_distribution Cauchy distribution]&lt;br /&gt;
** and [https://en.wikipedia.org/wiki/List_of_probability_distributions#Absolutely_continuous_distributions others]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Inverse_transform_sampling Inverse transform sampling]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stochastic_dominance Stochastic dominance] and [https://en.wikipedia.org/wiki/Coupling_(probability) Coupling] &lt;br /&gt;
* [https://en.wikipedia.org/wiki/Moment-generating_function Moment-generating function]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Convergence_of_random_variables Convergence of random variables]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Pointwise_convergence Pointwise convergence]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Convergence_of_measures#Weak_convergence_of_measures Weak convergence of measures]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Convergence_in_measure Convergence in measure]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Skorokhod%27s_representation_theorem Skorokhod&#039;s representation theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Continuous_mapping_theorem Continuous mapping theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Borel%E2%80%93Cantelli_lemma Borel–Cantelli lemma] and [https://en.wikipedia.org/wiki/Zero%E2%80%93one_law zero-one laws]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Law_of_large_numbers Law of large numbers]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Central_limit_theorem Central limit theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/De_Moivre%E2%80%93Laplace_theorem De Moivre–Laplace theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Berry%E2%80%93Esseen_theorem Berry–Esseen theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Characteristic_function_(probability_theory) Characteristic function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Lévy%27s_continuity_theorem Lévy&#039;s continuity theorem]&lt;br /&gt;
* Concentration of measures:&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Moment-generating_function Moment-generating function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chernoff_bound Chernoff bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hoeffding%27s_inequality Hoeffding&#039;s bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Doob_martingale#McDiarmid&#039;s_inequality McDiarmid&#039;s inequality]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Doob_martingale Doob martingale] and [https://en.wikipedia.org/wiki/Azuma%27s_inequality Azuma&#039;s inequality]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Sub-Gaussian_distribution Sub-Gaussian distribution]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stochastic_process Stochastic process]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stopping_time Stopping time]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Martingale_(probability_theory) Martingale]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Optional_stopping_theorem Optional stopping theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Wald%27s_equation Wald&#039;s equation]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Markov_chain Markov chain]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Markov_property Markov property] and [https://en.wikipedia.org/wiki/Memorylessness memorylessness]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Stochastic_matrix Stochastic matrix]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Stationary_distribution Stationary distribution]&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)&amp;diff=12864</id>
		<title>概率论与数理统计 (Spring 2025)</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)&amp;diff=12864"/>
		<updated>2025-02-17T06:25:33Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: /* Course info */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Infobox&lt;br /&gt;
|name         = Infobox&lt;br /&gt;
|bodystyle    = &lt;br /&gt;
|title        = &amp;lt;font size=3&amp;gt;&#039;&#039;&#039;概率论与数理统计&#039;&#039;&#039;&amp;lt;br&amp;gt;&lt;br /&gt;
&#039;&#039;&#039;Probability Theory&#039;&#039;&#039; &amp;lt;br&amp;gt; &amp;amp; &#039;&#039;&#039;Mathematical Statistics&#039;&#039;&#039;&amp;lt;/font&amp;gt;&lt;br /&gt;
|titlestyle   = &lt;br /&gt;
&lt;br /&gt;
|image        = &lt;br /&gt;
|imagestyle   = &lt;br /&gt;
|caption      = &lt;br /&gt;
|captionstyle = &lt;br /&gt;
|headerstyle  = background:#ccf;&lt;br /&gt;
|labelstyle   = background:#ddf;&lt;br /&gt;
|datastyle    = &lt;br /&gt;
&lt;br /&gt;
|header1 =Instructor&lt;br /&gt;
|label1  = &lt;br /&gt;
|data1   = &lt;br /&gt;
|header2 = &lt;br /&gt;
|label2  = &lt;br /&gt;
|data2   = &#039;&#039;&#039;尹一通&#039;&#039;&#039;&lt;br /&gt;
|header3 = &lt;br /&gt;
|label3  = Email&lt;br /&gt;
|data3   = yinyt@nju.edu.cn  &lt;br /&gt;
|header4 =&lt;br /&gt;
|label4  = office&lt;br /&gt;
|data4   = 计算机系 804&lt;br /&gt;
|header5 = &lt;br /&gt;
|label5  = &lt;br /&gt;
|data5   = &#039;&#039;&#039;刘景铖&#039;&#039;&#039;&lt;br /&gt;
|header6 = &lt;br /&gt;
|label6  = Email&lt;br /&gt;
|data6   = liu@nju.edu.cn  &lt;br /&gt;
|header7 =&lt;br /&gt;
|label7  = office&lt;br /&gt;
|data7   = 计算机系 516&lt;br /&gt;
|header8 = Class&lt;br /&gt;
|label8  = &lt;br /&gt;
|data8   = &lt;br /&gt;
|header9 =&lt;br /&gt;
|label9  = Class meeting&lt;br /&gt;
|data9   = Monday, 10am-12pm&amp;lt;br&amp;gt;&lt;br /&gt;
Wednesday (单), 10am-12pm&amp;lt;br&amp;gt;&lt;br /&gt;
仙Ⅰ-204&lt;br /&gt;
|header10=&lt;br /&gt;
|label10 = Office hour&lt;br /&gt;
|data10  = TBD &amp;lt;br&amp;gt;计算机系 804（尹一通）&amp;lt;br&amp;gt;计算机系 516（刘景铖）&lt;br /&gt;
|header11= Textbook&lt;br /&gt;
|label11 = &lt;br /&gt;
|data11  = &lt;br /&gt;
|header12=&lt;br /&gt;
|label12 = &lt;br /&gt;
|data12  = [[File:概率导论.jpeg|border|100px]]&lt;br /&gt;
|header13=&lt;br /&gt;
|label13 = &lt;br /&gt;
|data13  = &#039;&#039;&#039;概率导论&#039;&#039;&#039;（第2版·修订版）&amp;lt;br&amp;gt; Dimitri P. Bertsekas and John N. Tsitsiklis&amp;lt;br&amp;gt; 郑忠国 童行伟 译；人民邮电出版社 (2022)&lt;br /&gt;
|header14=&lt;br /&gt;
|label14 = &lt;br /&gt;
|data14  = [[File:Grimmett_probability.jpg|border|100px]]&lt;br /&gt;
|header15=&lt;br /&gt;
|label15 = &lt;br /&gt;
|data15  = &#039;&#039;&#039;Probability and Random Processes&#039;&#039;&#039; (4E) &amp;lt;br&amp;gt; Geoffrey Grimmett and David Stirzaker &amp;lt;br&amp;gt;  Oxford University Press (2020)&lt;br /&gt;
|header16=&lt;br /&gt;
|label16 = &lt;br /&gt;
|data16  = [[File:Probability_and_Computing_2ed.jpg|border|100px]]&lt;br /&gt;
|header17=&lt;br /&gt;
|label17 = &lt;br /&gt;
|data17  = &#039;&#039;&#039;Probability and Computing&#039;&#039;&#039; (2E) &amp;lt;br&amp;gt; Michael Mitzenmacher and Eli Upfal &amp;lt;br&amp;gt;   Cambridge University Press (2017)&lt;br /&gt;
|belowstyle = background:#ddf;&lt;br /&gt;
|below = &lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
This is the webpage for the &#039;&#039;Probability Theory and Mathematical Statistics&#039;&#039; (概率论与数理统计) class of Spring 2025. Students who take this class should check this page periodically for content updates and new announcements. &lt;br /&gt;
&lt;br /&gt;
= Announcement =&lt;br /&gt;
* TBA&lt;br /&gt;
&lt;br /&gt;
= Course info =&lt;br /&gt;
* &#039;&#039;&#039;Instructor &#039;&#039;&#039;: &lt;br /&gt;
:* [http://tcs.nju.edu.cn/yinyt/ 尹一通]：[mailto:yinyt@nju.edu.cn &amp;lt;yinyt@nju.edu.cn&amp;gt;]，计算机系 804 &lt;br /&gt;
:* [https://liuexp.github.io 刘景铖]：[mailto:liu@nju.edu.cn &amp;lt;liu@nju.edu.cn&amp;gt;]，计算机系 516 &lt;br /&gt;
* &#039;&#039;&#039;Teaching assistant&#039;&#039;&#039;:&lt;br /&gt;
** [https://sites.google.com/view/xinyuanzhang 张昕渊]：[mailto:zhangxy@smail.nju.edu.cn &amp;lt;zhangxy@smail.nju.edu.cn&amp;gt;]，计算机系 410&lt;br /&gt;
** 邹宗瑞：[mailto:zou.zongrui@smail.nju.edu.cn &amp;lt;zou.zongrui@smail.nju.edu.cn&amp;gt;]，计算机系 410&lt;br /&gt;
* &#039;&#039;&#039;Class meeting&#039;&#039;&#039;:&lt;br /&gt;
** 周一：10am-12pm，仙Ⅰ-204&lt;br /&gt;
** 周三（单）：10am-12pm，仙Ⅰ-204&lt;br /&gt;
* &#039;&#039;&#039;Office hour&#039;&#039;&#039;: &lt;br /&gt;
:* 周五 2pm-3pm, 计算机系 804（尹一通）&lt;br /&gt;
:* 周五 2pm-3pm, 计算机系 516（刘景铖）&lt;br /&gt;
:* &#039;&#039;&#039;QQ群&#039;&#039;&#039;: 974742320（申请加入需提供姓名、院系、学号）&lt;br /&gt;
&lt;br /&gt;
= Syllabus =&lt;br /&gt;
课程内容分为三大部分：&lt;br /&gt;
* &#039;&#039;&#039;经典概率论&#039;&#039;&#039;：概率空间、随机变量及其数字特征、多维与连续随机变量、极限定理等内容&lt;br /&gt;
* &#039;&#039;&#039;概率与计算&#039;&#039;&#039;：测度集中现象 (concentration of measure)、概率法 (the probabilistic method)、离散随机过程的相关专题&lt;br /&gt;
* &#039;&#039;&#039;数理统计&#039;&#039;&#039;：参数估计、假设检验、贝叶斯估计、线性回归等统计推断等概念&lt;br /&gt;
&lt;br /&gt;
对于第一和第二部分，要求清楚掌握基本概念，深刻理解关键的现象与规律以及背后的原理，并可以灵活运用所学方法求解相关问题。对于第三部分，要求熟悉数理统计的若干基本概念，以及典型的统计模型和统计推断问题。&lt;br /&gt;
&lt;br /&gt;
经过本课程的训练，力求使学生能够熟悉掌握概率的语言，并会利用概率思维来理解客观世界并对其建模，以及驾驭概率的数学工具来分析和求解专业问题。&lt;br /&gt;
&lt;br /&gt;
=== 教材与参考书 Course Materials ===&lt;br /&gt;
* &#039;&#039;&#039;[BT]&#039;&#039;&#039; 概率导论（第2版·修订版），[美]伯特瑟卡斯（Dimitri P.Bertsekas）[美]齐齐克利斯（John N.Tsitsiklis）著，郑忠国 童行伟 译，人民邮电出版社（2022）。&lt;br /&gt;
* &#039;&#039;&#039;[GS]&#039;&#039;&#039; &#039;&#039;Probability and Random Processes&#039;&#039;, by Geoffrey Grimmett and David Stirzaker; Oxford University Press; 4th edition (2020).&lt;br /&gt;
* &#039;&#039;&#039;[MU]&#039;&#039;&#039; &#039;&#039;Probability and Computing: Randomization and Probabilistic Techniques in Algorithms and Data Analysis&#039;&#039;, by Michael Mitzenmacher, Eli Upfal; Cambridge University Press; 2nd edition (2017).&lt;br /&gt;
&lt;br /&gt;
=== 成绩 Grading Policy ===&lt;br /&gt;
* 课程成绩：本课程将会有若干次作业和一次期末考试。最终成绩将由平时作业成绩和期末考试成绩综合得出。&lt;br /&gt;
* 迟交：如果有特殊的理由，无法按时完成作业，请提前联系授课老师，给出正当理由。否则迟交的作业将不被接受。&lt;br /&gt;
&lt;br /&gt;
=== &amp;lt;font color=red&amp;gt; 学术诚信 Academic Integrity &amp;lt;/font&amp;gt;===&lt;br /&gt;
学术诚信是所有从事学术活动的学生和学者最基本的职业道德底线，本课程将不遗余力的维护学术诚信规范，违反这一底线的行为将不会被容忍。&lt;br /&gt;
&lt;br /&gt;
作业完成的原则：署你名字的工作必须是你个人的贡献。在完成作业的过程中，允许讨论，前提是讨论的所有参与者均处于同等完成度。但关键想法的执行、以及作业文本的写作必须独立完成，并在作业中致谢（acknowledge）所有参与讨论的人。符合规则的讨论与致谢将不会影响得分。不允许其他任何形式的合作——尤其是与已经完成作业的同学“讨论”。&lt;br /&gt;
&lt;br /&gt;
本课程将对剽窃行为采取零容忍的态度。在完成作业过程中，对他人工作（出版物、互联网资料、其他人的作业等）直接的文本抄袭和对关键思想、关键元素的抄袭，按照 [http://www.acm.org/publications/policies/plagiarism_policy ACM Policy on Plagiarism]的解释，都将视为剽窃。剽窃者成绩将被取消。如果发现互相抄袭行为，&amp;lt;font color=red&amp;gt; 抄袭和被抄袭双方的成绩都将被取消&amp;lt;/font&amp;gt;。因此请主动防止自己的作业被他人抄袭。&lt;br /&gt;
&lt;br /&gt;
学术诚信影响学生个人的品行，也关乎整个教育系统的正常运转。为了一点分数而做出学术不端的行为，不仅使自己沦为一个欺骗者，也使他人的诚实努力失去意义。让我们一起努力维护一个诚信的环境。&lt;br /&gt;
&lt;br /&gt;
= Assignments =&lt;br /&gt;
TBA&lt;br /&gt;
&lt;br /&gt;
= Lectures =&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2025/Intro.pdf 课程简介]&lt;br /&gt;
&lt;br /&gt;
= Concepts =&lt;br /&gt;
* [https://plato.stanford.edu/entries/probability-interpret/ Interpretations of probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/History_of_probability History of probability]&lt;br /&gt;
* Example problems:&lt;br /&gt;
** [https://dornsifecms.usc.edu/assets/sites/520/docs/VonNeumann-ams12p36-38.pdf von Neumann&#039;s Bernoulli factory] and other [https://peteroupc.github.io/bernoulli.html Bernoulli factory algorithms]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Boy_or_Girl_paradox Boy or Girl paradox]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Monty_Hall_problem Monty Hall problem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bertrand_paradox_(probability) Bertrand paradox]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hard_spheres Hard spheres model] and [https://en.wikipedia.org/wiki/Ising_model Ising model]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/PageRank &#039;&#039;PageRank&#039;&#039;] and stationary [https://en.wikipedia.org/wiki/Random_walk random walk]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Diffusion_process Diffusion process] and [https://en.wikipedia.org/wiki/Diffusion_model diffusion model]&lt;br /&gt;
*[https://en.wikipedia.org/wiki/Probability_space Probability space]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Sample_space Sample space]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Event_(probability_theory) Event] and [https://en.wikipedia.org/wiki/Σ-algebra &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-algebra]&lt;br /&gt;
** Kolmogorov&#039;s [https://en.wikipedia.org/wiki/Probability_axioms axioms of probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Discrete_uniform_distribution Classical] and [https://en.wikipedia.org/wiki/Geometric_probability goemetric probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Boole%27s_inequality Union bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Inclusion%E2%80%93exclusion_principle Inclusion-Exclusion principle]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Boole%27s_inequality#Bonferroni_inequalities Bonferroni inequalities]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Conditional_probability Conditional probability]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chain_rule_(probability) Chain rule]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_total_probability Law of total probability]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bayes%27_theorem Bayes&#039; law]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Independence_(probability_theory) Independence] &lt;br /&gt;
** [https://en.wikipedia.org/wiki/Pairwise_independence Pairwise independence]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Random_variable Random variable]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cumulative_distribution_function Cumulative distribution function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_mass_function Probability mass function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_density_function Probability density function]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Multivariate_random_variable Random vector]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Joint_probability_distribution Joint probability distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Conditional_probability_distribution Conditional probability distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Marginal_distribution Marginal distribution]&lt;br /&gt;
* Some &#039;&#039;&#039;discrete&#039;&#039;&#039; probability distributions&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bernoulli_trial Bernoulli trial] and [https://en.wikipedia.org/wiki/Bernoulli_distribution Bernoulli distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Discrete_uniform_distribution Discrete uniform distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Binomial_distribution Binomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Geometric_distribution Geometric distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Negative_binomial_distribution Negative binomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hypergeometric_distribution Hypergeometric distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Poisson_distribution Poisson distribution]&lt;br /&gt;
** and [https://en.wikipedia.org/wiki/List_of_probability_distributions#Discrete_distributions others]&lt;br /&gt;
* Balls into bins model&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Multinomial_distribution Multinomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Birthday_problem Birthday problem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Coupon_collector%27s_problem Coupon collector]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Balls_into_bins_problem Occupancy problem]&lt;br /&gt;
* Random graphs&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Erd%C5%91s%E2%80%93R%C3%A9nyi_model Erdős–Rényi random graph model]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Galton%E2%80%93Watson_process Galton–Watson branching process]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Expected_value Expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_the_unconscious_statistician Law of the unconscious statistician, &#039;&#039;LOTUS&#039;&#039;]&lt;br /&gt;
** [https://dlsun.github.io/probability/linearity.html Linearity of expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Conditional_expectation Conditional expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_total_expectation Law of total expectation]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Markov%27s_inequality Markov&#039;s inequality]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Chebyshev%27s_inequality Chebyshev&#039;s inequality]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Moment_(mathematics) Moment] related&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Central_moment Central moment]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Variance Variance]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Standard_deviation Standard deviation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Covariance Covariance]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Correlation Correlation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Correlation_does_not_imply_causation Correlation does not imply causation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Skewness Skewness]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Kurtosis Kurtosis]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Moment_problem Moment problem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Measure_(mathematics) Measure]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Borel_set Borel set]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Lebesgue_integration Lebesgue integration]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Measurable_function Measurable function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cantor_set Cantor set]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Non-measurable_set Non-measurable set]&lt;br /&gt;
*[https://en.wikipedia.org/wiki/Probability_density_function Probability density function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_distribution#Absolutely_continuous_probability_distribution Continuous probability distribution]&lt;br /&gt;
**[https://en.wikipedia.org/wiki/Conditional_probability_distribution#Conditional_continuous_distributions Conditional continuous distributions]&lt;br /&gt;
**[https://en.wikipedia.org/wiki/Convolution_of_probability_distributions Convolution of probability distributions] and [https://en.wikipedia.org/wiki/List_of_convolutions_of_probability_distributions List of convolutions of probability distributions]&lt;br /&gt;
* Some &#039;&#039;&#039;continuous&#039;&#039;&#039; probability distributions&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Continuous_uniform_distribution Continuous uniform distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Exponential_distribution Exponential distribution] and [https://en.wikipedia.org/wiki/Poisson_point_process Poisson point process]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Normal_distribution Normal (Gaussion) distribution]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Gaussian_function Gaussian function] and [https://en.wikipedia.org/wiki/Gaussian_integral Gaussian integral]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Standard_normal_table Standard normal table]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/68%E2%80%9395%E2%80%9399.7_rule 68–95–99.7 rule]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Sum_of_normally_distributed_random_variables Sum of normally distributed random variables]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Multivariate_normal_distribution Multivariate normal distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chi-squared_distribution Chi-squared distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Gamma_distribution Gamma distribution] and [https://en.wikipedia.org/wiki/Gamma_function Gamma function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Beta_distribution Beta distribution] and [https://en.wikipedia.org/wiki/Beta_function Beta function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cauchy_distribution Cauchy distribution]&lt;br /&gt;
** and [https://en.wikipedia.org/wiki/List_of_probability_distributions#Absolutely_continuous_distributions others]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Inverse_transform_sampling Inverse transform sampling]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stochastic_dominance Stochastic dominance] and [https://en.wikipedia.org/wiki/Coupling_(probability) Coupling] &lt;br /&gt;
* [https://en.wikipedia.org/wiki/Moment-generating_function Moment-generating function]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Convergence_of_random_variables Convergence of random variables]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Pointwise_convergence Pointwise convergence]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Convergence_of_measures#Weak_convergence_of_measures Weak convergence of measures]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Convergence_in_measure Convergence in measure]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Skorokhod%27s_representation_theorem Skorokhod&#039;s representation theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Continuous_mapping_theorem Continuous mapping theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Borel%E2%80%93Cantelli_lemma Borel–Cantelli lemma] and [https://en.wikipedia.org/wiki/Zero%E2%80%93one_law zero-one laws]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Law_of_large_numbers Law of large numbers]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Central_limit_theorem Central limit theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/De_Moivre%E2%80%93Laplace_theorem De Moivre–Laplace theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Berry%E2%80%93Esseen_theorem Berry–Esseen theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Characteristic_function_(probability_theory) Characteristic function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Lévy%27s_continuity_theorem Lévy&#039;s continuity theorem]&lt;br /&gt;
* Concentration of measures:&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Moment-generating_function Moment-generating function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chernoff_bound Chernoff bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hoeffding%27s_inequality Hoeffding&#039;s bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Doob_martingale#McDiarmid&#039;s_inequality McDiarmid&#039;s inequality]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Doob_martingale Doob martingale] and [https://en.wikipedia.org/wiki/Azuma%27s_inequality Azuma&#039;s inequality]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Sub-Gaussian_distribution Sub-Gaussian distribution]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stochastic_process Stochastic process]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stopping_time Stopping time]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Martingale_(probability_theory) Martingale]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Optional_stopping_theorem Optional stopping theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Wald%27s_equation Wald&#039;s equation]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Markov_chain Markov chain]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Markov_property Markov property] and [https://en.wikipedia.org/wiki/Memorylessness memorylessness]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Stochastic_matrix Stochastic matrix]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Stationary_distribution Stationary distribution]&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)&amp;diff=12854</id>
		<title>概率论与数理统计 (Spring 2025)</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)&amp;diff=12854"/>
		<updated>2025-02-16T07:34:01Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: /* Assignments */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Infobox&lt;br /&gt;
|name         = Infobox&lt;br /&gt;
|bodystyle    = &lt;br /&gt;
|title        = &amp;lt;font size=3&amp;gt;&#039;&#039;&#039;概率论与数理统计&#039;&#039;&#039;&amp;lt;br&amp;gt;&lt;br /&gt;
&#039;&#039;&#039;Probability Theory&#039;&#039;&#039; &amp;lt;br&amp;gt; &amp;amp; &#039;&#039;&#039;Mathematical Statistics&#039;&#039;&#039;&amp;lt;/font&amp;gt;&lt;br /&gt;
|titlestyle   = &lt;br /&gt;
&lt;br /&gt;
|image        = &lt;br /&gt;
|imagestyle   = &lt;br /&gt;
|caption      = &lt;br /&gt;
|captionstyle = &lt;br /&gt;
|headerstyle  = background:#ccf;&lt;br /&gt;
|labelstyle   = background:#ddf;&lt;br /&gt;
|datastyle    = &lt;br /&gt;
&lt;br /&gt;
|header1 =Instructor&lt;br /&gt;
|label1  = &lt;br /&gt;
|data1   = &lt;br /&gt;
|header2 = &lt;br /&gt;
|label2  = &lt;br /&gt;
|data2   = &#039;&#039;&#039;尹一通&#039;&#039;&#039;&lt;br /&gt;
|header3 = &lt;br /&gt;
|label3  = Email&lt;br /&gt;
|data3   = yinyt@nju.edu.cn  &lt;br /&gt;
|header4 =&lt;br /&gt;
|label4  = office&lt;br /&gt;
|data4   = 计算机系 804&lt;br /&gt;
|header5 = &lt;br /&gt;
|label5  = &lt;br /&gt;
|data5   = &#039;&#039;&#039;刘景铖&#039;&#039;&#039;&lt;br /&gt;
|header6 = &lt;br /&gt;
|label6  = Email&lt;br /&gt;
|data6   = liu@nju.edu.cn  &lt;br /&gt;
|header7 =&lt;br /&gt;
|label7  = office&lt;br /&gt;
|data7   = 计算机系 516&lt;br /&gt;
|header8 = Class&lt;br /&gt;
|label8  = &lt;br /&gt;
|data8   = &lt;br /&gt;
|header9 =&lt;br /&gt;
|label9  = Class meeting&lt;br /&gt;
|data9   = Monday, 10am-12pm&amp;lt;br&amp;gt;&lt;br /&gt;
Wednesday (单), 10am-12pm&amp;lt;br&amp;gt;&lt;br /&gt;
仙Ⅰ-204&lt;br /&gt;
|header10=&lt;br /&gt;
|label10 = Office hour&lt;br /&gt;
|data10  = TBD &amp;lt;br&amp;gt;计算机系 804（尹一通）&amp;lt;br&amp;gt;计算机系 516（刘景铖）&lt;br /&gt;
|header11= Textbook&lt;br /&gt;
|label11 = &lt;br /&gt;
|data11  = &lt;br /&gt;
|header12=&lt;br /&gt;
|label12 = &lt;br /&gt;
|data12  = [[File:概率导论.jpeg|border|100px]]&lt;br /&gt;
|header13=&lt;br /&gt;
|label13 = &lt;br /&gt;
|data13  = &#039;&#039;&#039;概率导论&#039;&#039;&#039;（第2版·修订版）&amp;lt;br&amp;gt; Dimitri P. Bertsekas and John N. Tsitsiklis&amp;lt;br&amp;gt; 郑忠国 童行伟 译；人民邮电出版社 (2022)&lt;br /&gt;
|header14=&lt;br /&gt;
|label14 = &lt;br /&gt;
|data14  = [[File:Grimmett_probability.jpg|border|100px]]&lt;br /&gt;
|header15=&lt;br /&gt;
|label15 = &lt;br /&gt;
|data15  = &#039;&#039;&#039;Probability and Random Processes&#039;&#039;&#039; (4E) &amp;lt;br&amp;gt; Geoffrey Grimmett and David Stirzaker &amp;lt;br&amp;gt;  Oxford University Press (2020)&lt;br /&gt;
|header16=&lt;br /&gt;
|label16 = &lt;br /&gt;
|data16  = [[File:Probability_and_Computing_2ed.jpg|border|100px]]&lt;br /&gt;
|header17=&lt;br /&gt;
|label17 = &lt;br /&gt;
|data17  = &#039;&#039;&#039;Probability and Computing&#039;&#039;&#039; (2E) &amp;lt;br&amp;gt; Michael Mitzenmacher and Eli Upfal &amp;lt;br&amp;gt;   Cambridge University Press (2017)&lt;br /&gt;
|belowstyle = background:#ddf;&lt;br /&gt;
|below = &lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
This is the webpage for the &#039;&#039;Probability Theory and Mathematical Statistics&#039;&#039; (概率论与数理统计) class of Spring 2025. Students who take this class should check this page periodically for content updates and new announcements. &lt;br /&gt;
&lt;br /&gt;
= Announcement =&lt;br /&gt;
* TBA&lt;br /&gt;
&lt;br /&gt;
= Course info =&lt;br /&gt;
* &#039;&#039;&#039;Instructor &#039;&#039;&#039;: &lt;br /&gt;
:* [http://tcs.nju.edu.cn/yinyt/ 尹一通]：[mailto:yinyt@nju.edu.cn &amp;lt;yinyt@nju.edu.cn&amp;gt;]，计算机系 804 &lt;br /&gt;
:* [https://liuexp.github.io 刘景铖]：[mailto:liu@nju.edu.cn &amp;lt;liu@nju.edu.cn&amp;gt;]，计算机系 516 &lt;br /&gt;
* &#039;&#039;&#039;Teaching assistant&#039;&#039;&#039;:&lt;br /&gt;
** [https://sites.google.com/view/xinyuanzhang 张昕渊]：[mailto:zhangxy@smail.nju.edu.cn &amp;lt;zhangxy@smail.nju.edu.cn&amp;gt;]，计算机系 410&lt;br /&gt;
** 邹宗瑞：[mailto:zou.zongrui@smail.nju.edu.cn &amp;lt;zou.zongrui@smail.nju.edu.cn&amp;gt;]，计算机系 410&lt;br /&gt;
* &#039;&#039;&#039;Class meeting&#039;&#039;&#039;:&lt;br /&gt;
** 周一：10am-12pm，仙Ⅰ-204&lt;br /&gt;
** 周三（单）：10am-12pm，仙Ⅰ-204&lt;br /&gt;
* &#039;&#039;&#039;Office hour&#039;&#039;&#039;: &lt;br /&gt;
:* TBD, 计算机系 804（尹一通）&lt;br /&gt;
:* TBD, 计算机系 516（刘景铖）&lt;br /&gt;
:* &#039;&#039;&#039;QQ群&#039;&#039;&#039;: 974742320（申请加入需提供姓名、院系、学号）&lt;br /&gt;
&lt;br /&gt;
= Syllabus =&lt;br /&gt;
课程内容分为三大部分：&lt;br /&gt;
* &#039;&#039;&#039;经典概率论&#039;&#039;&#039;：概率空间、随机变量及其数字特征、多维与连续随机变量、极限定理等内容&lt;br /&gt;
* &#039;&#039;&#039;概率与计算&#039;&#039;&#039;：测度集中现象 (concentration of measure)、概率法 (the probabilistic method)、离散随机过程的相关专题&lt;br /&gt;
* &#039;&#039;&#039;数理统计&#039;&#039;&#039;：参数估计、假设检验、贝叶斯估计、线性回归等统计推断等概念&lt;br /&gt;
&lt;br /&gt;
对于第一和第二部分，要求清楚掌握基本概念，深刻理解关键的现象与规律以及背后的原理，并可以灵活运用所学方法求解相关问题。对于第三部分，要求熟悉数理统计的若干基本概念，以及典型的统计模型和统计推断问题。&lt;br /&gt;
&lt;br /&gt;
经过本课程的训练，力求使学生能够熟悉掌握概率的语言，并会利用概率思维来理解客观世界并对其建模，以及驾驭概率的数学工具来分析和求解专业问题。&lt;br /&gt;
&lt;br /&gt;
=== 教材与参考书 Course Materials ===&lt;br /&gt;
* &#039;&#039;&#039;[BT]&#039;&#039;&#039; 概率导论（第2版·修订版），[美]伯特瑟卡斯（Dimitri P.Bertsekas）[美]齐齐克利斯（John N.Tsitsiklis）著，郑忠国 童行伟 译，人民邮电出版社（2022）。&lt;br /&gt;
* &#039;&#039;&#039;[GS]&#039;&#039;&#039; &#039;&#039;Probability and Random Processes&#039;&#039;, by Geoffrey Grimmett and David Stirzaker; Oxford University Press; 4th edition (2020).&lt;br /&gt;
* &#039;&#039;&#039;[MU]&#039;&#039;&#039; &#039;&#039;Probability and Computing: Randomization and Probabilistic Techniques in Algorithms and Data Analysis&#039;&#039;, by Michael Mitzenmacher, Eli Upfal; Cambridge University Press; 2nd edition (2017).&lt;br /&gt;
&lt;br /&gt;
=== 成绩 Grading Policy ===&lt;br /&gt;
* 课程成绩：本课程将会有若干次作业和一次期末考试。最终成绩将由平时作业成绩和期末考试成绩综合得出。&lt;br /&gt;
* 迟交：如果有特殊的理由，无法按时完成作业，请提前联系授课老师，给出正当理由。否则迟交的作业将不被接受。&lt;br /&gt;
&lt;br /&gt;
=== &amp;lt;font color=red&amp;gt; 学术诚信 Academic Integrity &amp;lt;/font&amp;gt;===&lt;br /&gt;
学术诚信是所有从事学术活动的学生和学者最基本的职业道德底线，本课程将不遗余力的维护学术诚信规范，违反这一底线的行为将不会被容忍。&lt;br /&gt;
&lt;br /&gt;
作业完成的原则：署你名字的工作必须是你个人的贡献。在完成作业的过程中，允许讨论，前提是讨论的所有参与者均处于同等完成度。但关键想法的执行、以及作业文本的写作必须独立完成，并在作业中致谢（acknowledge）所有参与讨论的人。符合规则的讨论与致谢将不会影响得分。不允许其他任何形式的合作——尤其是与已经完成作业的同学“讨论”。&lt;br /&gt;
&lt;br /&gt;
本课程将对剽窃行为采取零容忍的态度。在完成作业过程中，对他人工作（出版物、互联网资料、其他人的作业等）直接的文本抄袭和对关键思想、关键元素的抄袭，按照 [http://www.acm.org/publications/policies/plagiarism_policy ACM Policy on Plagiarism]的解释，都将视为剽窃。剽窃者成绩将被取消。如果发现互相抄袭行为，&amp;lt;font color=red&amp;gt; 抄袭和被抄袭双方的成绩都将被取消&amp;lt;/font&amp;gt;。因此请主动防止自己的作业被他人抄袭。&lt;br /&gt;
&lt;br /&gt;
学术诚信影响学生个人的品行，也关乎整个教育系统的正常运转。为了一点分数而做出学术不端的行为，不仅使自己沦为一个欺骗者，也使他人的诚实努力失去意义。让我们一起努力维护一个诚信的环境。&lt;br /&gt;
&lt;br /&gt;
= Assignments =&lt;br /&gt;
TBA&lt;br /&gt;
&lt;br /&gt;
= Lectures =&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2024/Intro.pdf 课程简介]&lt;br /&gt;
&lt;br /&gt;
= Concepts =&lt;br /&gt;
* [https://plato.stanford.edu/entries/probability-interpret/ Interpretations of probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/History_of_probability History of probability]&lt;br /&gt;
* Example problems:&lt;br /&gt;
** [https://dornsifecms.usc.edu/assets/sites/520/docs/VonNeumann-ams12p36-38.pdf von Neumann&#039;s Bernoulli factory] and other [https://peteroupc.github.io/bernoulli.html Bernoulli factory algorithms]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Boy_or_Girl_paradox Boy or Girl paradox]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Monty_Hall_problem Monty Hall problem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bertrand_paradox_(probability) Bertrand paradox]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hard_spheres Hard spheres model] and [https://en.wikipedia.org/wiki/Ising_model Ising model]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/PageRank &#039;&#039;PageRank&#039;&#039;] and stationary [https://en.wikipedia.org/wiki/Random_walk random walk]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Diffusion_process Diffusion process] and [https://en.wikipedia.org/wiki/Diffusion_model diffusion model]&lt;br /&gt;
*[https://en.wikipedia.org/wiki/Probability_space Probability space]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Sample_space Sample space]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Event_(probability_theory) Event] and [https://en.wikipedia.org/wiki/Σ-algebra &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-algebra]&lt;br /&gt;
** Kolmogorov&#039;s [https://en.wikipedia.org/wiki/Probability_axioms axioms of probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Discrete_uniform_distribution Classical] and [https://en.wikipedia.org/wiki/Geometric_probability goemetric probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Boole%27s_inequality Union bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Inclusion%E2%80%93exclusion_principle Inclusion-Exclusion principle]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Boole%27s_inequality#Bonferroni_inequalities Bonferroni inequalities]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Conditional_probability Conditional probability]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chain_rule_(probability) Chain rule]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_total_probability Law of total probability]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bayes%27_theorem Bayes&#039; law]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Independence_(probability_theory) Independence] &lt;br /&gt;
** [https://en.wikipedia.org/wiki/Pairwise_independence Pairwise independence]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Random_variable Random variable]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cumulative_distribution_function Cumulative distribution function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_mass_function Probability mass function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_density_function Probability density function]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Multivariate_random_variable Random vector]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Joint_probability_distribution Joint probability distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Conditional_probability_distribution Conditional probability distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Marginal_distribution Marginal distribution]&lt;br /&gt;
* Some &#039;&#039;&#039;discrete&#039;&#039;&#039; probability distributions&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bernoulli_trial Bernoulli trial] and [https://en.wikipedia.org/wiki/Bernoulli_distribution Bernoulli distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Discrete_uniform_distribution Discrete uniform distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Binomial_distribution Binomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Geometric_distribution Geometric distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Negative_binomial_distribution Negative binomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hypergeometric_distribution Hypergeometric distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Poisson_distribution Poisson distribution]&lt;br /&gt;
** and [https://en.wikipedia.org/wiki/List_of_probability_distributions#Discrete_distributions others]&lt;br /&gt;
* Balls into bins model&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Multinomial_distribution Multinomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Birthday_problem Birthday problem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Coupon_collector%27s_problem Coupon collector]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Balls_into_bins_problem Occupancy problem]&lt;br /&gt;
* Random graphs&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Erd%C5%91s%E2%80%93R%C3%A9nyi_model Erdős–Rényi random graph model]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Galton%E2%80%93Watson_process Galton–Watson branching process]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Expected_value Expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_the_unconscious_statistician Law of the unconscious statistician, &#039;&#039;LOTUS&#039;&#039;]&lt;br /&gt;
** [https://dlsun.github.io/probability/linearity.html Linearity of expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Conditional_expectation Conditional expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_total_expectation Law of total expectation]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Markov%27s_inequality Markov&#039;s inequality]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Chebyshev%27s_inequality Chebyshev&#039;s inequality]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Moment_(mathematics) Moment] related&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Central_moment Central moment]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Variance Variance]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Standard_deviation Standard deviation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Covariance Covariance]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Correlation Correlation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Correlation_does_not_imply_causation Correlation does not imply causation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Skewness Skewness]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Kurtosis Kurtosis]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Moment_problem Moment problem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Measure_(mathematics) Measure]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Borel_set Borel set]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Lebesgue_integration Lebesgue integration]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Measurable_function Measurable function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cantor_set Cantor set]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Non-measurable_set Non-measurable set]&lt;br /&gt;
*[https://en.wikipedia.org/wiki/Probability_density_function Probability density function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_distribution#Absolutely_continuous_probability_distribution Continuous probability distribution]&lt;br /&gt;
**[https://en.wikipedia.org/wiki/Conditional_probability_distribution#Conditional_continuous_distributions Conditional continuous distributions]&lt;br /&gt;
**[https://en.wikipedia.org/wiki/Convolution_of_probability_distributions Convolution of probability distributions] and [https://en.wikipedia.org/wiki/List_of_convolutions_of_probability_distributions List of convolutions of probability distributions]&lt;br /&gt;
* Some &#039;&#039;&#039;continuous&#039;&#039;&#039; probability distributions&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Continuous_uniform_distribution Continuous uniform distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Exponential_distribution Exponential distribution] and [https://en.wikipedia.org/wiki/Poisson_point_process Poisson point process]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Normal_distribution Normal (Gaussion) distribution]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Gaussian_function Gaussian function] and [https://en.wikipedia.org/wiki/Gaussian_integral Gaussian integral]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Standard_normal_table Standard normal table]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/68%E2%80%9395%E2%80%9399.7_rule 68–95–99.7 rule]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Sum_of_normally_distributed_random_variables Sum of normally distributed random variables]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Multivariate_normal_distribution Multivariate normal distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chi-squared_distribution Chi-squared distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Gamma_distribution Gamma distribution] and [https://en.wikipedia.org/wiki/Gamma_function Gamma function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Beta_distribution Beta distribution] and [https://en.wikipedia.org/wiki/Beta_function Beta function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cauchy_distribution Cauchy distribution]&lt;br /&gt;
** and [https://en.wikipedia.org/wiki/List_of_probability_distributions#Absolutely_continuous_distributions others]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Inverse_transform_sampling Inverse transform sampling]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stochastic_dominance Stochastic dominance] and [https://en.wikipedia.org/wiki/Coupling_(probability) Coupling] &lt;br /&gt;
* [https://en.wikipedia.org/wiki/Moment-generating_function Moment-generating function]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Convergence_of_random_variables Convergence of random variables]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Pointwise_convergence Pointwise convergence]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Convergence_of_measures#Weak_convergence_of_measures Weak convergence of measures]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Convergence_in_measure Convergence in measure]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Skorokhod%27s_representation_theorem Skorokhod&#039;s representation theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Continuous_mapping_theorem Continuous mapping theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Borel%E2%80%93Cantelli_lemma Borel–Cantelli lemma] and [https://en.wikipedia.org/wiki/Zero%E2%80%93one_law zero-one laws]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Law_of_large_numbers Law of large numbers]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Central_limit_theorem Central limit theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/De_Moivre%E2%80%93Laplace_theorem De Moivre–Laplace theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Berry%E2%80%93Esseen_theorem Berry–Esseen theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Characteristic_function_(probability_theory) Characteristic function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Lévy%27s_continuity_theorem Lévy&#039;s continuity theorem]&lt;br /&gt;
* Concentration of measures:&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Moment-generating_function Moment-generating function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chernoff_bound Chernoff bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hoeffding%27s_inequality Hoeffding&#039;s bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Doob_martingale#McDiarmid&#039;s_inequality McDiarmid&#039;s inequality]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Doob_martingale Doob martingale] and [https://en.wikipedia.org/wiki/Azuma%27s_inequality Azuma&#039;s inequality]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Sub-Gaussian_distribution Sub-Gaussian distribution]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stochastic_process Stochastic process]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stopping_time Stopping time]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Martingale_(probability_theory) Martingale]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Optional_stopping_theorem Optional stopping theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Wald%27s_equation Wald&#039;s equation]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Markov_chain Markov chain]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Markov_property Markov property] and [https://en.wikipedia.org/wiki/Memorylessness memorylessness]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Stochastic_matrix Stochastic matrix]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Stationary_distribution Stationary distribution]&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)&amp;diff=12851</id>
		<title>概率论与数理统计 (Spring 2025)</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)&amp;diff=12851"/>
		<updated>2025-02-16T07:31:19Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: /* Course info */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Infobox&lt;br /&gt;
|name         = Infobox&lt;br /&gt;
|bodystyle    = &lt;br /&gt;
|title        = &amp;lt;font size=3&amp;gt;&#039;&#039;&#039;概率论与数理统计&#039;&#039;&#039;&amp;lt;br&amp;gt;&lt;br /&gt;
&#039;&#039;&#039;Probability Theory&#039;&#039;&#039; &amp;lt;br&amp;gt; &amp;amp; &#039;&#039;&#039;Mathematical Statistics&#039;&#039;&#039;&amp;lt;/font&amp;gt;&lt;br /&gt;
|titlestyle   = &lt;br /&gt;
&lt;br /&gt;
|image        = &lt;br /&gt;
|imagestyle   = &lt;br /&gt;
|caption      = &lt;br /&gt;
|captionstyle = &lt;br /&gt;
|headerstyle  = background:#ccf;&lt;br /&gt;
|labelstyle   = background:#ddf;&lt;br /&gt;
|datastyle    = &lt;br /&gt;
&lt;br /&gt;
|header1 =Instructor&lt;br /&gt;
|label1  = &lt;br /&gt;
|data1   = &lt;br /&gt;
|header2 = &lt;br /&gt;
|label2  = &lt;br /&gt;
|data2   = &#039;&#039;&#039;尹一通&#039;&#039;&#039;&lt;br /&gt;
|header3 = &lt;br /&gt;
|label3  = Email&lt;br /&gt;
|data3   = yinyt@nju.edu.cn  &lt;br /&gt;
|header4 =&lt;br /&gt;
|label4  = office&lt;br /&gt;
|data4   = 计算机系 804&lt;br /&gt;
|header5 = &lt;br /&gt;
|label5  = &lt;br /&gt;
|data5   = &#039;&#039;&#039;刘景铖&#039;&#039;&#039;&lt;br /&gt;
|header6 = &lt;br /&gt;
|label6  = Email&lt;br /&gt;
|data6   = liu@nju.edu.cn  &lt;br /&gt;
|header7 =&lt;br /&gt;
|label7  = office&lt;br /&gt;
|data7   = 计算机系 516&lt;br /&gt;
|header8 = Class&lt;br /&gt;
|label8  = &lt;br /&gt;
|data8   = &lt;br /&gt;
|header9 =&lt;br /&gt;
|label9  = Class meeting&lt;br /&gt;
|data9   = Monday, 10am-12pm&amp;lt;br&amp;gt;&lt;br /&gt;
Wednesday (单), 10am-12pm&amp;lt;br&amp;gt;&lt;br /&gt;
仙Ⅰ-204&lt;br /&gt;
|header10=&lt;br /&gt;
|label10 = Office hour&lt;br /&gt;
|data10  = TBD &amp;lt;br&amp;gt;计算机系 804（尹一通）&amp;lt;br&amp;gt;计算机系 516（刘景铖）&lt;br /&gt;
|header11= Textbook&lt;br /&gt;
|label11 = &lt;br /&gt;
|data11  = &lt;br /&gt;
|header12=&lt;br /&gt;
|label12 = &lt;br /&gt;
|data12  = [[File:概率导论.jpeg|border|100px]]&lt;br /&gt;
|header13=&lt;br /&gt;
|label13 = &lt;br /&gt;
|data13  = &#039;&#039;&#039;概率导论&#039;&#039;&#039;（第2版·修订版）&amp;lt;br&amp;gt; Dimitri P. Bertsekas and John N. Tsitsiklis&amp;lt;br&amp;gt; 郑忠国 童行伟 译；人民邮电出版社 (2022)&lt;br /&gt;
|header14=&lt;br /&gt;
|label14 = &lt;br /&gt;
|data14  = [[File:Grimmett_probability.jpg|border|100px]]&lt;br /&gt;
|header15=&lt;br /&gt;
|label15 = &lt;br /&gt;
|data15  = &#039;&#039;&#039;Probability and Random Processes&#039;&#039;&#039; (4E) &amp;lt;br&amp;gt; Geoffrey Grimmett and David Stirzaker &amp;lt;br&amp;gt;  Oxford University Press (2020)&lt;br /&gt;
|header16=&lt;br /&gt;
|label16 = &lt;br /&gt;
|data16  = [[File:Probability_and_Computing_2ed.jpg|border|100px]]&lt;br /&gt;
|header17=&lt;br /&gt;
|label17 = &lt;br /&gt;
|data17  = &#039;&#039;&#039;Probability and Computing&#039;&#039;&#039; (2E) &amp;lt;br&amp;gt; Michael Mitzenmacher and Eli Upfal &amp;lt;br&amp;gt;   Cambridge University Press (2017)&lt;br /&gt;
|belowstyle = background:#ddf;&lt;br /&gt;
|below = &lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
This is the webpage for the &#039;&#039;Probability Theory and Mathematical Statistics&#039;&#039; (概率论与数理统计) class of Spring 2025. Students who take this class should check this page periodically for content updates and new announcements. &lt;br /&gt;
&lt;br /&gt;
= Announcement =&lt;br /&gt;
* TBA&lt;br /&gt;
&lt;br /&gt;
= Course info =&lt;br /&gt;
* &#039;&#039;&#039;Instructor &#039;&#039;&#039;: &lt;br /&gt;
:* [http://tcs.nju.edu.cn/yinyt/ 尹一通]：[mailto:yinyt@nju.edu.cn &amp;lt;yinyt@nju.edu.cn&amp;gt;]，计算机系 804 &lt;br /&gt;
:* [https://liuexp.github.io 刘景铖]：[mailto:liu@nju.edu.cn &amp;lt;liu@nju.edu.cn&amp;gt;]，计算机系 516 &lt;br /&gt;
* &#039;&#039;&#039;Teaching assistant&#039;&#039;&#039;:&lt;br /&gt;
** [https://sites.google.com/view/xinyuanzhang 张昕渊]：[mailto:zhangxy@smail.nju.edu.cn &amp;lt;zhangxy@smail.nju.edu.cn&amp;gt;]，计算机系 410&lt;br /&gt;
** 邹宗瑞：[mailto:zou.zongrui@smail.nju.edu.cn &amp;lt;zou.zongrui@smail.nju.edu.cn&amp;gt;]，计算机系 410&lt;br /&gt;
* &#039;&#039;&#039;Class meeting&#039;&#039;&#039;:&lt;br /&gt;
** 周一：10am-12pm，仙Ⅰ-204&lt;br /&gt;
** 周三（单）：10am-12pm，仙Ⅰ-204&lt;br /&gt;
* &#039;&#039;&#039;Office hour&#039;&#039;&#039;: &lt;br /&gt;
:* TBD, 计算机系 804（尹一通）&lt;br /&gt;
:* TBD, 计算机系 516（刘景铖）&lt;br /&gt;
:* &#039;&#039;&#039;QQ群&#039;&#039;&#039;: 974742320（申请加入需提供姓名、院系、学号）&lt;br /&gt;
&lt;br /&gt;
= Syllabus =&lt;br /&gt;
课程内容分为三大部分：&lt;br /&gt;
* &#039;&#039;&#039;经典概率论&#039;&#039;&#039;：概率空间、随机变量及其数字特征、多维与连续随机变量、极限定理等内容&lt;br /&gt;
* &#039;&#039;&#039;概率与计算&#039;&#039;&#039;：测度集中现象 (concentration of measure)、概率法 (the probabilistic method)、离散随机过程的相关专题&lt;br /&gt;
* &#039;&#039;&#039;数理统计&#039;&#039;&#039;：参数估计、假设检验、贝叶斯估计、线性回归等统计推断等概念&lt;br /&gt;
&lt;br /&gt;
对于第一和第二部分，要求清楚掌握基本概念，深刻理解关键的现象与规律以及背后的原理，并可以灵活运用所学方法求解相关问题。对于第三部分，要求熟悉数理统计的若干基本概念，以及典型的统计模型和统计推断问题。&lt;br /&gt;
&lt;br /&gt;
经过本课程的训练，力求使学生能够熟悉掌握概率的语言，并会利用概率思维来理解客观世界并对其建模，以及驾驭概率的数学工具来分析和求解专业问题。&lt;br /&gt;
&lt;br /&gt;
=== 教材与参考书 Course Materials ===&lt;br /&gt;
* &#039;&#039;&#039;[BT]&#039;&#039;&#039; 概率导论（第2版·修订版），[美]伯特瑟卡斯（Dimitri P.Bertsekas）[美]齐齐克利斯（John N.Tsitsiklis）著，郑忠国 童行伟 译，人民邮电出版社（2022）。&lt;br /&gt;
* &#039;&#039;&#039;[GS]&#039;&#039;&#039; &#039;&#039;Probability and Random Processes&#039;&#039;, by Geoffrey Grimmett and David Stirzaker; Oxford University Press; 4th edition (2020).&lt;br /&gt;
* &#039;&#039;&#039;[MU]&#039;&#039;&#039; &#039;&#039;Probability and Computing: Randomization and Probabilistic Techniques in Algorithms and Data Analysis&#039;&#039;, by Michael Mitzenmacher, Eli Upfal; Cambridge University Press; 2nd edition (2017).&lt;br /&gt;
&lt;br /&gt;
=== 成绩 Grading Policy ===&lt;br /&gt;
* 课程成绩：本课程将会有若干次作业和一次期末考试。最终成绩将由平时作业成绩和期末考试成绩综合得出。&lt;br /&gt;
* 迟交：如果有特殊的理由，无法按时完成作业，请提前联系授课老师，给出正当理由。否则迟交的作业将不被接受。&lt;br /&gt;
&lt;br /&gt;
=== &amp;lt;font color=red&amp;gt; 学术诚信 Academic Integrity &amp;lt;/font&amp;gt;===&lt;br /&gt;
学术诚信是所有从事学术活动的学生和学者最基本的职业道德底线，本课程将不遗余力的维护学术诚信规范，违反这一底线的行为将不会被容忍。&lt;br /&gt;
&lt;br /&gt;
作业完成的原则：署你名字的工作必须是你个人的贡献。在完成作业的过程中，允许讨论，前提是讨论的所有参与者均处于同等完成度。但关键想法的执行、以及作业文本的写作必须独立完成，并在作业中致谢（acknowledge）所有参与讨论的人。符合规则的讨论与致谢将不会影响得分。不允许其他任何形式的合作——尤其是与已经完成作业的同学“讨论”。&lt;br /&gt;
&lt;br /&gt;
本课程将对剽窃行为采取零容忍的态度。在完成作业过程中，对他人工作（出版物、互联网资料、其他人的作业等）直接的文本抄袭和对关键思想、关键元素的抄袭，按照 [http://www.acm.org/publications/policies/plagiarism_policy ACM Policy on Plagiarism]的解释，都将视为剽窃。剽窃者成绩将被取消。如果发现互相抄袭行为，&amp;lt;font color=red&amp;gt; 抄袭和被抄袭双方的成绩都将被取消&amp;lt;/font&amp;gt;。因此请主动防止自己的作业被他人抄袭。&lt;br /&gt;
&lt;br /&gt;
学术诚信影响学生个人的品行，也关乎整个教育系统的正常运转。为了一点分数而做出学术不端的行为，不仅使自己沦为一个欺骗者，也使他人的诚实努力失去意义。让我们一起努力维护一个诚信的环境。&lt;br /&gt;
&lt;br /&gt;
= Assignments =&lt;br /&gt;
TBD&lt;br /&gt;
&lt;br /&gt;
= Lectures =&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2024/Intro.pdf 课程简介]&lt;br /&gt;
&lt;br /&gt;
= Concepts =&lt;br /&gt;
* [https://plato.stanford.edu/entries/probability-interpret/ Interpretations of probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/History_of_probability History of probability]&lt;br /&gt;
* Example problems:&lt;br /&gt;
** [https://dornsifecms.usc.edu/assets/sites/520/docs/VonNeumann-ams12p36-38.pdf von Neumann&#039;s Bernoulli factory] and other [https://peteroupc.github.io/bernoulli.html Bernoulli factory algorithms]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Boy_or_Girl_paradox Boy or Girl paradox]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Monty_Hall_problem Monty Hall problem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bertrand_paradox_(probability) Bertrand paradox]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hard_spheres Hard spheres model] and [https://en.wikipedia.org/wiki/Ising_model Ising model]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/PageRank &#039;&#039;PageRank&#039;&#039;] and stationary [https://en.wikipedia.org/wiki/Random_walk random walk]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Diffusion_process Diffusion process] and [https://en.wikipedia.org/wiki/Diffusion_model diffusion model]&lt;br /&gt;
*[https://en.wikipedia.org/wiki/Probability_space Probability space]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Sample_space Sample space]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Event_(probability_theory) Event] and [https://en.wikipedia.org/wiki/Σ-algebra &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-algebra]&lt;br /&gt;
** Kolmogorov&#039;s [https://en.wikipedia.org/wiki/Probability_axioms axioms of probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Discrete_uniform_distribution Classical] and [https://en.wikipedia.org/wiki/Geometric_probability goemetric probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Boole%27s_inequality Union bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Inclusion%E2%80%93exclusion_principle Inclusion-Exclusion principle]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Boole%27s_inequality#Bonferroni_inequalities Bonferroni inequalities]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Conditional_probability Conditional probability]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chain_rule_(probability) Chain rule]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_total_probability Law of total probability]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bayes%27_theorem Bayes&#039; law]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Independence_(probability_theory) Independence] &lt;br /&gt;
** [https://en.wikipedia.org/wiki/Pairwise_independence Pairwise independence]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Random_variable Random variable]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cumulative_distribution_function Cumulative distribution function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_mass_function Probability mass function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_density_function Probability density function]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Multivariate_random_variable Random vector]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Joint_probability_distribution Joint probability distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Conditional_probability_distribution Conditional probability distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Marginal_distribution Marginal distribution]&lt;br /&gt;
* Some &#039;&#039;&#039;discrete&#039;&#039;&#039; probability distributions&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bernoulli_trial Bernoulli trial] and [https://en.wikipedia.org/wiki/Bernoulli_distribution Bernoulli distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Discrete_uniform_distribution Discrete uniform distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Binomial_distribution Binomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Geometric_distribution Geometric distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Negative_binomial_distribution Negative binomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hypergeometric_distribution Hypergeometric distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Poisson_distribution Poisson distribution]&lt;br /&gt;
** and [https://en.wikipedia.org/wiki/List_of_probability_distributions#Discrete_distributions others]&lt;br /&gt;
* Balls into bins model&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Multinomial_distribution Multinomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Birthday_problem Birthday problem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Coupon_collector%27s_problem Coupon collector]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Balls_into_bins_problem Occupancy problem]&lt;br /&gt;
* Random graphs&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Erd%C5%91s%E2%80%93R%C3%A9nyi_model Erdős–Rényi random graph model]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Galton%E2%80%93Watson_process Galton–Watson branching process]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Expected_value Expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_the_unconscious_statistician Law of the unconscious statistician, &#039;&#039;LOTUS&#039;&#039;]&lt;br /&gt;
** [https://dlsun.github.io/probability/linearity.html Linearity of expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Conditional_expectation Conditional expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_total_expectation Law of total expectation]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Markov%27s_inequality Markov&#039;s inequality]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Chebyshev%27s_inequality Chebyshev&#039;s inequality]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Moment_(mathematics) Moment] related&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Central_moment Central moment]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Variance Variance]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Standard_deviation Standard deviation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Covariance Covariance]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Correlation Correlation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Correlation_does_not_imply_causation Correlation does not imply causation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Skewness Skewness]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Kurtosis Kurtosis]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Moment_problem Moment problem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Measure_(mathematics) Measure]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Borel_set Borel set]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Lebesgue_integration Lebesgue integration]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Measurable_function Measurable function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cantor_set Cantor set]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Non-measurable_set Non-measurable set]&lt;br /&gt;
*[https://en.wikipedia.org/wiki/Probability_density_function Probability density function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_distribution#Absolutely_continuous_probability_distribution Continuous probability distribution]&lt;br /&gt;
**[https://en.wikipedia.org/wiki/Conditional_probability_distribution#Conditional_continuous_distributions Conditional continuous distributions]&lt;br /&gt;
**[https://en.wikipedia.org/wiki/Convolution_of_probability_distributions Convolution of probability distributions] and [https://en.wikipedia.org/wiki/List_of_convolutions_of_probability_distributions List of convolutions of probability distributions]&lt;br /&gt;
* Some &#039;&#039;&#039;continuous&#039;&#039;&#039; probability distributions&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Continuous_uniform_distribution Continuous uniform distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Exponential_distribution Exponential distribution] and [https://en.wikipedia.org/wiki/Poisson_point_process Poisson point process]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Normal_distribution Normal (Gaussion) distribution]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Gaussian_function Gaussian function] and [https://en.wikipedia.org/wiki/Gaussian_integral Gaussian integral]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Standard_normal_table Standard normal table]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/68%E2%80%9395%E2%80%9399.7_rule 68–95–99.7 rule]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Sum_of_normally_distributed_random_variables Sum of normally distributed random variables]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Multivariate_normal_distribution Multivariate normal distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chi-squared_distribution Chi-squared distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Gamma_distribution Gamma distribution] and [https://en.wikipedia.org/wiki/Gamma_function Gamma function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Beta_distribution Beta distribution] and [https://en.wikipedia.org/wiki/Beta_function Beta function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cauchy_distribution Cauchy distribution]&lt;br /&gt;
** and [https://en.wikipedia.org/wiki/List_of_probability_distributions#Absolutely_continuous_distributions others]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Inverse_transform_sampling Inverse transform sampling]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stochastic_dominance Stochastic dominance] and [https://en.wikipedia.org/wiki/Coupling_(probability) Coupling] &lt;br /&gt;
* [https://en.wikipedia.org/wiki/Moment-generating_function Moment-generating function]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Convergence_of_random_variables Convergence of random variables]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Pointwise_convergence Pointwise convergence]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Convergence_of_measures#Weak_convergence_of_measures Weak convergence of measures]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Convergence_in_measure Convergence in measure]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Skorokhod%27s_representation_theorem Skorokhod&#039;s representation theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Continuous_mapping_theorem Continuous mapping theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Borel%E2%80%93Cantelli_lemma Borel–Cantelli lemma] and [https://en.wikipedia.org/wiki/Zero%E2%80%93one_law zero-one laws]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Law_of_large_numbers Law of large numbers]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Central_limit_theorem Central limit theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/De_Moivre%E2%80%93Laplace_theorem De Moivre–Laplace theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Berry%E2%80%93Esseen_theorem Berry–Esseen theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Characteristic_function_(probability_theory) Characteristic function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Lévy%27s_continuity_theorem Lévy&#039;s continuity theorem]&lt;br /&gt;
* Concentration of measures:&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Moment-generating_function Moment-generating function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chernoff_bound Chernoff bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hoeffding%27s_inequality Hoeffding&#039;s bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Doob_martingale#McDiarmid&#039;s_inequality McDiarmid&#039;s inequality]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Doob_martingale Doob martingale] and [https://en.wikipedia.org/wiki/Azuma%27s_inequality Azuma&#039;s inequality]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Sub-Gaussian_distribution Sub-Gaussian distribution]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stochastic_process Stochastic process]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stopping_time Stopping time]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Martingale_(probability_theory) Martingale]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Optional_stopping_theorem Optional stopping theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Wald%27s_equation Wald&#039;s equation]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Markov_chain Markov chain]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Markov_property Markov property] and [https://en.wikipedia.org/wiki/Memorylessness memorylessness]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Stochastic_matrix Stochastic matrix]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Stationary_distribution Stationary distribution]&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)&amp;diff=12849</id>
		<title>概率论与数理统计 (Spring 2025)</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)&amp;diff=12849"/>
		<updated>2025-02-16T07:30:11Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: Created page with &amp;quot;{{Infobox |name         = Infobox |bodystyle    =  |title        = &amp;lt;font size=3&amp;gt;&amp;#039;&amp;#039;&amp;#039;概率论与数理统计&amp;#039;&amp;#039;&amp;#039;&amp;lt;br&amp;gt; &amp;#039;&amp;#039;&amp;#039;Probability Theory&amp;#039;&amp;#039;&amp;#039; &amp;lt;br&amp;gt; &amp;amp; &amp;#039;&amp;#039;&amp;#039;Mathematical Statistics&amp;#039;&amp;#039;&amp;#039;&amp;lt;/font&amp;gt; |titlestyle   =   |image        =  |imagestyle   =  |caption      =  |captionstyle =  |headerstyle  = background:#ccf; |labelstyle   = background:#ddf; |datastyle    =   |header1 =Instructor |label1  =  |data1   =  |header2 =  |label2  =  |data2   = &amp;#039;&amp;#039;&amp;#039;尹一通&amp;#039;&amp;#039;&amp;#039; |header3 =  |label3  = Em...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Infobox&lt;br /&gt;
|name         = Infobox&lt;br /&gt;
|bodystyle    = &lt;br /&gt;
|title        = &amp;lt;font size=3&amp;gt;&#039;&#039;&#039;概率论与数理统计&#039;&#039;&#039;&amp;lt;br&amp;gt;&lt;br /&gt;
&#039;&#039;&#039;Probability Theory&#039;&#039;&#039; &amp;lt;br&amp;gt; &amp;amp; &#039;&#039;&#039;Mathematical Statistics&#039;&#039;&#039;&amp;lt;/font&amp;gt;&lt;br /&gt;
|titlestyle   = &lt;br /&gt;
&lt;br /&gt;
|image        = &lt;br /&gt;
|imagestyle   = &lt;br /&gt;
|caption      = &lt;br /&gt;
|captionstyle = &lt;br /&gt;
|headerstyle  = background:#ccf;&lt;br /&gt;
|labelstyle   = background:#ddf;&lt;br /&gt;
|datastyle    = &lt;br /&gt;
&lt;br /&gt;
|header1 =Instructor&lt;br /&gt;
|label1  = &lt;br /&gt;
|data1   = &lt;br /&gt;
|header2 = &lt;br /&gt;
|label2  = &lt;br /&gt;
|data2   = &#039;&#039;&#039;尹一通&#039;&#039;&#039;&lt;br /&gt;
|header3 = &lt;br /&gt;
|label3  = Email&lt;br /&gt;
|data3   = yinyt@nju.edu.cn  &lt;br /&gt;
|header4 =&lt;br /&gt;
|label4  = office&lt;br /&gt;
|data4   = 计算机系 804&lt;br /&gt;
|header5 = &lt;br /&gt;
|label5  = &lt;br /&gt;
|data5   = &#039;&#039;&#039;刘景铖&#039;&#039;&#039;&lt;br /&gt;
|header6 = &lt;br /&gt;
|label6  = Email&lt;br /&gt;
|data6   = liu@nju.edu.cn  &lt;br /&gt;
|header7 =&lt;br /&gt;
|label7  = office&lt;br /&gt;
|data7   = 计算机系 516&lt;br /&gt;
|header8 = Class&lt;br /&gt;
|label8  = &lt;br /&gt;
|data8   = &lt;br /&gt;
|header9 =&lt;br /&gt;
|label9  = Class meeting&lt;br /&gt;
|data9   = Monday, 10am-12pm&amp;lt;br&amp;gt;&lt;br /&gt;
Wednesday (单), 10am-12pm&amp;lt;br&amp;gt;&lt;br /&gt;
仙Ⅰ-204&lt;br /&gt;
|header10=&lt;br /&gt;
|label10 = Office hour&lt;br /&gt;
|data10  = TBD &amp;lt;br&amp;gt;计算机系 804（尹一通）&amp;lt;br&amp;gt;计算机系 516（刘景铖）&lt;br /&gt;
|header11= Textbook&lt;br /&gt;
|label11 = &lt;br /&gt;
|data11  = &lt;br /&gt;
|header12=&lt;br /&gt;
|label12 = &lt;br /&gt;
|data12  = [[File:概率导论.jpeg|border|100px]]&lt;br /&gt;
|header13=&lt;br /&gt;
|label13 = &lt;br /&gt;
|data13  = &#039;&#039;&#039;概率导论&#039;&#039;&#039;（第2版·修订版）&amp;lt;br&amp;gt; Dimitri P. Bertsekas and John N. Tsitsiklis&amp;lt;br&amp;gt; 郑忠国 童行伟 译；人民邮电出版社 (2022)&lt;br /&gt;
|header14=&lt;br /&gt;
|label14 = &lt;br /&gt;
|data14  = [[File:Grimmett_probability.jpg|border|100px]]&lt;br /&gt;
|header15=&lt;br /&gt;
|label15 = &lt;br /&gt;
|data15  = &#039;&#039;&#039;Probability and Random Processes&#039;&#039;&#039; (4E) &amp;lt;br&amp;gt; Geoffrey Grimmett and David Stirzaker &amp;lt;br&amp;gt;  Oxford University Press (2020)&lt;br /&gt;
|header16=&lt;br /&gt;
|label16 = &lt;br /&gt;
|data16  = [[File:Probability_and_Computing_2ed.jpg|border|100px]]&lt;br /&gt;
|header17=&lt;br /&gt;
|label17 = &lt;br /&gt;
|data17  = &#039;&#039;&#039;Probability and Computing&#039;&#039;&#039; (2E) &amp;lt;br&amp;gt; Michael Mitzenmacher and Eli Upfal &amp;lt;br&amp;gt;   Cambridge University Press (2017)&lt;br /&gt;
|belowstyle = background:#ddf;&lt;br /&gt;
|below = &lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
This is the webpage for the &#039;&#039;Probability Theory and Mathematical Statistics&#039;&#039; (概率论与数理统计) class of Spring 2025. Students who take this class should check this page periodically for content updates and new announcements. &lt;br /&gt;
&lt;br /&gt;
= Announcement =&lt;br /&gt;
* TBA&lt;br /&gt;
&lt;br /&gt;
= Course info =&lt;br /&gt;
* &#039;&#039;&#039;Instructor &#039;&#039;&#039;: &lt;br /&gt;
:* [http://tcs.nju.edu.cn/yinyt/ 尹一通]：[mailto:yinyt@nju.edu.cn &amp;lt;yinyt@nju.edu.cn&amp;gt;]，计算机系 804 &lt;br /&gt;
:* [https://liuexp.github.io 刘景铖]：[mailto:liu@nju.edu.cn &amp;lt;liu@nju.edu.cn&amp;gt;]，计算机系 516 &lt;br /&gt;
* &#039;&#039;&#039;Teaching assistant&#039;&#039;&#039;:&lt;br /&gt;
** [https://sites.google.com/view/xinyuanzhang 张昕渊]：[mailto:zhangxy@smail.nju.edu.cn &amp;lt;zhangxy@smail.nju.edu.cn&amp;gt;]，计算机系 410&lt;br /&gt;
** 邹宗瑞：[mailto:zou.zongrui@smail.nju.edu.cn &amp;lt;zou.zongrui@smail.nju.edu.cn&amp;gt;]，计算机系 410&lt;br /&gt;
* &#039;&#039;&#039;Class meeting&#039;&#039;&#039;:&lt;br /&gt;
** 周一：10am-12pm，仙Ⅰ-204&lt;br /&gt;
** 周三（单）：10am-12pm，仙Ⅰ-204&lt;br /&gt;
* &#039;&#039;&#039;Office hour&#039;&#039;&#039;: &lt;br /&gt;
:* Monday, 4pm-5pm, 计算机系 804（尹一通）&lt;br /&gt;
:* Tuesday, 3pm-4pm, 计算机系 516（刘景铖）&lt;br /&gt;
:* &#039;&#039;&#039;QQ群&#039;&#039;&#039;: 629856946（申请加入需提供姓名、院系、学号）&lt;br /&gt;
&lt;br /&gt;
= Syllabus =&lt;br /&gt;
课程内容分为三大部分：&lt;br /&gt;
* &#039;&#039;&#039;经典概率论&#039;&#039;&#039;：概率空间、随机变量及其数字特征、多维与连续随机变量、极限定理等内容&lt;br /&gt;
* &#039;&#039;&#039;概率与计算&#039;&#039;&#039;：测度集中现象 (concentration of measure)、概率法 (the probabilistic method)、离散随机过程的相关专题&lt;br /&gt;
* &#039;&#039;&#039;数理统计&#039;&#039;&#039;：参数估计、假设检验、贝叶斯估计、线性回归等统计推断等概念&lt;br /&gt;
&lt;br /&gt;
对于第一和第二部分，要求清楚掌握基本概念，深刻理解关键的现象与规律以及背后的原理，并可以灵活运用所学方法求解相关问题。对于第三部分，要求熟悉数理统计的若干基本概念，以及典型的统计模型和统计推断问题。&lt;br /&gt;
&lt;br /&gt;
经过本课程的训练，力求使学生能够熟悉掌握概率的语言，并会利用概率思维来理解客观世界并对其建模，以及驾驭概率的数学工具来分析和求解专业问题。&lt;br /&gt;
&lt;br /&gt;
=== 教材与参考书 Course Materials ===&lt;br /&gt;
* &#039;&#039;&#039;[BT]&#039;&#039;&#039; 概率导论（第2版·修订版），[美]伯特瑟卡斯（Dimitri P.Bertsekas）[美]齐齐克利斯（John N.Tsitsiklis）著，郑忠国 童行伟 译，人民邮电出版社（2022）。&lt;br /&gt;
* &#039;&#039;&#039;[GS]&#039;&#039;&#039; &#039;&#039;Probability and Random Processes&#039;&#039;, by Geoffrey Grimmett and David Stirzaker; Oxford University Press; 4th edition (2020).&lt;br /&gt;
* &#039;&#039;&#039;[MU]&#039;&#039;&#039; &#039;&#039;Probability and Computing: Randomization and Probabilistic Techniques in Algorithms and Data Analysis&#039;&#039;, by Michael Mitzenmacher, Eli Upfal; Cambridge University Press; 2nd edition (2017).&lt;br /&gt;
&lt;br /&gt;
=== 成绩 Grading Policy ===&lt;br /&gt;
* 课程成绩：本课程将会有若干次作业和一次期末考试。最终成绩将由平时作业成绩和期末考试成绩综合得出。&lt;br /&gt;
* 迟交：如果有特殊的理由，无法按时完成作业，请提前联系授课老师，给出正当理由。否则迟交的作业将不被接受。&lt;br /&gt;
&lt;br /&gt;
=== &amp;lt;font color=red&amp;gt; 学术诚信 Academic Integrity &amp;lt;/font&amp;gt;===&lt;br /&gt;
学术诚信是所有从事学术活动的学生和学者最基本的职业道德底线，本课程将不遗余力的维护学术诚信规范，违反这一底线的行为将不会被容忍。&lt;br /&gt;
&lt;br /&gt;
作业完成的原则：署你名字的工作必须是你个人的贡献。在完成作业的过程中，允许讨论，前提是讨论的所有参与者均处于同等完成度。但关键想法的执行、以及作业文本的写作必须独立完成，并在作业中致谢（acknowledge）所有参与讨论的人。符合规则的讨论与致谢将不会影响得分。不允许其他任何形式的合作——尤其是与已经完成作业的同学“讨论”。&lt;br /&gt;
&lt;br /&gt;
本课程将对剽窃行为采取零容忍的态度。在完成作业过程中，对他人工作（出版物、互联网资料、其他人的作业等）直接的文本抄袭和对关键思想、关键元素的抄袭，按照 [http://www.acm.org/publications/policies/plagiarism_policy ACM Policy on Plagiarism]的解释，都将视为剽窃。剽窃者成绩将被取消。如果发现互相抄袭行为，&amp;lt;font color=red&amp;gt; 抄袭和被抄袭双方的成绩都将被取消&amp;lt;/font&amp;gt;。因此请主动防止自己的作业被他人抄袭。&lt;br /&gt;
&lt;br /&gt;
学术诚信影响学生个人的品行，也关乎整个教育系统的正常运转。为了一点分数而做出学术不端的行为，不仅使自己沦为一个欺骗者，也使他人的诚实努力失去意义。让我们一起努力维护一个诚信的环境。&lt;br /&gt;
&lt;br /&gt;
= Assignments =&lt;br /&gt;
TBD&lt;br /&gt;
&lt;br /&gt;
= Lectures =&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2024/Intro.pdf 课程简介]&lt;br /&gt;
&lt;br /&gt;
= Concepts =&lt;br /&gt;
* [https://plato.stanford.edu/entries/probability-interpret/ Interpretations of probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/History_of_probability History of probability]&lt;br /&gt;
* Example problems:&lt;br /&gt;
** [https://dornsifecms.usc.edu/assets/sites/520/docs/VonNeumann-ams12p36-38.pdf von Neumann&#039;s Bernoulli factory] and other [https://peteroupc.github.io/bernoulli.html Bernoulli factory algorithms]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Boy_or_Girl_paradox Boy or Girl paradox]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Monty_Hall_problem Monty Hall problem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bertrand_paradox_(probability) Bertrand paradox]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hard_spheres Hard spheres model] and [https://en.wikipedia.org/wiki/Ising_model Ising model]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/PageRank &#039;&#039;PageRank&#039;&#039;] and stationary [https://en.wikipedia.org/wiki/Random_walk random walk]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Diffusion_process Diffusion process] and [https://en.wikipedia.org/wiki/Diffusion_model diffusion model]&lt;br /&gt;
*[https://en.wikipedia.org/wiki/Probability_space Probability space]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Sample_space Sample space]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Event_(probability_theory) Event] and [https://en.wikipedia.org/wiki/Σ-algebra &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-algebra]&lt;br /&gt;
** Kolmogorov&#039;s [https://en.wikipedia.org/wiki/Probability_axioms axioms of probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Discrete_uniform_distribution Classical] and [https://en.wikipedia.org/wiki/Geometric_probability goemetric probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Boole%27s_inequality Union bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Inclusion%E2%80%93exclusion_principle Inclusion-Exclusion principle]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Boole%27s_inequality#Bonferroni_inequalities Bonferroni inequalities]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Conditional_probability Conditional probability]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chain_rule_(probability) Chain rule]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_total_probability Law of total probability]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bayes%27_theorem Bayes&#039; law]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Independence_(probability_theory) Independence] &lt;br /&gt;
** [https://en.wikipedia.org/wiki/Pairwise_independence Pairwise independence]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Random_variable Random variable]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cumulative_distribution_function Cumulative distribution function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_mass_function Probability mass function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_density_function Probability density function]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Multivariate_random_variable Random vector]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Joint_probability_distribution Joint probability distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Conditional_probability_distribution Conditional probability distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Marginal_distribution Marginal distribution]&lt;br /&gt;
* Some &#039;&#039;&#039;discrete&#039;&#039;&#039; probability distributions&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bernoulli_trial Bernoulli trial] and [https://en.wikipedia.org/wiki/Bernoulli_distribution Bernoulli distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Discrete_uniform_distribution Discrete uniform distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Binomial_distribution Binomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Geometric_distribution Geometric distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Negative_binomial_distribution Negative binomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hypergeometric_distribution Hypergeometric distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Poisson_distribution Poisson distribution]&lt;br /&gt;
** and [https://en.wikipedia.org/wiki/List_of_probability_distributions#Discrete_distributions others]&lt;br /&gt;
* Balls into bins model&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Multinomial_distribution Multinomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Birthday_problem Birthday problem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Coupon_collector%27s_problem Coupon collector]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Balls_into_bins_problem Occupancy problem]&lt;br /&gt;
* Random graphs&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Erd%C5%91s%E2%80%93R%C3%A9nyi_model Erdős–Rényi random graph model]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Galton%E2%80%93Watson_process Galton–Watson branching process]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Expected_value Expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_the_unconscious_statistician Law of the unconscious statistician, &#039;&#039;LOTUS&#039;&#039;]&lt;br /&gt;
** [https://dlsun.github.io/probability/linearity.html Linearity of expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Conditional_expectation Conditional expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_total_expectation Law of total expectation]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Markov%27s_inequality Markov&#039;s inequality]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Chebyshev%27s_inequality Chebyshev&#039;s inequality]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Moment_(mathematics) Moment] related&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Central_moment Central moment]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Variance Variance]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Standard_deviation Standard deviation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Covariance Covariance]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Correlation Correlation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Correlation_does_not_imply_causation Correlation does not imply causation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Skewness Skewness]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Kurtosis Kurtosis]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Moment_problem Moment problem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Measure_(mathematics) Measure]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Borel_set Borel set]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Lebesgue_integration Lebesgue integration]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Measurable_function Measurable function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cantor_set Cantor set]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Non-measurable_set Non-measurable set]&lt;br /&gt;
*[https://en.wikipedia.org/wiki/Probability_density_function Probability density function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_distribution#Absolutely_continuous_probability_distribution Continuous probability distribution]&lt;br /&gt;
**[https://en.wikipedia.org/wiki/Conditional_probability_distribution#Conditional_continuous_distributions Conditional continuous distributions]&lt;br /&gt;
**[https://en.wikipedia.org/wiki/Convolution_of_probability_distributions Convolution of probability distributions] and [https://en.wikipedia.org/wiki/List_of_convolutions_of_probability_distributions List of convolutions of probability distributions]&lt;br /&gt;
* Some &#039;&#039;&#039;continuous&#039;&#039;&#039; probability distributions&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Continuous_uniform_distribution Continuous uniform distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Exponential_distribution Exponential distribution] and [https://en.wikipedia.org/wiki/Poisson_point_process Poisson point process]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Normal_distribution Normal (Gaussion) distribution]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Gaussian_function Gaussian function] and [https://en.wikipedia.org/wiki/Gaussian_integral Gaussian integral]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Standard_normal_table Standard normal table]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/68%E2%80%9395%E2%80%9399.7_rule 68–95–99.7 rule]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Sum_of_normally_distributed_random_variables Sum of normally distributed random variables]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Multivariate_normal_distribution Multivariate normal distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chi-squared_distribution Chi-squared distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Gamma_distribution Gamma distribution] and [https://en.wikipedia.org/wiki/Gamma_function Gamma function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Beta_distribution Beta distribution] and [https://en.wikipedia.org/wiki/Beta_function Beta function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cauchy_distribution Cauchy distribution]&lt;br /&gt;
** and [https://en.wikipedia.org/wiki/List_of_probability_distributions#Absolutely_continuous_distributions others]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Inverse_transform_sampling Inverse transform sampling]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stochastic_dominance Stochastic dominance] and [https://en.wikipedia.org/wiki/Coupling_(probability) Coupling] &lt;br /&gt;
* [https://en.wikipedia.org/wiki/Moment-generating_function Moment-generating function]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Convergence_of_random_variables Convergence of random variables]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Pointwise_convergence Pointwise convergence]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Convergence_of_measures#Weak_convergence_of_measures Weak convergence of measures]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Convergence_in_measure Convergence in measure]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Skorokhod%27s_representation_theorem Skorokhod&#039;s representation theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Continuous_mapping_theorem Continuous mapping theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Borel%E2%80%93Cantelli_lemma Borel–Cantelli lemma] and [https://en.wikipedia.org/wiki/Zero%E2%80%93one_law zero-one laws]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Law_of_large_numbers Law of large numbers]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Central_limit_theorem Central limit theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/De_Moivre%E2%80%93Laplace_theorem De Moivre–Laplace theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Berry%E2%80%93Esseen_theorem Berry–Esseen theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Characteristic_function_(probability_theory) Characteristic function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Lévy%27s_continuity_theorem Lévy&#039;s continuity theorem]&lt;br /&gt;
* Concentration of measures:&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Moment-generating_function Moment-generating function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chernoff_bound Chernoff bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hoeffding%27s_inequality Hoeffding&#039;s bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Doob_martingale#McDiarmid&#039;s_inequality McDiarmid&#039;s inequality]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Doob_martingale Doob martingale] and [https://en.wikipedia.org/wiki/Azuma%27s_inequality Azuma&#039;s inequality]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Sub-Gaussian_distribution Sub-Gaussian distribution]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stochastic_process Stochastic process]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stopping_time Stopping time]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Martingale_(probability_theory) Martingale]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Optional_stopping_theorem Optional stopping theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Wald%27s_equation Wald&#039;s equation]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Markov_chain Markov chain]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Markov_property Markov property] and [https://en.wikipedia.org/wiki/Memorylessness memorylessness]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Stochastic_matrix Stochastic matrix]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Stationary_distribution Stationary distribution]&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=Main_Page&amp;diff=12847</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=Main_Page&amp;diff=12847"/>
		<updated>2025-02-16T07:26:07Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: /* Home Pages for Courses and Seminars */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;This is a course/seminar wiki run by the [http://tcs.nju.edu.cn theory group] in the Department of Computer Science and Technology at Nanjing University.&lt;br /&gt;
&lt;br /&gt;
== Home Pages for Courses and Seminars==&lt;br /&gt;
*[[高级算法 (Fall 2024)|高级算法 Advanced Algorithms (Fall 2024)]]&lt;br /&gt;
&lt;br /&gt;
*[[高级算法 (Spring 2025)|高级算法 Advanced Algorithms (Spring 2025 苏州校区)]]&lt;br /&gt;
&lt;br /&gt;
* [[概率论与数理统计 (Spring 2025) | 概率论与数理统计 Probability Theory (Spring 2025)]]&lt;br /&gt;
&lt;br /&gt;
*[[Theory Seminar|理论计算机科学讨论班]]&lt;br /&gt;
&lt;br /&gt;
*[[Study Group|理论计算机科学学习小组]]&lt;br /&gt;
&lt;br /&gt;
;Past courses&lt;br /&gt;
&lt;br /&gt;
* Advanced Algorithms: [[高级算法 (Fall 2024)|Fall 2024]], [[高级算法 (Fall 2023)|Fall 2023]], [[高级算法 (Fall 2022)|Fall 2022]], [[高级算法 (Fall 2021)|Fall 2021]], [[高级算法 (Fall 2020)|Fall 2020]], [[高级算法 (Fall 2019)|Fall 2019]], [[高级算法 (Fall 2018)|Fall 2018]], [[高级算法 (Fall 2017)|Fall 2017]], [[随机算法 \ 高级算法 (Fall 2016)|Fall 2016]].&lt;br /&gt;
&lt;br /&gt;
*Algorithm Design and Analysis: [https://tcs.nju.edu.cn/shili/courses/2024spring-algo/ Spring 2024]&lt;br /&gt;
&lt;br /&gt;
* Combinatorics: [[组合数学 (Spring 2024)|Spring 2024]], [[组合数学 (Spring 2023)|Spring 2023]], [[组合数学 (Fall 2019)|Fall 2019]], [[组合数学 (Fall 2017)|Fall 2017]], [[组合数学 (Fall 2016)|Fall 2016]], [[组合数学 (Fall 2015)|Fall 2015]], [[组合数学 (Spring 2014)|Spring 2014]], [[组合数学 (Spring 2013)|Spring 2013]], [[组合数学 (Fall 2011)|Fall 2011]], [[Combinatorics (Fall 2010)|Fall 2010]].&lt;br /&gt;
&lt;br /&gt;
* Computational Complexity: [[计算复杂性 (Spring 2025)|Spring 2025]], [[计算复杂性 (Spring 2024)|Spring 2024]], [[计算复杂性 (Spring 2023)|Spring 2023]], [[计算复杂性 (Fall 2019)|Fall 2019]], [[计算复杂性 (Fall 2018)|Fall 2018]].&lt;br /&gt;
&lt;br /&gt;
* Numerical Method: [[计算方法 Numerical method (Spring 2024)|Spring 2024]], [[计算方法 Numerical method (Spring 2023)|Spring 2023]], [https://liuexp.github.io/numerical.html Spring 2022].&lt;br /&gt;
&lt;br /&gt;
* Probability Theory: [[概率论与数理统计 (Spring 2024)|Spring 2024]], [[概率论与数理统计 (Spring 2023)|Spring 2023]].&lt;br /&gt;
&lt;br /&gt;
* Quantum Computation: [[量子计算 (Spring 2022)|Spring 2022]], [[量子计算 (Spring 2021)|Spring 2021]], [[量子计算 (Fall 2019)|Fall 2019]].&lt;br /&gt;
&lt;br /&gt;
* Randomized Algorithms:  [[随机算法 (Fall 2015)|Fall 2015]], [[随机算法 (Spring 2014)|Spring 2014]], [[随机算法 (Spring 2013)|Spring 2013]], [[随机算法 (Fall 2011)|Fall 2011]], [[Randomized Algorithms (Spring 2010)|Spring 2010]].&lt;br /&gt;
&lt;br /&gt;
;Past seminars, workshops and summer schools&lt;br /&gt;
*计算理论之美暑期学校: [[计算理论之美 (Summer 2024)|2024]], [[计算理论之美 (Summer 2023)|2023]], [[计算理论之美 (Summer 2021)|2021]]&lt;br /&gt;
*[[TCSPhD2020| 理论计算机科学优秀博士生论坛2020]]&lt;br /&gt;
*[[Quantum|量子算法与物理实现研讨会]]&lt;br /&gt;
*Nanjing Theory Day: [[Theory@Nanjing 2019|2019]], [[Theory@Nanjing 2018|2018]], [[Theory@Nanjing 2017|2017]]&lt;br /&gt;
*[[\Delta Seminar on Logic, Philosophy, and Computer Science|Δ Seminar on Logic, Philosophy, and Computer Science]]&lt;br /&gt;
*[[近似算法讨论班 (Fall 2011)|近似算法 Approximation Algorithms, Fall 2011.]]&lt;br /&gt;
&lt;br /&gt;
; 其它链接&lt;br /&gt;
* [[General Circulation(Fall 2024)|大气环流 General Circulation of the Atmosphere, Fall 2024]]&lt;br /&gt;
* [[General Circulation(Fall 2023)|大气环流 General Circulation of the Atmosphere, Fall 2023]]&lt;br /&gt;
&lt;br /&gt;
* [[概率论 (Summer 2014)| 概率与计算 (上海交大 Summer 2014)]]&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=Main_Page&amp;diff=12846</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=Main_Page&amp;diff=12846"/>
		<updated>2025-02-16T07:25:41Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: /* Home Pages for Courses and Seminars */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;This is a course/seminar wiki run by the [http://tcs.nju.edu.cn theory group] in the Department of Computer Science and Technology at Nanjing University.&lt;br /&gt;
&lt;br /&gt;
== Home Pages for Courses and Seminars==&lt;br /&gt;
*[[高级算法 (Fall 2024)|高级算法 Advanced Algorithms (Fall 2024)]]&lt;br /&gt;
&lt;br /&gt;
*[[高级算法 (Spring 2025)|高级算法 Advanced Algorithms (Spring 2025 苏州校区)]]&lt;br /&gt;
&lt;br /&gt;
* [[概率论与数理统计 (Spring 2025) | 概率论与数理统计 (Spring 2025)]]&lt;br /&gt;
&lt;br /&gt;
*[[Theory Seminar|理论计算机科学讨论班]]&lt;br /&gt;
&lt;br /&gt;
*[[Study Group|理论计算机科学学习小组]]&lt;br /&gt;
&lt;br /&gt;
;Past courses&lt;br /&gt;
&lt;br /&gt;
* Advanced Algorithms: [[高级算法 (Fall 2024)|Fall 2024]], [[高级算法 (Fall 2023)|Fall 2023]], [[高级算法 (Fall 2022)|Fall 2022]], [[高级算法 (Fall 2021)|Fall 2021]], [[高级算法 (Fall 2020)|Fall 2020]], [[高级算法 (Fall 2019)|Fall 2019]], [[高级算法 (Fall 2018)|Fall 2018]], [[高级算法 (Fall 2017)|Fall 2017]], [[随机算法 \ 高级算法 (Fall 2016)|Fall 2016]].&lt;br /&gt;
&lt;br /&gt;
*Algorithm Design and Analysis: [https://tcs.nju.edu.cn/shili/courses/2024spring-algo/ Spring 2024]&lt;br /&gt;
&lt;br /&gt;
* Combinatorics: [[组合数学 (Spring 2024)|Spring 2024]], [[组合数学 (Spring 2023)|Spring 2023]], [[组合数学 (Fall 2019)|Fall 2019]], [[组合数学 (Fall 2017)|Fall 2017]], [[组合数学 (Fall 2016)|Fall 2016]], [[组合数学 (Fall 2015)|Fall 2015]], [[组合数学 (Spring 2014)|Spring 2014]], [[组合数学 (Spring 2013)|Spring 2013]], [[组合数学 (Fall 2011)|Fall 2011]], [[Combinatorics (Fall 2010)|Fall 2010]].&lt;br /&gt;
&lt;br /&gt;
* Computational Complexity: [[计算复杂性 (Spring 2025)|Spring 2025]], [[计算复杂性 (Spring 2024)|Spring 2024]], [[计算复杂性 (Spring 2023)|Spring 2023]], [[计算复杂性 (Fall 2019)|Fall 2019]], [[计算复杂性 (Fall 2018)|Fall 2018]].&lt;br /&gt;
&lt;br /&gt;
* Numerical Method: [[计算方法 Numerical method (Spring 2024)|Spring 2024]], [[计算方法 Numerical method (Spring 2023)|Spring 2023]], [https://liuexp.github.io/numerical.html Spring 2022].&lt;br /&gt;
&lt;br /&gt;
* Probability Theory: [[概率论与数理统计 (Spring 2024)|Spring 2024]], [[概率论与数理统计 (Spring 2023)|Spring 2023]].&lt;br /&gt;
&lt;br /&gt;
* Quantum Computation: [[量子计算 (Spring 2022)|Spring 2022]], [[量子计算 (Spring 2021)|Spring 2021]], [[量子计算 (Fall 2019)|Fall 2019]].&lt;br /&gt;
&lt;br /&gt;
* Randomized Algorithms:  [[随机算法 (Fall 2015)|Fall 2015]], [[随机算法 (Spring 2014)|Spring 2014]], [[随机算法 (Spring 2013)|Spring 2013]], [[随机算法 (Fall 2011)|Fall 2011]], [[Randomized Algorithms (Spring 2010)|Spring 2010]].&lt;br /&gt;
&lt;br /&gt;
;Past seminars, workshops and summer schools&lt;br /&gt;
*计算理论之美暑期学校: [[计算理论之美 (Summer 2024)|2024]], [[计算理论之美 (Summer 2023)|2023]], [[计算理论之美 (Summer 2021)|2021]]&lt;br /&gt;
*[[TCSPhD2020| 理论计算机科学优秀博士生论坛2020]]&lt;br /&gt;
*[[Quantum|量子算法与物理实现研讨会]]&lt;br /&gt;
*Nanjing Theory Day: [[Theory@Nanjing 2019|2019]], [[Theory@Nanjing 2018|2018]], [[Theory@Nanjing 2017|2017]]&lt;br /&gt;
*[[\Delta Seminar on Logic, Philosophy, and Computer Science|Δ Seminar on Logic, Philosophy, and Computer Science]]&lt;br /&gt;
*[[近似算法讨论班 (Fall 2011)|近似算法 Approximation Algorithms, Fall 2011.]]&lt;br /&gt;
&lt;br /&gt;
; 其它链接&lt;br /&gt;
* [[General Circulation(Fall 2024)|大气环流 General Circulation of the Atmosphere, Fall 2024]]&lt;br /&gt;
* [[General Circulation(Fall 2023)|大气环流 General Circulation of the Atmosphere, Fall 2023]]&lt;br /&gt;
&lt;br /&gt;
* [[概率论 (Summer 2014)| 概率与计算 (上海交大 Summer 2014)]]&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2024)/Problem_Set_4&amp;diff=12488</id>
		<title>概率论与数理统计 (Spring 2024)/Problem Set 4</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2024)/Problem_Set_4&amp;diff=12488"/>
		<updated>2024-06-03T07:26:13Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: /* Problem 4 (LLN and CLT, 15 points + 10 points) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
*Bonus problem为附加题（选做）。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 4==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Continuous Random Variables, 30 points)==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Density function&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Determine the value of &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;f(x) = C\exp(-x-e^{-x}), x\in \mathbb{R}&amp;lt;/math&amp;gt; is a probability density function (PDF) for a continuous random variable.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Independence&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be independent and identically distributed continuous random variables with cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; and probability density function (PDF) &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt;. Find out the density functions of &amp;lt;math&amp;gt;V = \max\{X,Y\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;U = \min\{X,Y\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Correlation&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be uniformly distributed on &amp;lt;math&amp;gt;(-1,1)&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_k = \cos(k \pi X)&amp;lt;/math&amp;gt; for &amp;lt;math&amp;gt;k=1,2,\ldots,n&amp;lt;/math&amp;gt;. Are the random variables &amp;lt;math&amp;gt;Y_1, Y_2, \ldots, Y_n&amp;lt;/math&amp;gt; correlated? independent? You should prove your claim rigorously.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Expectation of random variables (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a continuous random variable with mean &amp;lt;math&amp;gt;\mu&amp;lt;/math&amp;gt; and cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X \ge 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\int_{0}^a F(x) dx = \int_{a}^{\infty} [1-F(x)] dx&amp;lt;/math&amp;gt; if and only if &amp;lt;math&amp;gt;a = \mu&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; has finite variance. Show that &amp;lt;math&amp;gt;g(a) = \mathbb{E}((X-a)^2)&amp;lt;/math&amp;gt; achieves the minimum when &amp;lt;math&amp;gt;a = \mu&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Expectation of random variables (II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X, Y&amp;lt;/math&amp;gt; be two independent and identically distributed continuous random variables with cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;. Furthermore, &amp;lt;math&amp;gt;X,Y \ge 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbb{E}[|X-Y|] = 2 \left(\mathbb{E}[X] - \int_{0}^{\infty} (1-F(x))^2 dx\right)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Conditional distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be two random variables. The joint density of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; is given by &amp;lt;math&amp;gt;f(x,y) = c(x^2 - y^2)e^{-x}&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;0\leq x &amp;lt;\infty&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-x\leq y \leq x&amp;lt;/math&amp;gt;. Here, &amp;lt;math&amp;gt;c\in \mathbb{R}_+&amp;lt;/math&amp;gt; is a constant. Find out the conditional distribution of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;, given &amp;lt;math&amp;gt;X = x&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Uniform Distribution (I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;P_i = (X_i,Y_i), 1\leq i\leq n&amp;lt;/math&amp;gt;, be independent, uniformly distributed points in the unit square &amp;lt;math&amp;gt;[0,1]^2&amp;lt;/math&amp;gt;. A point &amp;lt;math&amp;gt;P_i&amp;lt;/math&amp;gt; is called &amp;quot;peripheral&amp;quot; if, for all &amp;lt;math&amp;gt;r = 1,2,\cdots,n&amp;lt;/math&amp;gt;, either &amp;lt;math&amp;gt;X_r \leq X_i&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;Y_r \leq Y_i&amp;lt;/math&amp;gt;, or both. Find out the expected number of peripheral points.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Uniform Distribution (II)&amp;lt;/strong&amp;gt;] Derive the moment generating function of the standard uniform distribution, i.e., uniform distribution on &amp;lt;math&amp;gt;(0,1)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Exponential distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; have an exponential distribution. Show that &amp;lt;math&amp;gt;\textbf{Pr}[X&amp;gt;s+x|X&amp;gt;s] = \textbf{Pr}[X&amp;gt;x]&amp;lt;/math&amp;gt;, for &amp;lt;math&amp;gt;x,s\geq 0&amp;lt;/math&amp;gt;. This is the memoryless property. Show that the exponential distribution is the only continuous distribution with this property.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Normal distribution(I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X,Y\sim N(0,1)&amp;lt;/math&amp;gt; be two independent and identically distributed normal random variables. Let &amp;lt;math&amp;gt;Z = X-Y&amp;lt;/math&amp;gt;. Find the density function of &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;|Z|&amp;lt;/math&amp;gt; respectively.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Normal distribution(II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; have the &amp;lt;math&amp;gt;N(0,1)&amp;lt;/math&amp;gt; distribution and let &amp;lt;math&amp;gt;a&amp;gt;0&amp;lt;/math&amp;gt;. Show that the random variable &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; given by&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{equation*}&lt;br /&gt;
Y = \begin{cases}&lt;br /&gt;
X, &amp;amp; |X|&amp;lt; a \\&lt;br /&gt;
-X, &amp;amp; |X|\geq a&lt;br /&gt;
\end{cases}&lt;br /&gt;
\end{equation*}&amp;lt;/math&amp;gt;&lt;br /&gt;
has the &amp;lt;math&amp;gt;N(0,1)&amp;lt;/math&amp;gt; distribution, and find an expression for &amp;lt;math&amp;gt;\rho(a) = \textbf{Cov}(X,Y)&amp;lt;/math&amp;gt; in terms of the density function &amp;lt;math&amp;gt;\phi&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random process&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Given a real number &amp;lt;math&amp;gt;U&amp;lt;1&amp;lt;/math&amp;gt; as input of the following process, find out the expected returning value.&lt;br /&gt;
{{Theorem|&#039;&#039;Process&#039;&#039;|&lt;br /&gt;
:&#039;&#039;&#039;Input:&#039;&#039;&#039;  real numbers &amp;lt;math&amp;gt;U &amp;lt; 1&amp;lt;/math&amp;gt;;&lt;br /&gt;
----&lt;br /&gt;
:initialize &amp;lt;math&amp;gt;x = 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;count = 0&amp;lt;/math&amp;gt;;&lt;br /&gt;
:while &amp;lt;math&amp;gt; x &amp;lt; U &amp;lt;/math&amp;gt; do&lt;br /&gt;
:* choose &amp;lt;math&amp;gt;y \in (0,1)&amp;lt;/math&amp;gt; uniformly at random;&lt;br /&gt;
:* update &amp;lt;math&amp;gt;x = x + y&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;count = count + 1&amp;lt;/math&amp;gt;;&lt;br /&gt;
:return &amp;lt;math&amp;gt;count&amp;lt;/math&amp;gt;;&lt;br /&gt;
}}&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Random semicircle&amp;lt;/strong&amp;gt;] We sample &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; points within a circle &amp;lt;math&amp;gt;C=\{(x,y) \in \mathbb{R}^2 \mid x^2+y^2 \le 1\}&amp;lt;/math&amp;gt; independently and uniformly at random (i.e., the density function &amp;lt;math&amp;gt;f(x,y) \propto 1_{(x,y) \in C}&amp;lt;/math&amp;gt;). Find out the probability that they all lie within some semicircle of the original circle &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;. (Hint: you may apply the technique of change of variables, see [https://en.wikipedia.org/wiki/Random_variable#Functions_of_random_variables function of random variables] or Chapter 4.7 in [GS])&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Stochastic domination&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X, Y&amp;lt;/math&amp;gt; be continuous random variables. Show that &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; dominates &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; stochastically if and only if &amp;lt;math&amp;gt;\mathbb{E}[f(X)]\geq \mathbb{E}[f(Y)]&amp;lt;/math&amp;gt; for any non-decreasing function &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; for which the expectations exist.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Concentration of measure, 10 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Tossing coins&amp;lt;/strong&amp;gt;] We repeatedly toss a fair coin (with an equal probability of heads and tails). Let the random variable &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of throws required to obtain a total of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; heads. Show that &amp;lt;math&amp;gt;\textbf{Pr}[X &amp;gt; 2n + 2\sqrt{n\log n}]\leq O(1/n)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Chernoff vs Chebyshev&amp;lt;/strong&amp;gt;] We have a standard six-sided die. Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of times a 6 occurs in &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; throws off the die. Compare the best upper bounds on &amp;lt;math&amp;gt;\textbf{Pr}[X\geq n/4]&amp;lt;/math&amp;gt; that you can obtain using Chebyshev&#039;s inequality and Chernoff bounds.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;&amp;lt;strong&amp;gt;-th moment bound]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a random variable with expectation &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; such that moment generating function &amp;lt;math&amp;gt;\mathbf{E}[\exp(t|X|)]&amp;lt;/math&amp;gt; is finite for some &amp;lt;math&amp;gt; t &amp;gt; 0 &amp;lt;/math&amp;gt;. We can use the following two kinds of tail inequalities for &amp;lt;math&amp;gt; X &amp;lt;/math&amp;gt;: &lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&amp;lt;nowiki&amp;gt;     &amp;lt;/nowiki&amp;gt;&#039;&#039;&#039;&#039;&#039;Chernoff Bound&#039;&#039;&#039;&#039;&#039;&lt;br /&gt;
:&amp;lt;math&amp;gt;&lt;br /&gt;
\begin{align}&lt;br /&gt;
\mathbf{Pr}[|X| \geq \delta] \leq \min_{t \geq 0} \frac{\mathbf{E}[e^{t|X|}]}{e^{t\delta}}&lt;br /&gt;
\end{align}&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&#039;&#039;&#039;&#039;&#039;&amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;th-Moment Bound&#039;&#039;&#039;&#039;&#039;&lt;br /&gt;
:&amp;lt;math&amp;gt;&lt;br /&gt;
\begin{align}&lt;br /&gt;
\mathbf{Pr}[|X| \geq \delta] \leq \frac{\mathbf{E}[|X|^k]}{\delta^k}&lt;br /&gt;
\end{align}&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
# Show that for each &amp;lt;math&amp;gt;\delta&amp;lt;/math&amp;gt;, there exists a choice of &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; such that the &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;th-moment bound is no weaker than the Chernoff bound. (Hint: Use the probabilistic method.)&lt;br /&gt;
# Why would we still prefer the Chernoff bound to the (seemingly) stronger &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;-th moment bound?&lt;br /&gt;
&lt;br /&gt;
* [&#039;&#039;&#039;Cut size in random graph&#039;&#039;&#039;] Show that with probability at least &amp;lt;math&amp;gt;2/3&amp;lt;/math&amp;gt;, the size of the max-cut in [[wikipedia:Erdős–Rényi_model#Definition|Erdős–Rényi random graph]] &amp;lt;math&amp;gt;G(n,1/2)&amp;lt;/math&amp;gt; is at most &amp;lt;math&amp;gt;n^2/8 + O(n^{1.5})&amp;lt;/math&amp;gt;. In the &amp;lt;math&amp;gt;G(n,1/2)&amp;lt;/math&amp;gt; model, each edge is included in the graph with probability &amp;lt;math&amp;gt;1/2&amp;lt;/math&amp;gt;, independently of every other edge.&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Modes of Convergence and Characteristic Function, 15 points) (&amp;lt;strong&amp;gt;Bonus problem&amp;lt;/strong&amp;gt;)==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Connection of convergence modes (I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;(X_n)_{n \ge 1}, (Y_n)_{n \ge 1}, X, Y&amp;lt;/math&amp;gt; be random variables and &amp;lt;math&amp;gt;c\in\mathbb{R}&amp;lt;/math&amp;gt; be a real number. &lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; Suppose &amp;lt;math&amp;gt;X_n \overset{D}{\to} X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_n \overset{D}{\to} c&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;X_nY_n \overset{D}{\to} cX&amp;lt;/math&amp;gt;. &lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Construct an example such that &amp;lt;math&amp;gt;X_n \overset{D}{\to} X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_n \overset{D}{\to} Y&amp;lt;/math&amp;gt; but &amp;lt;math&amp;gt;X_nY_n&amp;lt;/math&amp;gt; does not converge to &amp;lt;math&amp;gt;XY&amp;lt;/math&amp;gt; in distribution.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Connection of convergence modes (II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;(X_n)_{n \ge 1}, X&amp;lt;/math&amp;gt; be random variables. Prove that &amp;lt;math&amp;gt;X_n \overset{P}{\to} X&amp;lt;/math&amp;gt; if and only if for every subsequence &amp;lt;math&amp;gt;X_{n(m)}&amp;lt;/math&amp;gt;, there exists a further subsequence &amp;lt;math&amp;gt;Y_k = X_{n(m_k)}&amp;lt;/math&amp;gt; that converges almost surely to &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;. (Hint: you may use the first Borel-Cantelli lemma.)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Characteristic Function&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; is a discrete random variable only taking integer values and &amp;lt;math&amp;gt;\mathbf{E}[X]=0&amp;lt;/math&amp;gt;. Suppose furthermore that there is no infinite subprogression &amp;lt;math&amp;gt;a+q\mathbb{Z}&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;\mathbb{Z}&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;q&amp;gt;1&amp;lt;/math&amp;gt; for which &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; takes values almost surely in &amp;lt;math&amp;gt;a+q \mathbb{Z}&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;|\phi_X(t)|&amp;lt;1&amp;lt;/math&amp;gt; for &amp;lt;math&amp;gt;0 &amp;lt; t \le \pi&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\phi_X(t)=\mathbf{E}[\mathrm{e}^{itX}]&amp;lt;/math&amp;gt; is the characteristic function of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 4 (LLN and CLT, 15 points + 10 points) ==&lt;br /&gt;
&amp;lt;strong&amp;gt;In this problem, you may apply the results of Laws of Large Numbers (LLN) and the Central Limit Theorem (CLT) to solve the problems.&amp;lt;/strong&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;St. Petersburg paradox&amp;lt;/strong&amp;gt;] Consider the well-known game involving a fair coin. In this game, if it takes &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; tosses to obtain a head, you will win &amp;lt;math&amp;gt;2^k&amp;lt;/math&amp;gt; dollars as the reward. Despite the game&#039;s expected reward being infinite, people tend to offer relatively modest amounts to participate. The following provides a mathematical explanation for this phenomenon.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
For each &amp;lt;math&amp;gt;n \ge 1&amp;lt;/math&amp;gt;, let &amp;lt;math&amp;gt;X_{n,1}, X_{n,2},\ldots, X_{n,k}&amp;lt;/math&amp;gt; be independent random variables. Furthermore, let &amp;lt;math&amp;gt;b_n &amp;gt; 0&amp;lt;/math&amp;gt; be real numbers with &amp;lt;math&amp;gt;b_n \to \infty&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widetilde{X}_{n,k} = X_{n,k} \mathbf{1}_{|X_{n,k}| \le b_n}&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;1 \le k \le n&amp;lt;/math&amp;gt;. If &amp;lt;math&amp;gt;\sum_{k=1}^n \mathbf{Pr}(|X_{n,k}| &amp;gt; b_n) \to 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;b_n^{-2} \sum_{k=1}^n \mathbf{E}[\widetilde{X}_{n,k}^2] \to 0&amp;lt;/math&amp;gt; when &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, then &amp;lt;math&amp;gt;(S_n-a_n)/b_n \overset{P}{\to} 0 &amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;S_n = \sum_{k=1}^n X_{n,k}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;a_n = \sum_{k=1}^n \mathbf{E}[\widetilde{X}_{n,k}]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;S_n&amp;lt;/math&amp;gt; be the total winnings after playing &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds of the game. Prove that &amp;lt;math&amp;gt;\frac{S_n}{n \log_2 n} \overset{P}{\to} 1&amp;lt;/math&amp;gt;. (Therefore, a fair price to play this game &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; times is roughly &amp;lt;math&amp;gt;n \log_2 n&amp;lt;/math&amp;gt; dollars)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; (&amp;lt;strong&amp;gt;Bonus problem, 5 points&amp;lt;/strong&amp;gt;)&lt;br /&gt;
Let &amp;lt;math&amp;gt;S_n&amp;lt;/math&amp;gt; be the total winnings after playing &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds of the game. Prove that &amp;lt;math&amp;gt; \limsup_{n \to \infty} \frac{S_n}{n \log_2 n} = \infty&amp;lt;/math&amp;gt; almost surely. (Hint: You may use Borel-Cantelli lemmas)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; (&amp;lt;strong&amp;gt;Bonus problem, 5 points&amp;lt;/strong&amp;gt;)&lt;br /&gt;
Let &amp;lt;math&amp;gt;X_1,X_2,\ldots,X_n&amp;lt;/math&amp;gt; are i.i.d. copies of a random variable &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; with the standard Cauchy distribution (i.e., the probability density function of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; is &amp;lt;math&amp;gt;\frac{1}{\pi} \frac{1}{1+x^2}&amp;lt;/math&amp;gt;), show that &amp;lt;math&amp;gt;\frac{S_n}{n \log n}&amp;lt;/math&amp;gt; converges in probability to a constant &amp;lt;math&amp;gt;\frac{2}{\pi}&amp;lt;/math&amp;gt; but is almost surely unbounded, where &amp;lt;math&amp;gt;S_n = \sum_{i=1}^n |X_i|&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Monte Carlo Integration&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; be a continuous function on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;U_1,U_2,\ldots,U_n \sim U(0,1)&amp;lt;/math&amp;gt; be independent random variables. Show that &amp;lt;math&amp;gt;I = \frac{1}{n}\sum_{i=1}^n f(U_i) \to \int_0^1 f(x) \mathrm{d}x&amp;lt;/math&amp;gt; in probability. (Remark: This holds so long as &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; is a measurable function on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;\int_0^1 |f(x)| \mathrm{d}x &amp;lt; \infty&amp;lt;/math&amp;gt;)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Square root&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X_1,X_2,\ldots&amp;lt;/math&amp;gt; be i.i.d. random variables with &amp;lt;math&amp;gt;X_i \ge 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{E}[X_1] = 1&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{Var}[X_1] = \sigma^2 \in (0,+\infty)&amp;lt;/math&amp;gt;. Show &amp;lt;math&amp;gt;2(\sqrt{S_n} - \sqrt{n}) \overset{D}{\to} \sigma N(0,1)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;S_n = \sum_{i=1}^n X_i&amp;lt;/math&amp;gt;. (Hint: You may use the statement in Problem 3.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2024)/Problem_Set_4&amp;diff=12468</id>
		<title>概率论与数理统计 (Spring 2024)/Problem Set 4</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2024)/Problem_Set_4&amp;diff=12468"/>
		<updated>2024-05-29T04:02:01Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: /* Problem 4 (LLN and CLT, 15 points + 10 points) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
*Bonus problem为附加题（选做）。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 4==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Continuous Random Variables, 30 points)==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Density function&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Determine the value of &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;f(x) = C\exp(-x-e^{-x}), x\in \mathbb{R}&amp;lt;/math&amp;gt; is a probability density function (PDF) for a continuous random variable.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Independence&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be independent and identically distributed continuous random variables with cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; and probability density function (PDF) &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt;. Find out the density functions of &amp;lt;math&amp;gt;V = \max\{X,Y\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;U = \min\{X,Y\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Correlation&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be uniformly distributed on &amp;lt;math&amp;gt;(-1,1)&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_k = \cos(k \pi X)&amp;lt;/math&amp;gt; for &amp;lt;math&amp;gt;k=1,2,\ldots,n&amp;lt;/math&amp;gt;. Are the random variables &amp;lt;math&amp;gt;Y_1, Y_2, \ldots, Y_n&amp;lt;/math&amp;gt; correlated? independent? You should prove your claim rigorously.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Expectation of random variables (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a continuous random variable with mean &amp;lt;math&amp;gt;\mu&amp;lt;/math&amp;gt; and cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X \ge 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\int_{0}^a F(x) dx = \int_{a}^{\infty} [1-F(x)] dx&amp;lt;/math&amp;gt; if and only if &amp;lt;math&amp;gt;a = \mu&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; has finite variance. Show that &amp;lt;math&amp;gt;g(a) = \mathbb{E}((X-a)^2)&amp;lt;/math&amp;gt; achieves the minimum when &amp;lt;math&amp;gt;a = \mu&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Expectation of random variables (II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X, Y&amp;lt;/math&amp;gt; be two independent and identically distributed continuous random variables with cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;. Furthermore, &amp;lt;math&amp;gt;X,Y \ge 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbb{E}[|X-Y|] = 2 \left(\mathbb{E}[X] - \int_{0}^{\infty} (1-F(x))^2 dx\right)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Conditional distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be two random variables. The joint density of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; is given by &amp;lt;math&amp;gt;f(x,y) = c(x^2 - y^2)e^{-x}&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;0\leq x &amp;lt;\infty&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-x\leq y \leq x&amp;lt;/math&amp;gt;. Here, &amp;lt;math&amp;gt;c\in \mathbb{R}_+&amp;lt;/math&amp;gt; is a constant. Find out the conditional distribution of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;, given &amp;lt;math&amp;gt;X = x&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Uniform Distribution (I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;P_i = (X_i,Y_i), 1\leq i\leq n&amp;lt;/math&amp;gt;, be independent, uniformly distributed points in the unit square &amp;lt;math&amp;gt;[0,1]^2&amp;lt;/math&amp;gt;. A point &amp;lt;math&amp;gt;P_i&amp;lt;/math&amp;gt; is called &amp;quot;peripheral&amp;quot; if, for all &amp;lt;math&amp;gt;r = 1,2,\cdots,n&amp;lt;/math&amp;gt;, either &amp;lt;math&amp;gt;X_r \leq X_i&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;Y_r \leq Y_i&amp;lt;/math&amp;gt;, or both. Find out the expected number of peripheral points.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Uniform Distribution (II)&amp;lt;/strong&amp;gt;] Derive the moment generating function of the standard uniform distribution, i.e., uniform distribution on &amp;lt;math&amp;gt;(0,1)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Exponential distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; have an exponential distribution. Show that &amp;lt;math&amp;gt;\textbf{Pr}[X&amp;gt;s+x|X&amp;gt;s] = \textbf{Pr}[X&amp;gt;x]&amp;lt;/math&amp;gt;, for &amp;lt;math&amp;gt;x,s\geq 0&amp;lt;/math&amp;gt;. This is the memoryless property. Show that the exponential distribution is the only continuous distribution with this property.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Normal distribution(I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X,Y\sim N(0,1)&amp;lt;/math&amp;gt; be two independent and identically distributed normal random variables. Let &amp;lt;math&amp;gt;Z = X-Y&amp;lt;/math&amp;gt;. Find the density function of &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;|Z|&amp;lt;/math&amp;gt; respectively.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Normal distribution(II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; have the &amp;lt;math&amp;gt;N(0,1)&amp;lt;/math&amp;gt; distribution and let &amp;lt;math&amp;gt;a&amp;gt;0&amp;lt;/math&amp;gt;. Show that the random variable &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; given by&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{equation*}&lt;br /&gt;
Y = \begin{cases}&lt;br /&gt;
X, &amp;amp; |X|&amp;lt; a \\&lt;br /&gt;
-X, &amp;amp; |X|\geq a&lt;br /&gt;
\end{cases}&lt;br /&gt;
\end{equation*}&amp;lt;/math&amp;gt;&lt;br /&gt;
has the &amp;lt;math&amp;gt;N(0,1)&amp;lt;/math&amp;gt; distribution, and find an expression for &amp;lt;math&amp;gt;\rho(a) = \textbf{Cov}(X,Y)&amp;lt;/math&amp;gt; in terms of the density function &amp;lt;math&amp;gt;\phi&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random process&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Given a real number &amp;lt;math&amp;gt;U&amp;lt;1&amp;lt;/math&amp;gt; as input of the following process, find out the expected returning value.&lt;br /&gt;
{{Theorem|&#039;&#039;Process&#039;&#039;|&lt;br /&gt;
:&#039;&#039;&#039;Input:&#039;&#039;&#039;  real numbers &amp;lt;math&amp;gt;U &amp;lt; 1&amp;lt;/math&amp;gt;;&lt;br /&gt;
----&lt;br /&gt;
:initialize &amp;lt;math&amp;gt;x = 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;count = 0&amp;lt;/math&amp;gt;;&lt;br /&gt;
:while &amp;lt;math&amp;gt; x &amp;lt; U &amp;lt;/math&amp;gt; do&lt;br /&gt;
:* choose &amp;lt;math&amp;gt;y \in (0,1)&amp;lt;/math&amp;gt; uniformly at random;&lt;br /&gt;
:* update &amp;lt;math&amp;gt;x = x + y&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;count = count + 1&amp;lt;/math&amp;gt;;&lt;br /&gt;
:return &amp;lt;math&amp;gt;count&amp;lt;/math&amp;gt;;&lt;br /&gt;
}}&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Random semicircle&amp;lt;/strong&amp;gt;] We sample &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; points within a circle &amp;lt;math&amp;gt;C=\{(x,y) \in \mathbb{R}^2 \mid x^2+y^2 \le 1\}&amp;lt;/math&amp;gt; independently and uniformly at random (i.e., the density function &amp;lt;math&amp;gt;f(x,y) \propto 1_{(x,y) \in C}&amp;lt;/math&amp;gt;). Find out the probability that they all lie within some semicircle of the original circle &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;. (Hint: you may apply the technique of change of variables, see [https://en.wikipedia.org/wiki/Random_variable#Functions_of_random_variables function of random variables] or Chapter 4.7 in [GS])&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Stochastic domination&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X, Y&amp;lt;/math&amp;gt; be continuous random variables. Show that &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; dominates &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; stochastically if and only if &amp;lt;math&amp;gt;\mathbb{E}[f(X)]\geq \mathbb{E}[f(Y)]&amp;lt;/math&amp;gt; for any non-decreasing function &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; for which the expectations exist.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Concentration of measure, 10 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Tossing coins&amp;lt;/strong&amp;gt;] We repeatedly toss a fair coin (with an equal probability of heads and tails). Let the random variable &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of throws required to obtain a total of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; heads. Show that &amp;lt;math&amp;gt;\textbf{Pr}[X &amp;gt; 2n + 2\sqrt{n\log n}]\leq O(1/n)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Chernoff vs Chebyshev&amp;lt;/strong&amp;gt;] We have a standard six-sided die. Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of times a 6 occurs in &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; throws off the die. Compare the best upper bounds on &amp;lt;math&amp;gt;\textbf{Pr}[X\geq n/4]&amp;lt;/math&amp;gt; that you can obtain using Chebyshev&#039;s inequality and Chernoff bounds.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;&amp;lt;strong&amp;gt;-th moment bound]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a random variable with expectation &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; such that moment generating function &amp;lt;math&amp;gt;\mathbf{E}[\exp(t|X|)]&amp;lt;/math&amp;gt; is finite for some &amp;lt;math&amp;gt; t &amp;gt; 0 &amp;lt;/math&amp;gt;. We can use the following two kinds of tail inequalities for &amp;lt;math&amp;gt; X &amp;lt;/math&amp;gt;: &lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&amp;lt;nowiki&amp;gt;     &amp;lt;/nowiki&amp;gt;&#039;&#039;&#039;&#039;&#039;Chernoff Bound&#039;&#039;&#039;&#039;&#039;&lt;br /&gt;
:&amp;lt;math&amp;gt;&lt;br /&gt;
\begin{align}&lt;br /&gt;
\mathbf{Pr}[|X| \geq \delta] \leq \min_{t \geq 0} \frac{\mathbf{E}[e^{t|X|}]}{e^{t\delta}}&lt;br /&gt;
\end{align}&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&#039;&#039;&#039;&#039;&#039;&amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;th-Moment Bound&#039;&#039;&#039;&#039;&#039;&lt;br /&gt;
:&amp;lt;math&amp;gt;&lt;br /&gt;
\begin{align}&lt;br /&gt;
\mathbf{Pr}[|X| \geq \delta] \leq \frac{\mathbf{E}[|X|^k]}{\delta^k}&lt;br /&gt;
\end{align}&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
# Show that for each &amp;lt;math&amp;gt;\delta&amp;lt;/math&amp;gt;, there exists a choice of &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; such that the &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;th-moment bound is no weaker than the Chernoff bound. (Hint: Use the probabilistic method.)&lt;br /&gt;
# Why would we still prefer the Chernoff bound to the (seemingly) stronger &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;-th moment bound?&lt;br /&gt;
&lt;br /&gt;
* [&#039;&#039;&#039;Cut size in random graph&#039;&#039;&#039;] Show that with probability at least &amp;lt;math&amp;gt;2/3&amp;lt;/math&amp;gt;, the size of the max-cut in [[wikipedia:Erdős–Rényi_model#Definition|Erdős–Rényi random graph]] &amp;lt;math&amp;gt;G(n,1/2)&amp;lt;/math&amp;gt; is at most &amp;lt;math&amp;gt;n^2/8 + O(n^{1.5})&amp;lt;/math&amp;gt;. In the &amp;lt;math&amp;gt;G(n,1/2)&amp;lt;/math&amp;gt; model, each edge is included in the graph with probability &amp;lt;math&amp;gt;1/2&amp;lt;/math&amp;gt;, independently of every other edge.&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Modes of Convergence and Characteristic Function, 15 points) (&amp;lt;strong&amp;gt;Bonus problem&amp;lt;/strong&amp;gt;)==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Connection of convergence modes (I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;(X_n)_{n \ge 1}, (Y_n)_{n \ge 1}, X, Y&amp;lt;/math&amp;gt; be random variables and &amp;lt;math&amp;gt;c\in\mathbb{R}&amp;lt;/math&amp;gt; be a real number. &lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; Suppose &amp;lt;math&amp;gt;X_n \overset{D}{\to} X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_n \overset{D}{\to} c&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;X_nY_n \overset{D}{\to} cX&amp;lt;/math&amp;gt;. &lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Construct an example such that &amp;lt;math&amp;gt;X_n \overset{D}{\to} X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_n \overset{D}{\to} Y&amp;lt;/math&amp;gt; but &amp;lt;math&amp;gt;X_nY_n&amp;lt;/math&amp;gt; does not converge to &amp;lt;math&amp;gt;XY&amp;lt;/math&amp;gt; in distribution.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Connection of convergence modes (II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;(X_n)_{n \ge 1}, X&amp;lt;/math&amp;gt; be random variables. Prove that &amp;lt;math&amp;gt;X_n \overset{P}{\to} X&amp;lt;/math&amp;gt; if and only if for every subsequence &amp;lt;math&amp;gt;X_{n(m)}&amp;lt;/math&amp;gt;, there exists a further subsequence &amp;lt;math&amp;gt;Y_k = X_{n(m_k)}&amp;lt;/math&amp;gt; that converges almost surely to &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;. (Hint: you may use the first Borel-Cantelli lemma.)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Characteristic Function&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; is a discrete random variable only taking integer values and &amp;lt;math&amp;gt;\mathbf{E}[X]=0&amp;lt;/math&amp;gt;. Suppose furthermore that there is no infinite subprogression &amp;lt;math&amp;gt;a+q\mathbb{Z}&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;\mathbb{Z}&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;q&amp;gt;1&amp;lt;/math&amp;gt; for which &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; takes values almost surely in &amp;lt;math&amp;gt;a+q \mathbb{Z}&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;|\phi_X(t)|&amp;lt;1&amp;lt;/math&amp;gt; for &amp;lt;math&amp;gt;0 &amp;lt; t \le \pi&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\phi_X(t)=\mathbf{E}[\mathrm{e}^{itX}]&amp;lt;/math&amp;gt; is the characteristic function of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 4 (LLN and CLT, 15 points + 10 points) ==&lt;br /&gt;
&amp;lt;strong&amp;gt;In this problem, you may apply the results of Laws of Large Numbers (LLN) and the Central Limit Theorem (CLT) to solve the problems.&amp;lt;/strong&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;St. Petersburg paradox&amp;lt;/strong&amp;gt;] Consider the well-known game involving a fair coin. In this game, if it takes &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; tosses to obtain a head, you will win &amp;lt;math&amp;gt;2^k&amp;lt;/math&amp;gt; dollars as the reward. Despite the game&#039;s expected reward being infinite, people tend to offer relatively modest amounts to participate. The following provides a mathematical explanation for this phenomenon.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
For each &amp;lt;math&amp;gt;n \ge 1&amp;lt;/math&amp;gt;, let &amp;lt;math&amp;gt;X_{n,1}, X_{n,2},\ldots, X_{n,k}&amp;lt;/math&amp;gt; be independent random variables. Furthermore, let &amp;lt;math&amp;gt;b_n &amp;gt; 0&amp;lt;/math&amp;gt; be real numbers with &amp;lt;math&amp;gt;b_n \to \infty&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widetilde{X}_{n,k} = X_{n,k} \mathbf{1}_{|X_{n,k}| \le b_n}&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;1 \le k \le n&amp;lt;/math&amp;gt;. If &amp;lt;math&amp;gt;\sum_{k=1}^n \mathbf{Pr}(|X_{n,k}| &amp;gt; b_n) \to 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;b_n^{-2} \sum_{k=1}^n \mathbf{E}[\widetilde{X}_{n,k}^2] \to 0&amp;lt;/math&amp;gt; when &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, then &amp;lt;math&amp;gt;(S_n-a_n)/b_n \overset{P}{\to} 0 &amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;S_n = \sum_{k=1}^n X_{n,k}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;a_n = \sum_{k=1}^n \mathbf{E}[\widetilde{X}_{n,k}]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;S_n&amp;lt;/math&amp;gt; be the total winnings after playing &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds of the game. Prove that &amp;lt;math&amp;gt;\frac{S_n}{n \log_2 n} \overset{P}{\to} 1&amp;lt;/math&amp;gt;. (Therefore, a fair price to play this game &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; times is roughly &amp;lt;math&amp;gt;n \log_2 n&amp;lt;/math&amp;gt; dollars)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; (&amp;lt;strong&amp;gt;Bonus problem, 5 points&amp;lt;/strong&amp;gt;)&lt;br /&gt;
Let &amp;lt;math&amp;gt;S_n&amp;lt;/math&amp;gt; be the total winnings after playing &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds of the game. Prove that &amp;lt;math&amp;gt; \limsup_{n \to \infty} \frac{S_n}{n \log_2 n} = \infty&amp;lt;/math&amp;gt; almost surely. (Hint: You may use Borel-Cantelli lemmas)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; (&amp;lt;strong&amp;gt;Bonus problem, 5 points&amp;lt;/strong&amp;gt;)&lt;br /&gt;
Let &amp;lt;math&amp;gt;X_1,X_2,\ldots,X_n&amp;lt;/math&amp;gt; are i.i.d. copies of a random variable &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; with the standard Cauchy distribution (i.e., the probability density function of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; is &amp;lt;math&amp;gt;\frac{1}{\pi} \frac{1}{1+x^2}&amp;lt;/math&amp;gt;), show that &amp;lt;math&amp;gt;\frac{S_n}{n \log n}&amp;lt;/math&amp;gt; converges in probability to a constant &amp;lt;math&amp;gt;\frac{2}{\pi}&amp;lt;/math&amp;gt; but is almost surely unbounded.  &lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Monte Carlo Integration&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; be a continuous function on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;U_1,U_2,\ldots,U_n \sim U(0,1)&amp;lt;/math&amp;gt; be independent random variables. Show that &amp;lt;math&amp;gt;I = \frac{1}{n}\sum_{i=1}^n f(U_i) \to \int_0^1 f(x) \mathrm{d}x&amp;lt;/math&amp;gt; in probability. (Remark: This holds so long as &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; is a measurable function on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;\int_0^1 |f(x)| \mathrm{d}x &amp;lt; \infty&amp;lt;/math&amp;gt;)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Square root&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X_1,X_2,\ldots&amp;lt;/math&amp;gt; be i.i.d. random variables with &amp;lt;math&amp;gt;X_i \ge 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{E}[X_1] = 1&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{Var}[X_1] = \sigma^2 \in (0,+\infty)&amp;lt;/math&amp;gt;. Show &amp;lt;math&amp;gt;2(\sqrt{S_n} - \sqrt{n}) \overset{D}{\to} \sigma N(0,1)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;S_n = \sum_{i=1}^n X_i&amp;lt;/math&amp;gt;. (Hint: You may use the statement in Problem 3.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2024)&amp;diff=12450</id>
		<title>概率论与数理统计 (Spring 2024)</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2024)&amp;diff=12450"/>
		<updated>2024-05-25T05:48:55Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: /* Assignments */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Infobox&lt;br /&gt;
|name         = Infobox&lt;br /&gt;
|bodystyle    = &lt;br /&gt;
|title        = &amp;lt;font size=3&amp;gt;&#039;&#039;&#039;概率论与数理统计&#039;&#039;&#039;&amp;lt;br&amp;gt;&lt;br /&gt;
&#039;&#039;&#039;Probability Theory&#039;&#039;&#039; &amp;lt;br&amp;gt; &amp;amp; &#039;&#039;&#039;Mathematical Statistics&#039;&#039;&#039;&amp;lt;/font&amp;gt;&lt;br /&gt;
|titlestyle   = &lt;br /&gt;
&lt;br /&gt;
|image        = &lt;br /&gt;
|imagestyle   = &lt;br /&gt;
|caption      = &lt;br /&gt;
|captionstyle = &lt;br /&gt;
|headerstyle  = background:#ccf;&lt;br /&gt;
|labelstyle   = background:#ddf;&lt;br /&gt;
|datastyle    = &lt;br /&gt;
&lt;br /&gt;
|header1 =Instructor&lt;br /&gt;
|label1  = &lt;br /&gt;
|data1   = &lt;br /&gt;
|header2 = &lt;br /&gt;
|label2  = &lt;br /&gt;
|data2   = &#039;&#039;&#039;尹一通&#039;&#039;&#039;&lt;br /&gt;
|header3 = &lt;br /&gt;
|label3  = Email&lt;br /&gt;
|data3   = yinyt@nju.edu.cn  &lt;br /&gt;
|header4 =&lt;br /&gt;
|label4  = office&lt;br /&gt;
|data4   = 计算机系 804&lt;br /&gt;
|header5 = &lt;br /&gt;
|label5  = &lt;br /&gt;
|data5   = &#039;&#039;&#039;刘景铖&#039;&#039;&#039;&lt;br /&gt;
|header6 = &lt;br /&gt;
|label6  = Email&lt;br /&gt;
|data6   = liu@nju.edu.cn  &lt;br /&gt;
|header7 =&lt;br /&gt;
|label7  = office&lt;br /&gt;
|data7   = 计算机系 516&lt;br /&gt;
|header8 = Class&lt;br /&gt;
|label8  = &lt;br /&gt;
|data8   = &lt;br /&gt;
|header9 =&lt;br /&gt;
|label9  = Class meeting&lt;br /&gt;
|data9   = Monday, 2pm-4pm&amp;lt;br&amp;gt;&lt;br /&gt;
Wednesday (双), 10am-12pm&amp;lt;br&amp;gt;&lt;br /&gt;
仙Ⅰ-204&lt;br /&gt;
|header10=&lt;br /&gt;
|label10 = Office hour&lt;br /&gt;
|data10  = Monday, 4pm-5pm &amp;lt;br&amp;gt;计算机系 804（尹一通）&amp;lt;br&amp;gt;计算机系 516（刘景铖）&lt;br /&gt;
|header11= Textbook&lt;br /&gt;
|label11 = &lt;br /&gt;
|data11  = &lt;br /&gt;
|header12=&lt;br /&gt;
|label12 = &lt;br /&gt;
|data12  = [[File:概率导论.jpeg|border|100px]]&lt;br /&gt;
|header13=&lt;br /&gt;
|label13 = &lt;br /&gt;
|data13  = &#039;&#039;&#039;概率导论&#039;&#039;&#039;（第2版·修订版）&amp;lt;br&amp;gt; Dimitri P. Bertsekas and John N. Tsitsiklis&amp;lt;br&amp;gt; 郑忠国 童行伟 译；人民邮电出版社 (2022)&lt;br /&gt;
|header14=&lt;br /&gt;
|label14 = &lt;br /&gt;
|data14  = [[File:Grimmett_probability.jpg|border|100px]]&lt;br /&gt;
|header15=&lt;br /&gt;
|label15 = &lt;br /&gt;
|data15  = &#039;&#039;&#039;Probability and Random Processes&#039;&#039;&#039; (4E) &amp;lt;br&amp;gt; Geoffrey Grimmett and David Stirzaker &amp;lt;br&amp;gt;  Oxford University Press (2020)&lt;br /&gt;
|header16=&lt;br /&gt;
|label16 = &lt;br /&gt;
|data16  = [[File:Probability_and_Computing_2ed.jpg|border|100px]]&lt;br /&gt;
|header17=&lt;br /&gt;
|label17 = &lt;br /&gt;
|data17  = &#039;&#039;&#039;Probability and Computing&#039;&#039;&#039; (2E) &amp;lt;br&amp;gt; Michael Mitzenmacher and Eli Upfal &amp;lt;br&amp;gt;   Cambridge University Press (2017)&lt;br /&gt;
|belowstyle = background:#ddf;&lt;br /&gt;
|below = &lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
This is the webpage for the &#039;&#039;Probability Theory and Mathematical Statistics&#039;&#039; (概率论与数理统计) class of Spring 2024. Students who take this class should check this page periodically for content updates and new announcements. &lt;br /&gt;
&lt;br /&gt;
= Announcement =&lt;br /&gt;
* TBA&lt;br /&gt;
&lt;br /&gt;
= Course info =&lt;br /&gt;
* &#039;&#039;&#039;Instructor &#039;&#039;&#039;: &lt;br /&gt;
:* [http://tcs.nju.edu.cn/yinyt/ 尹一通]：[mailto:yinyt@nju.edu.cn &amp;lt;yinyt@nju.edu.cn&amp;gt;]，计算机系 804 &lt;br /&gt;
:* [https://liuexp.github.io 刘景铖]：[mailto:liu@nju.edu.cn &amp;lt;liu@nju.edu.cn&amp;gt;]，计算机系 516 &lt;br /&gt;
* &#039;&#039;&#039;Teaching assistant&#039;&#039;&#039;:&lt;br /&gt;
** [https://sites.google.com/view/xinyuanzhang 张昕渊]：[mailto:zhangxy@smail.nju.edu.cn &amp;lt;zhangxy@smail.nju.edu.cn&amp;gt;]，计算机系 410&lt;br /&gt;
** 邹宗瑞：[mailto:zou.zongrui@smail.nju.edu.cn &amp;lt;zou.zongrui@smail.nju.edu.cn&amp;gt;]，计算机系 410&lt;br /&gt;
* &#039;&#039;&#039;Class meeting&#039;&#039;&#039;:&lt;br /&gt;
** 周一：2pm-4pm，仙Ⅰ-204&lt;br /&gt;
** 周三（双）：10am-12pm，仙Ⅰ-204&lt;br /&gt;
* &#039;&#039;&#039;Office hour&#039;&#039;&#039;: &lt;br /&gt;
:* Monday, 4pm-5pm, 计算机系 804（尹一通）&lt;br /&gt;
:* Tuesday, 3pm-4pm, 计算机系 516（刘景铖）&lt;br /&gt;
:* &#039;&#039;&#039;QQ群&#039;&#039;&#039;: 629856946（申请加入需提供姓名、院系、学号）&lt;br /&gt;
&lt;br /&gt;
= Syllabus =&lt;br /&gt;
课程内容分为三大部分：&lt;br /&gt;
* &#039;&#039;&#039;经典概率论&#039;&#039;&#039;：概率空间、随机变量及其数字特征、多维与连续随机变量、极限定理等内容&lt;br /&gt;
* &#039;&#039;&#039;概率与计算&#039;&#039;&#039;：测度集中现象 (concentration of measure)、概率法 (the probabilistic method)、离散随机过程的相关专题&lt;br /&gt;
* &#039;&#039;&#039;数理统计&#039;&#039;&#039;：参数估计、假设检验、贝叶斯估计、线性回归等统计推断等概念&lt;br /&gt;
&lt;br /&gt;
对于第一和第二部分，要求清楚掌握基本概念，深刻理解关键的现象与规律以及背后的原理，并可以灵活运用所学方法求解相关问题。对于第三部分，要求熟悉数理统计的若干基本概念，以及典型的统计模型和统计推断问题。&lt;br /&gt;
&lt;br /&gt;
经过本课程的训练，力求使学生能够熟悉掌握概率的语言，并会利用概率思维来理解客观世界并对其建模，以及驾驭概率的数学工具来分析和求解专业问题。&lt;br /&gt;
&lt;br /&gt;
=== 教材与参考书 Course Materials ===&lt;br /&gt;
* &#039;&#039;&#039;[BT]&#039;&#039;&#039; 概率导论（第2版·修订版），[美]伯特瑟卡斯（Dimitri P.Bertsekas）[美]齐齐克利斯（John N.Tsitsiklis）著，郑忠国 童行伟 译，人民邮电出版社（2022）。&lt;br /&gt;
* &#039;&#039;&#039;[GS]&#039;&#039;&#039; &#039;&#039;Probability and Random Processes&#039;&#039;, by Geoffrey Grimmett and David Stirzaker; Oxford University Press; 4th edition (2020).&lt;br /&gt;
* &#039;&#039;&#039;[MU]&#039;&#039;&#039; &#039;&#039;Probability and Computing: Randomization and Probabilistic Techniques in Algorithms and Data Analysis&#039;&#039;, by Michael Mitzenmacher, Eli Upfal; Cambridge University Press; 2nd edition (2017).&lt;br /&gt;
&lt;br /&gt;
=== 成绩 Grading Policy ===&lt;br /&gt;
* 课程成绩：本课程将会有若干次作业和一次期末考试。最终成绩将由平时作业成绩和期末考试成绩综合得出。&lt;br /&gt;
* 迟交：如果有特殊的理由，无法按时完成作业，请提前联系授课老师，给出正当理由。否则迟交的作业将不被接受。&lt;br /&gt;
&lt;br /&gt;
=== &amp;lt;font color=red&amp;gt; 学术诚信 Academic Integrity &amp;lt;/font&amp;gt;===&lt;br /&gt;
学术诚信是所有从事学术活动的学生和学者最基本的职业道德底线，本课程将不遗余力的维护学术诚信规范，违反这一底线的行为将不会被容忍。&lt;br /&gt;
&lt;br /&gt;
作业完成的原则：署你名字的工作必须是你个人的贡献。在完成作业的过程中，允许讨论，前提是讨论的所有参与者均处于同等完成度。但关键想法的执行、以及作业文本的写作必须独立完成，并在作业中致谢（acknowledge）所有参与讨论的人。符合规则的讨论与致谢将不会影响得分。不允许其他任何形式的合作——尤其是与已经完成作业的同学“讨论”。&lt;br /&gt;
&lt;br /&gt;
本课程将对剽窃行为采取零容忍的态度。在完成作业过程中，对他人工作（出版物、互联网资料、其他人的作业等）直接的文本抄袭和对关键思想、关键元素的抄袭，按照 [http://www.acm.org/publications/policies/plagiarism_policy ACM Policy on Plagiarism]的解释，都将视为剽窃。剽窃者成绩将被取消。如果发现互相抄袭行为，&amp;lt;font color=red&amp;gt; 抄袭和被抄袭双方的成绩都将被取消&amp;lt;/font&amp;gt;。因此请主动防止自己的作业被他人抄袭。&lt;br /&gt;
&lt;br /&gt;
学术诚信影响学生个人的品行，也关乎整个教育系统的正常运转。为了一点分数而做出学术不端的行为，不仅使自己沦为一个欺骗者，也使他人的诚实努力失去意义。让我们一起努力维护一个诚信的环境。&lt;br /&gt;
&lt;br /&gt;
= Assignments =&lt;br /&gt;
*[[概率论与数理统计 (Spring 2024)/Problem Set 1|Problem Set 1]]  请在 &amp;lt;strike&amp;gt;2024/4/3&amp;lt;/strike&amp;gt;&amp;lt;font color=red&amp;gt;2024/4/8&amp;lt;/font&amp;gt; 上课之前(2pm UTC+8)提交到 [mailto:pr2024_nju@163.com pr2024_nju@163.com] (文件名为&#039;&amp;lt;font color=red &amp;gt;学号_姓名_A1.pdf&amp;lt;/font&amp;gt;&#039;). &lt;br /&gt;
** [[概率论与数理统计 (Spring 2024)/第一次作业提交名单|第一次作业提交名单]]&lt;br /&gt;
*[[概率论与数理统计 (Spring 2024)/Problem Set 2|Problem Set 2]]  请在 &amp;lt;strike&amp;gt;2024/4/29&amp;lt;/strike&amp;gt;&amp;lt;font color=red&amp;gt;2024/5/6&amp;lt;/font&amp;gt; 上课之前(2pm UTC+8)提交到 [mailto:pr2024_nju@163.com pr2024_nju@163.com] (文件名为&#039;&amp;lt;font color=red &amp;gt;学号_姓名_A2.pdf&amp;lt;/font&amp;gt;&#039;).&lt;br /&gt;
** [[概率论与数理统计 (Spring 2024)/第二次作业提交名单|第二次作业提交名单]]&lt;br /&gt;
*[[概率论与数理统计 (Spring 2024)/Problem Set 3|Problem Set 3]]  请在 2024/5/20 上课之前(2pm UTC+8)提交到 [mailto:pr2024_nju@163.com pr2024_nju@163.com] (文件名为&#039;&amp;lt;font color=red &amp;gt;学号_姓名_A3.pdf&amp;lt;/font&amp;gt;&#039;).&lt;br /&gt;
** [[概率论与数理统计 (Spring 2024)/第三次作业提交名单|第三次作业提交名单]]&lt;br /&gt;
*[[概率论与数理统计 (Spring 2024)/Problem Set 4|Problem Set 4]]  请在 2024/6/12 上课之前(10am UTC+8)提交到 [mailto:pr2024_nju@163.com pr2024_nju@163.com] (文件名为&#039;&amp;lt;font color=red &amp;gt;学号_姓名_A4.pdf&amp;lt;/font&amp;gt;&#039;).&lt;br /&gt;
&lt;br /&gt;
= Lectures =&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2024/Intro.pdf 课程简介]&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2024/ProbSpace.pdf 概率空间]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 第1章&#039;&#039;&#039; 或 &#039;&#039;&#039;[GS] Chapter 1&#039;&#039;&#039;&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2024)/Karger&#039;s min-cut algorithm| Karger&#039;s min-cut algorithm]]&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2024/RandVar.pdf 随机变量]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 第2章&#039;&#039;&#039; 或 &#039;&#039;&#039;[GS] Chapter 2, Sections 3.1~3.5, 3.7&#039;&#039;&#039;&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[MU] Chapter 2&#039;&#039;&#039;&lt;br /&gt;
#* the [http://tcs.nju.edu.cn/slides/prob2024/discrete-pmf.nb &#039;&#039;&#039;Mathematica notebook file&#039;&#039;&#039;] for the PMFs of basic discrete distributions&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2024)/Average-case analysis of QuickSort|Average-case analysis of &#039;&#039;&#039;&#039;&#039;QuickSort&#039;&#039;&#039;&#039;&#039;]]&lt;br /&gt;
#* [https://www.bilibili.com/video/BV1ta411A7fp/ 高尔顿板（Galton board）视频] 和 [https://en.wikipedia.org/wiki/Galton_board 维基百科页面]&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2024/Deviation.pdf 矩与偏差]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[MU] Chapter 3&#039;&#039;&#039;&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 章节 2.4, 4.2, 4.3, 5.1&#039;&#039;&#039; 或 &#039;&#039;&#039;[GS] Sections 3.3, 3.6, 7.3&#039;&#039;&#039;&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2024)/Two-point sampling|Two-point sampling]]&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2024)/Threshold of k-clique in random graph|Threshold of &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;-clique in random graph]]&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2024)/Weierstrass Approximation Theorem|Weierstrass approximation]]&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2024/Continuous.pdf 连续分布]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 第3章, 和4.1节&#039;&#039;&#039; 或 &#039;&#039;&#039;[GS] Chapter 4&#039;&#039;&#039;&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[MU] Chapters 8, 9&#039;&#039;&#039;&lt;br /&gt;
#* [https://measure.axler.net/MIRA.pdf Measure, Integration &amp;amp; Real Analysis] by Sheldon Axler&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2024/Convergence.pdf 极限定理]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 第5章&#039;&#039;&#039; &lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[GS] Sections 5.7~5.10, 7.1~7.5&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= Concepts =&lt;br /&gt;
* [https://plato.stanford.edu/entries/probability-interpret/ Interpretations of probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/History_of_probability History of probability]&lt;br /&gt;
* Example problems:&lt;br /&gt;
** [https://dornsifecms.usc.edu/assets/sites/520/docs/VonNeumann-ams12p36-38.pdf von Neumann&#039;s Bernoulli factory] and other [https://peteroupc.github.io/bernoulli.html Bernoulli factory algorithms]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Boy_or_Girl_paradox Boy or Girl paradox]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Monty_Hall_problem Monty Hall problem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bertrand_paradox_(probability) Bertrand paradox]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hard_spheres Hard spheres model] and [https://en.wikipedia.org/wiki/Ising_model Ising model]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/PageRank &#039;&#039;PageRank&#039;&#039;] and stationary [https://en.wikipedia.org/wiki/Random_walk random walk]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Diffusion_process Diffusion process] and [https://en.wikipedia.org/wiki/Diffusion_model diffusion model]&lt;br /&gt;
*[https://en.wikipedia.org/wiki/Probability_space Probability space]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Sample_space Sample space]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Event_(probability_theory) Event] and [https://en.wikipedia.org/wiki/Σ-algebra &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-algebra]&lt;br /&gt;
** Kolmogorov&#039;s [https://en.wikipedia.org/wiki/Probability_axioms axioms of probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Discrete_uniform_distribution Classical] and [https://en.wikipedia.org/wiki/Geometric_probability goemetric probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Boole%27s_inequality Union bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Inclusion%E2%80%93exclusion_principle Inclusion-Exclusion principle]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Boole%27s_inequality#Bonferroni_inequalities Bonferroni inequalities]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Conditional_probability Conditional probability]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chain_rule_(probability) Chain rule]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_total_probability Law of total probability]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bayes%27_theorem Bayes&#039; law]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Independence_(probability_theory) Independence] &lt;br /&gt;
** [https://en.wikipedia.org/wiki/Pairwise_independence Pairwise independence]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Random_variable Random variable]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cumulative_distribution_function Cumulative distribution function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_mass_function Probability mass function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_density_function Probability density function]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Multivariate_random_variable Random vector]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Joint_probability_distribution Joint probability distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Conditional_probability_distribution Conditional probability distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Marginal_distribution Marginal distribution]&lt;br /&gt;
* Some &#039;&#039;&#039;discrete&#039;&#039;&#039; probability distributions&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bernoulli_trial Bernoulli trial] and [https://en.wikipedia.org/wiki/Bernoulli_distribution Bernoulli distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Discrete_uniform_distribution Discrete uniform distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Binomial_distribution Binomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Geometric_distribution Geometric distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Negative_binomial_distribution Negative binomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hypergeometric_distribution Hypergeometric distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Poisson_distribution Poisson distribution]&lt;br /&gt;
** and [https://en.wikipedia.org/wiki/List_of_probability_distributions#Discrete_distributions others]&lt;br /&gt;
* Balls into bins model&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Multinomial_distribution Multinomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Birthday_problem Birthday problem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Coupon_collector%27s_problem Coupon collector]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Balls_into_bins_problem Occupancy problem]&lt;br /&gt;
* Random graphs&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Erd%C5%91s%E2%80%93R%C3%A9nyi_model Erdős–Rényi random graph model]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Galton%E2%80%93Watson_process Galton–Watson branching process]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Expected_value Expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_the_unconscious_statistician Law of the unconscious statistician, &#039;&#039;LOTUS&#039;&#039;]&lt;br /&gt;
** [https://dlsun.github.io/probability/linearity.html Linearity of expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Conditional_expectation Conditional expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_total_expectation Law of total expectation]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Markov%27s_inequality Markov&#039;s inequality]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Chebyshev%27s_inequality Chebyshev&#039;s inequality]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Moment_(mathematics) Moment] related&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Central_moment Central moment]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Variance Variance]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Standard_deviation Standard deviation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Covariance Covariance]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Correlation Correlation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Correlation_does_not_imply_causation Correlation does not imply causation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Skewness Skewness]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Kurtosis Kurtosis]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Moment_problem Moment problem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Measure_(mathematics) Measure]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Borel_set Borel set]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Lebesgue_integration Lebesgue integration]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Measurable_function Measurable function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cantor_set Cantor set]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Non-measurable_set Non-measurable set]&lt;br /&gt;
*[https://en.wikipedia.org/wiki/Probability_density_function Probability density function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_distribution#Absolutely_continuous_probability_distribution Continuous probability distribution]&lt;br /&gt;
**[https://en.wikipedia.org/wiki/Conditional_probability_distribution#Conditional_continuous_distributions Conditional continuous distributions]&lt;br /&gt;
**[https://en.wikipedia.org/wiki/Convolution_of_probability_distributions Convolution of probability distributions] and [https://en.wikipedia.org/wiki/List_of_convolutions_of_probability_distributions List of convolutions of probability distributions]&lt;br /&gt;
* Some &#039;&#039;&#039;continuous&#039;&#039;&#039; probability distributions&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Continuous_uniform_distribution Continuous uniform distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Exponential_distribution Exponential distribution] and [https://en.wikipedia.org/wiki/Poisson_point_process Poisson point process]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Normal_distribution Normal (Gaussion) distribution]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Gaussian_function Gaussian function] and [https://en.wikipedia.org/wiki/Gaussian_integral Gaussian integral]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Standard_normal_table Standard normal table]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/68%E2%80%9395%E2%80%9399.7_rule 68–95–99.7 rule]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Sum_of_normally_distributed_random_variables Sum of normally distributed random variables]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Multivariate_normal_distribution Multivariate normal distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chi-squared_distribution Chi-squared distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Gamma_distribution Gamma distribution] and [https://en.wikipedia.org/wiki/Gamma_function Gamma function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Beta_distribution Beta distribution] and [https://en.wikipedia.org/wiki/Beta_function Beta function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cauchy_distribution Cauchy distribution]&lt;br /&gt;
** and [https://en.wikipedia.org/wiki/List_of_probability_distributions#Absolutely_continuous_distributions others]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Inverse_transform_sampling Inverse transform sampling]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stochastic_dominance Stochastic dominance] and [https://en.wikipedia.org/wiki/Coupling_(probability) Coupling] &lt;br /&gt;
* [https://en.wikipedia.org/wiki/Moment-generating_function Moment-generating function]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Convergence_of_random_variables Convergence of random variables]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Pointwise_convergence Pointwise convergence]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Convergence_of_measures#Weak_convergence_of_measures Weak convergence of measures]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Convergence_in_measure Convergence in measure]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Skorokhod%27s_representation_theorem Skorokhod&#039;s representation theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Continuous_mapping_theorem Continuous mapping theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Borel%E2%80%93Cantelli_lemma Borel–Cantelli lemma] and [https://en.wikipedia.org/wiki/Zero%E2%80%93one_law zero-one laws]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Law_of_large_numbers Law of large numbers]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Central_limit_theorem Central limit theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/De_Moivre%E2%80%93Laplace_theorem De Moivre–Laplace theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Berry%E2%80%93Esseen_theorem Berry–Esseen theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Characteristic_function_(probability_theory) Characteristic function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Lévy%27s_continuity_theorem Lévy&#039;s continuity theorem]&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2024)&amp;diff=12449</id>
		<title>概率论与数理统计 (Spring 2024)</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2024)&amp;diff=12449"/>
		<updated>2024-05-24T15:34:55Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: /* Assignments */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Infobox&lt;br /&gt;
|name         = Infobox&lt;br /&gt;
|bodystyle    = &lt;br /&gt;
|title        = &amp;lt;font size=3&amp;gt;&#039;&#039;&#039;概率论与数理统计&#039;&#039;&#039;&amp;lt;br&amp;gt;&lt;br /&gt;
&#039;&#039;&#039;Probability Theory&#039;&#039;&#039; &amp;lt;br&amp;gt; &amp;amp; &#039;&#039;&#039;Mathematical Statistics&#039;&#039;&#039;&amp;lt;/font&amp;gt;&lt;br /&gt;
|titlestyle   = &lt;br /&gt;
&lt;br /&gt;
|image        = &lt;br /&gt;
|imagestyle   = &lt;br /&gt;
|caption      = &lt;br /&gt;
|captionstyle = &lt;br /&gt;
|headerstyle  = background:#ccf;&lt;br /&gt;
|labelstyle   = background:#ddf;&lt;br /&gt;
|datastyle    = &lt;br /&gt;
&lt;br /&gt;
|header1 =Instructor&lt;br /&gt;
|label1  = &lt;br /&gt;
|data1   = &lt;br /&gt;
|header2 = &lt;br /&gt;
|label2  = &lt;br /&gt;
|data2   = &#039;&#039;&#039;尹一通&#039;&#039;&#039;&lt;br /&gt;
|header3 = &lt;br /&gt;
|label3  = Email&lt;br /&gt;
|data3   = yinyt@nju.edu.cn  &lt;br /&gt;
|header4 =&lt;br /&gt;
|label4  = office&lt;br /&gt;
|data4   = 计算机系 804&lt;br /&gt;
|header5 = &lt;br /&gt;
|label5  = &lt;br /&gt;
|data5   = &#039;&#039;&#039;刘景铖&#039;&#039;&#039;&lt;br /&gt;
|header6 = &lt;br /&gt;
|label6  = Email&lt;br /&gt;
|data6   = liu@nju.edu.cn  &lt;br /&gt;
|header7 =&lt;br /&gt;
|label7  = office&lt;br /&gt;
|data7   = 计算机系 516&lt;br /&gt;
|header8 = Class&lt;br /&gt;
|label8  = &lt;br /&gt;
|data8   = &lt;br /&gt;
|header9 =&lt;br /&gt;
|label9  = Class meeting&lt;br /&gt;
|data9   = Monday, 2pm-4pm&amp;lt;br&amp;gt;&lt;br /&gt;
Wednesday (双), 10am-12pm&amp;lt;br&amp;gt;&lt;br /&gt;
仙Ⅰ-204&lt;br /&gt;
|header10=&lt;br /&gt;
|label10 = Office hour&lt;br /&gt;
|data10  = Monday, 4pm-5pm &amp;lt;br&amp;gt;计算机系 804（尹一通）&amp;lt;br&amp;gt;计算机系 516（刘景铖）&lt;br /&gt;
|header11= Textbook&lt;br /&gt;
|label11 = &lt;br /&gt;
|data11  = &lt;br /&gt;
|header12=&lt;br /&gt;
|label12 = &lt;br /&gt;
|data12  = [[File:概率导论.jpeg|border|100px]]&lt;br /&gt;
|header13=&lt;br /&gt;
|label13 = &lt;br /&gt;
|data13  = &#039;&#039;&#039;概率导论&#039;&#039;&#039;（第2版·修订版）&amp;lt;br&amp;gt; Dimitri P. Bertsekas and John N. Tsitsiklis&amp;lt;br&amp;gt; 郑忠国 童行伟 译；人民邮电出版社 (2022)&lt;br /&gt;
|header14=&lt;br /&gt;
|label14 = &lt;br /&gt;
|data14  = [[File:Grimmett_probability.jpg|border|100px]]&lt;br /&gt;
|header15=&lt;br /&gt;
|label15 = &lt;br /&gt;
|data15  = &#039;&#039;&#039;Probability and Random Processes&#039;&#039;&#039; (4E) &amp;lt;br&amp;gt; Geoffrey Grimmett and David Stirzaker &amp;lt;br&amp;gt;  Oxford University Press (2020)&lt;br /&gt;
|header16=&lt;br /&gt;
|label16 = &lt;br /&gt;
|data16  = [[File:Probability_and_Computing_2ed.jpg|border|100px]]&lt;br /&gt;
|header17=&lt;br /&gt;
|label17 = &lt;br /&gt;
|data17  = &#039;&#039;&#039;Probability and Computing&#039;&#039;&#039; (2E) &amp;lt;br&amp;gt; Michael Mitzenmacher and Eli Upfal &amp;lt;br&amp;gt;   Cambridge University Press (2017)&lt;br /&gt;
|belowstyle = background:#ddf;&lt;br /&gt;
|below = &lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
This is the webpage for the &#039;&#039;Probability Theory and Mathematical Statistics&#039;&#039; (概率论与数理统计) class of Spring 2024. Students who take this class should check this page periodically for content updates and new announcements. &lt;br /&gt;
&lt;br /&gt;
= Announcement =&lt;br /&gt;
* TBA&lt;br /&gt;
&lt;br /&gt;
= Course info =&lt;br /&gt;
* &#039;&#039;&#039;Instructor &#039;&#039;&#039;: &lt;br /&gt;
:* [http://tcs.nju.edu.cn/yinyt/ 尹一通]：[mailto:yinyt@nju.edu.cn &amp;lt;yinyt@nju.edu.cn&amp;gt;]，计算机系 804 &lt;br /&gt;
:* [https://liuexp.github.io 刘景铖]：[mailto:liu@nju.edu.cn &amp;lt;liu@nju.edu.cn&amp;gt;]，计算机系 516 &lt;br /&gt;
* &#039;&#039;&#039;Teaching assistant&#039;&#039;&#039;:&lt;br /&gt;
** [https://sites.google.com/view/xinyuanzhang 张昕渊]：[mailto:zhangxy@smail.nju.edu.cn &amp;lt;zhangxy@smail.nju.edu.cn&amp;gt;]，计算机系 410&lt;br /&gt;
** 邹宗瑞：[mailto:zou.zongrui@smail.nju.edu.cn &amp;lt;zou.zongrui@smail.nju.edu.cn&amp;gt;]，计算机系 410&lt;br /&gt;
* &#039;&#039;&#039;Class meeting&#039;&#039;&#039;:&lt;br /&gt;
** 周一：2pm-4pm，仙Ⅰ-204&lt;br /&gt;
** 周三（双）：10am-12pm，仙Ⅰ-204&lt;br /&gt;
* &#039;&#039;&#039;Office hour&#039;&#039;&#039;: &lt;br /&gt;
:* Monday, 4pm-5pm, 计算机系 804（尹一通）&lt;br /&gt;
:* Tuesday, 3pm-4pm, 计算机系 516（刘景铖）&lt;br /&gt;
:* &#039;&#039;&#039;QQ群&#039;&#039;&#039;: 629856946（申请加入需提供姓名、院系、学号）&lt;br /&gt;
&lt;br /&gt;
= Syllabus =&lt;br /&gt;
课程内容分为三大部分：&lt;br /&gt;
* &#039;&#039;&#039;经典概率论&#039;&#039;&#039;：概率空间、随机变量及其数字特征、多维与连续随机变量、极限定理等内容&lt;br /&gt;
* &#039;&#039;&#039;概率与计算&#039;&#039;&#039;：测度集中现象 (concentration of measure)、概率法 (the probabilistic method)、离散随机过程的相关专题&lt;br /&gt;
* &#039;&#039;&#039;数理统计&#039;&#039;&#039;：参数估计、假设检验、贝叶斯估计、线性回归等统计推断等概念&lt;br /&gt;
&lt;br /&gt;
对于第一和第二部分，要求清楚掌握基本概念，深刻理解关键的现象与规律以及背后的原理，并可以灵活运用所学方法求解相关问题。对于第三部分，要求熟悉数理统计的若干基本概念，以及典型的统计模型和统计推断问题。&lt;br /&gt;
&lt;br /&gt;
经过本课程的训练，力求使学生能够熟悉掌握概率的语言，并会利用概率思维来理解客观世界并对其建模，以及驾驭概率的数学工具来分析和求解专业问题。&lt;br /&gt;
&lt;br /&gt;
=== 教材与参考书 Course Materials ===&lt;br /&gt;
* &#039;&#039;&#039;[BT]&#039;&#039;&#039; 概率导论（第2版·修订版），[美]伯特瑟卡斯（Dimitri P.Bertsekas）[美]齐齐克利斯（John N.Tsitsiklis）著，郑忠国 童行伟 译，人民邮电出版社（2022）。&lt;br /&gt;
* &#039;&#039;&#039;[GS]&#039;&#039;&#039; &#039;&#039;Probability and Random Processes&#039;&#039;, by Geoffrey Grimmett and David Stirzaker; Oxford University Press; 4th edition (2020).&lt;br /&gt;
* &#039;&#039;&#039;[MU]&#039;&#039;&#039; &#039;&#039;Probability and Computing: Randomization and Probabilistic Techniques in Algorithms and Data Analysis&#039;&#039;, by Michael Mitzenmacher, Eli Upfal; Cambridge University Press; 2nd edition (2017).&lt;br /&gt;
&lt;br /&gt;
=== 成绩 Grading Policy ===&lt;br /&gt;
* 课程成绩：本课程将会有若干次作业和一次期末考试。最终成绩将由平时作业成绩和期末考试成绩综合得出。&lt;br /&gt;
* 迟交：如果有特殊的理由，无法按时完成作业，请提前联系授课老师，给出正当理由。否则迟交的作业将不被接受。&lt;br /&gt;
&lt;br /&gt;
=== &amp;lt;font color=red&amp;gt; 学术诚信 Academic Integrity &amp;lt;/font&amp;gt;===&lt;br /&gt;
学术诚信是所有从事学术活动的学生和学者最基本的职业道德底线，本课程将不遗余力的维护学术诚信规范，违反这一底线的行为将不会被容忍。&lt;br /&gt;
&lt;br /&gt;
作业完成的原则：署你名字的工作必须是你个人的贡献。在完成作业的过程中，允许讨论，前提是讨论的所有参与者均处于同等完成度。但关键想法的执行、以及作业文本的写作必须独立完成，并在作业中致谢（acknowledge）所有参与讨论的人。符合规则的讨论与致谢将不会影响得分。不允许其他任何形式的合作——尤其是与已经完成作业的同学“讨论”。&lt;br /&gt;
&lt;br /&gt;
本课程将对剽窃行为采取零容忍的态度。在完成作业过程中，对他人工作（出版物、互联网资料、其他人的作业等）直接的文本抄袭和对关键思想、关键元素的抄袭，按照 [http://www.acm.org/publications/policies/plagiarism_policy ACM Policy on Plagiarism]的解释，都将视为剽窃。剽窃者成绩将被取消。如果发现互相抄袭行为，&amp;lt;font color=red&amp;gt; 抄袭和被抄袭双方的成绩都将被取消&amp;lt;/font&amp;gt;。因此请主动防止自己的作业被他人抄袭。&lt;br /&gt;
&lt;br /&gt;
学术诚信影响学生个人的品行，也关乎整个教育系统的正常运转。为了一点分数而做出学术不端的行为，不仅使自己沦为一个欺骗者，也使他人的诚实努力失去意义。让我们一起努力维护一个诚信的环境。&lt;br /&gt;
&lt;br /&gt;
= Assignments =&lt;br /&gt;
*[[概率论与数理统计 (Spring 2024)/Problem Set 1|Problem Set 1]]  请在 &amp;lt;strike&amp;gt;2024/4/3&amp;lt;/strike&amp;gt;&amp;lt;font color=red&amp;gt;2024/4/8&amp;lt;/font&amp;gt; 上课之前(2pm UTC+8)提交到 [mailto:pr2024_nju@163.com pr2024_nju@163.com] (文件名为&#039;&amp;lt;font color=red &amp;gt;学号_姓名_A1.pdf&amp;lt;/font&amp;gt;&#039;). &lt;br /&gt;
** [[概率论与数理统计 (Spring 2024)/第一次作业提交名单|第一次作业提交名单]]&lt;br /&gt;
*[[概率论与数理统计 (Spring 2024)/Problem Set 2|Problem Set 2]]  请在 &amp;lt;strike&amp;gt;2024/4/29&amp;lt;/strike&amp;gt;&amp;lt;font color=red&amp;gt;2024/5/6&amp;lt;/font&amp;gt; 上课之前(2pm UTC+8)提交到 [mailto:pr2024_nju@163.com pr2024_nju@163.com] (文件名为&#039;&amp;lt;font color=red &amp;gt;学号_姓名_A2.pdf&amp;lt;/font&amp;gt;&#039;).&lt;br /&gt;
** [[概率论与数理统计 (Spring 2024)/第二次作业提交名单|第二次作业提交名单]]&lt;br /&gt;
*[[概率论与数理统计 (Spring 2024)/Problem Set 3|Problem Set 3]]  请在 2024/5/20 上课之前(2pm UTC+8)提交到 [mailto:pr2024_nju@163.com pr2024_nju@163.com] (文件名为&#039;&amp;lt;font color=red &amp;gt;学号_姓名_A3.pdf&amp;lt;/font&amp;gt;&#039;).&lt;br /&gt;
** [[概率论与数理统计 (Spring 2024)/第三次作业提交名单|第三次作业提交名单]]&lt;br /&gt;
*[[概率论与数理统计 (Spring 2024)/Problem Set 4|Problem Set 4]]  请在 2024/6/10 上课之前(2pm UTC+8)提交到 [mailto:pr2024_nju@163.com pr2024_nju@163.com] (文件名为&#039;&amp;lt;font color=red &amp;gt;学号_姓名_A4.pdf&amp;lt;/font&amp;gt;&#039;).&lt;br /&gt;
&lt;br /&gt;
= Lectures =&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2024/Intro.pdf 课程简介]&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2024/ProbSpace.pdf 概率空间]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 第1章&#039;&#039;&#039; 或 &#039;&#039;&#039;[GS] Chapter 1&#039;&#039;&#039;&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2024)/Karger&#039;s min-cut algorithm| Karger&#039;s min-cut algorithm]]&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2024/RandVar.pdf 随机变量]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 第2章&#039;&#039;&#039; 或 &#039;&#039;&#039;[GS] Chapter 2, Sections 3.1~3.5, 3.7&#039;&#039;&#039;&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[MU] Chapter 2&#039;&#039;&#039;&lt;br /&gt;
#* the [http://tcs.nju.edu.cn/slides/prob2024/discrete-pmf.nb &#039;&#039;&#039;Mathematica notebook file&#039;&#039;&#039;] for the PMFs of basic discrete distributions&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2024)/Average-case analysis of QuickSort|Average-case analysis of &#039;&#039;&#039;&#039;&#039;QuickSort&#039;&#039;&#039;&#039;&#039;]]&lt;br /&gt;
#* [https://www.bilibili.com/video/BV1ta411A7fp/ 高尔顿板（Galton board）视频] 和 [https://en.wikipedia.org/wiki/Galton_board 维基百科页面]&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2024/Deviation.pdf 矩与偏差]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[MU] Chapter 3&#039;&#039;&#039;&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 章节 2.4, 4.2, 4.3, 5.1&#039;&#039;&#039; 或 &#039;&#039;&#039;[GS] Sections 3.3, 3.6, 7.3&#039;&#039;&#039;&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2024)/Two-point sampling|Two-point sampling]]&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2024)/Threshold of k-clique in random graph|Threshold of &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;-clique in random graph]]&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2024)/Weierstrass Approximation Theorem|Weierstrass approximation]]&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2024/Continuous.pdf 连续分布]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 第3章, 和4.1节&#039;&#039;&#039; 或 &#039;&#039;&#039;[GS] Chapter 4&#039;&#039;&#039;&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[MU] Chapters 8, 9&#039;&#039;&#039;&lt;br /&gt;
#* [https://measure.axler.net/MIRA.pdf Measure, Integration &amp;amp; Real Analysis] by Sheldon Axler&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2024/Convergence.pdf 极限定理]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 第5章&#039;&#039;&#039; &lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[GS] Sections 5.7~5.10, 7.1~7.5&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= Concepts =&lt;br /&gt;
* [https://plato.stanford.edu/entries/probability-interpret/ Interpretations of probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/History_of_probability History of probability]&lt;br /&gt;
* Example problems:&lt;br /&gt;
** [https://dornsifecms.usc.edu/assets/sites/520/docs/VonNeumann-ams12p36-38.pdf von Neumann&#039;s Bernoulli factory] and other [https://peteroupc.github.io/bernoulli.html Bernoulli factory algorithms]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Boy_or_Girl_paradox Boy or Girl paradox]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Monty_Hall_problem Monty Hall problem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bertrand_paradox_(probability) Bertrand paradox]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hard_spheres Hard spheres model] and [https://en.wikipedia.org/wiki/Ising_model Ising model]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/PageRank &#039;&#039;PageRank&#039;&#039;] and stationary [https://en.wikipedia.org/wiki/Random_walk random walk]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Diffusion_process Diffusion process] and [https://en.wikipedia.org/wiki/Diffusion_model diffusion model]&lt;br /&gt;
*[https://en.wikipedia.org/wiki/Probability_space Probability space]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Sample_space Sample space]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Event_(probability_theory) Event] and [https://en.wikipedia.org/wiki/Σ-algebra &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-algebra]&lt;br /&gt;
** Kolmogorov&#039;s [https://en.wikipedia.org/wiki/Probability_axioms axioms of probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Discrete_uniform_distribution Classical] and [https://en.wikipedia.org/wiki/Geometric_probability goemetric probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Boole%27s_inequality Union bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Inclusion%E2%80%93exclusion_principle Inclusion-Exclusion principle]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Boole%27s_inequality#Bonferroni_inequalities Bonferroni inequalities]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Conditional_probability Conditional probability]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chain_rule_(probability) Chain rule]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_total_probability Law of total probability]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bayes%27_theorem Bayes&#039; law]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Independence_(probability_theory) Independence] &lt;br /&gt;
** [https://en.wikipedia.org/wiki/Pairwise_independence Pairwise independence]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Random_variable Random variable]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cumulative_distribution_function Cumulative distribution function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_mass_function Probability mass function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_density_function Probability density function]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Multivariate_random_variable Random vector]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Joint_probability_distribution Joint probability distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Conditional_probability_distribution Conditional probability distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Marginal_distribution Marginal distribution]&lt;br /&gt;
* Some &#039;&#039;&#039;discrete&#039;&#039;&#039; probability distributions&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bernoulli_trial Bernoulli trial] and [https://en.wikipedia.org/wiki/Bernoulli_distribution Bernoulli distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Discrete_uniform_distribution Discrete uniform distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Binomial_distribution Binomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Geometric_distribution Geometric distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Negative_binomial_distribution Negative binomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hypergeometric_distribution Hypergeometric distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Poisson_distribution Poisson distribution]&lt;br /&gt;
** and [https://en.wikipedia.org/wiki/List_of_probability_distributions#Discrete_distributions others]&lt;br /&gt;
* Balls into bins model&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Multinomial_distribution Multinomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Birthday_problem Birthday problem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Coupon_collector%27s_problem Coupon collector]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Balls_into_bins_problem Occupancy problem]&lt;br /&gt;
* Random graphs&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Erd%C5%91s%E2%80%93R%C3%A9nyi_model Erdős–Rényi random graph model]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Galton%E2%80%93Watson_process Galton–Watson branching process]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Expected_value Expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_the_unconscious_statistician Law of the unconscious statistician, &#039;&#039;LOTUS&#039;&#039;]&lt;br /&gt;
** [https://dlsun.github.io/probability/linearity.html Linearity of expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Conditional_expectation Conditional expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_total_expectation Law of total expectation]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Markov%27s_inequality Markov&#039;s inequality]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Chebyshev%27s_inequality Chebyshev&#039;s inequality]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Moment_(mathematics) Moment] related&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Central_moment Central moment]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Variance Variance]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Standard_deviation Standard deviation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Covariance Covariance]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Correlation Correlation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Correlation_does_not_imply_causation Correlation does not imply causation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Skewness Skewness]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Kurtosis Kurtosis]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Moment_problem Moment problem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Measure_(mathematics) Measure]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Borel_set Borel set]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Lebesgue_integration Lebesgue integration]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Measurable_function Measurable function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cantor_set Cantor set]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Non-measurable_set Non-measurable set]&lt;br /&gt;
*[https://en.wikipedia.org/wiki/Probability_density_function Probability density function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_distribution#Absolutely_continuous_probability_distribution Continuous probability distribution]&lt;br /&gt;
**[https://en.wikipedia.org/wiki/Conditional_probability_distribution#Conditional_continuous_distributions Conditional continuous distributions]&lt;br /&gt;
**[https://en.wikipedia.org/wiki/Convolution_of_probability_distributions Convolution of probability distributions] and [https://en.wikipedia.org/wiki/List_of_convolutions_of_probability_distributions List of convolutions of probability distributions]&lt;br /&gt;
* Some &#039;&#039;&#039;continuous&#039;&#039;&#039; probability distributions&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Continuous_uniform_distribution Continuous uniform distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Exponential_distribution Exponential distribution] and [https://en.wikipedia.org/wiki/Poisson_point_process Poisson point process]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Normal_distribution Normal (Gaussion) distribution]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Gaussian_function Gaussian function] and [https://en.wikipedia.org/wiki/Gaussian_integral Gaussian integral]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Standard_normal_table Standard normal table]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/68%E2%80%9395%E2%80%9399.7_rule 68–95–99.7 rule]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Sum_of_normally_distributed_random_variables Sum of normally distributed random variables]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Multivariate_normal_distribution Multivariate normal distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chi-squared_distribution Chi-squared distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Gamma_distribution Gamma distribution] and [https://en.wikipedia.org/wiki/Gamma_function Gamma function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Beta_distribution Beta distribution] and [https://en.wikipedia.org/wiki/Beta_function Beta function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cauchy_distribution Cauchy distribution]&lt;br /&gt;
** and [https://en.wikipedia.org/wiki/List_of_probability_distributions#Absolutely_continuous_distributions others]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Inverse_transform_sampling Inverse transform sampling]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stochastic_dominance Stochastic dominance] and [https://en.wikipedia.org/wiki/Coupling_(probability) Coupling] &lt;br /&gt;
* [https://en.wikipedia.org/wiki/Moment-generating_function Moment-generating function]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Convergence_of_random_variables Convergence of random variables]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Pointwise_convergence Pointwise convergence]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Convergence_of_measures#Weak_convergence_of_measures Weak convergence of measures]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Convergence_in_measure Convergence in measure]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Skorokhod%27s_representation_theorem Skorokhod&#039;s representation theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Continuous_mapping_theorem Continuous mapping theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Borel%E2%80%93Cantelli_lemma Borel–Cantelli lemma] and [https://en.wikipedia.org/wiki/Zero%E2%80%93one_law zero-one laws]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Law_of_large_numbers Law of large numbers]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Central_limit_theorem Central limit theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/De_Moivre%E2%80%93Laplace_theorem De Moivre–Laplace theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Berry%E2%80%93Esseen_theorem Berry–Esseen theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Characteristic_function_(probability_theory) Characteristic function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Lévy%27s_continuity_theorem Lévy&#039;s continuity theorem]&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2024)/Problem_Set_4&amp;diff=12448</id>
		<title>概率论与数理统计 (Spring 2024)/Problem Set 4</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2024)/Problem_Set_4&amp;diff=12448"/>
		<updated>2024-05-24T15:33:12Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: /* Problem 4 (LLN and CLT, 15 points + 5 points) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
*Bonus problem为附加题（选做）。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 4==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Continuous Random Variables, 30 points)==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Density function&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Determine the value of &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;f(x) = C\exp(-x-e^{-x}), x\in \mathbb{R}&amp;lt;/math&amp;gt; is a probability density function (PDF) for a continuous random variable.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Independence&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be independent and identically distributed continuous random variables with cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; and probability density function (PDF) &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt;. Find out the density functions of &amp;lt;math&amp;gt;V = \max\{X,Y\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;U = \min\{X,Y\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Correlation&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be uniformly distributed on &amp;lt;math&amp;gt;(-1,1)&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_k = \cos(k \pi X)&amp;lt;/math&amp;gt; for &amp;lt;math&amp;gt;k=1,2,\ldots,n&amp;lt;/math&amp;gt;. Are the random variables &amp;lt;math&amp;gt;Y_1, Y_2, \ldots, Y_n&amp;lt;/math&amp;gt; correlated? independent? You should prove your claim rigorously.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Expectation of random variables (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a continuous random variable with mean &amp;lt;math&amp;gt;\mu&amp;lt;/math&amp;gt; and cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X \ge 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\int_{0}^a F(x) dx = \int_{a}^{\infty} [1-F(x)] dx&amp;lt;/math&amp;gt; if and only if &amp;lt;math&amp;gt;a = \mu&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; has finite variance. Show that &amp;lt;math&amp;gt;g(a) = \mathbb{E}((X-a)^2)&amp;lt;/math&amp;gt; achieves the minimum when &amp;lt;math&amp;gt;a = \mu&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Expectation of random variables (II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X, Y&amp;lt;/math&amp;gt; be two independent and identically distributed continuous random variables with cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;. Furthermore, &amp;lt;math&amp;gt;X,Y \ge 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbb{E}[|X-Y|] = 2 \left(\mathbb{E}[X] - \int_{0}^{\infty} (1-F(x))^2 dx\right)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Conditional distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be two random variables. The joint density of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; is given by &amp;lt;math&amp;gt;f(x,y) = c(x^2 - y^2)e^{-x}&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;0\leq x &amp;lt;\infty&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-x\leq y \leq x&amp;lt;/math&amp;gt;. Here, &amp;lt;math&amp;gt;c\in \mathbb{R}_+&amp;lt;/math&amp;gt; is a constant. Find out the conditional distribution of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;, given &amp;lt;math&amp;gt;X = x&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Uniform Distribution (I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;P_i = (X_i,Y_i), 1\leq i\leq n&amp;lt;/math&amp;gt;, be independent, uniformly distributed points in the unit square &amp;lt;math&amp;gt;[0,1]^2&amp;lt;/math&amp;gt;. A point &amp;lt;math&amp;gt;P_i&amp;lt;/math&amp;gt; is called &amp;quot;peripheral&amp;quot; if, for all &amp;lt;math&amp;gt;r = 1,2,\cdots,n&amp;lt;/math&amp;gt;, either &amp;lt;math&amp;gt;X_r \leq X_i&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;Y_r \leq Y_i&amp;lt;/math&amp;gt;, or both. Find out the expected number of peripheral points.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Uniform Distribution (II)&amp;lt;/strong&amp;gt;] Derive the moment generating function of the standard uniform distribution, i.e., uniform distribution on &amp;lt;math&amp;gt;(0,1)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Exponential distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; have an exponential distribution. Show that &amp;lt;math&amp;gt;\textbf{Pr}[X&amp;gt;s+x|X&amp;gt;s] = \textbf{Pr}[X&amp;gt;x]&amp;lt;/math&amp;gt;, for &amp;lt;math&amp;gt;x,s\geq 0&amp;lt;/math&amp;gt;. This is the memoryless property. Show that the exponential distribution is the only continuous distribution with this property.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Normal distribution(I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X,Y\sim N(0,1)&amp;lt;/math&amp;gt; be two independent and identically distributed normal random variables. Let &amp;lt;math&amp;gt;Z = X-Y&amp;lt;/math&amp;gt;. Find the density function of &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;|Z|&amp;lt;/math&amp;gt; respectively.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Normal distribution(II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; have the &amp;lt;math&amp;gt;N(0,1)&amp;lt;/math&amp;gt; distribution and let &amp;lt;math&amp;gt;a&amp;gt;0&amp;lt;/math&amp;gt;. Show that the random variable &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; given by&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{equation*}&lt;br /&gt;
Y = \begin{cases}&lt;br /&gt;
X, &amp;amp; |X|&amp;lt; a \\&lt;br /&gt;
-X, &amp;amp; |X|\geq a&lt;br /&gt;
\end{cases}&lt;br /&gt;
\end{equation*}&amp;lt;/math&amp;gt;&lt;br /&gt;
has the &amp;lt;math&amp;gt;N(0,1)&amp;lt;/math&amp;gt; distribution, and find an expression for &amp;lt;math&amp;gt;\rho(a) = \textbf{Cov}(X,Y)&amp;lt;/math&amp;gt; in terms of the density function &amp;lt;math&amp;gt;\phi&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random process&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Given a real number &amp;lt;math&amp;gt;U&amp;lt;1&amp;lt;/math&amp;gt; as input of the following process, find out the expected returning value.&lt;br /&gt;
{{Theorem|&#039;&#039;Process&#039;&#039;|&lt;br /&gt;
:&#039;&#039;&#039;Input:&#039;&#039;&#039;  real numbers &amp;lt;math&amp;gt;U &amp;lt; 1&amp;lt;/math&amp;gt;;&lt;br /&gt;
----&lt;br /&gt;
:initialize &amp;lt;math&amp;gt;x = 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;count = 0&amp;lt;/math&amp;gt;;&lt;br /&gt;
:while &amp;lt;math&amp;gt; x &amp;lt; U &amp;lt;/math&amp;gt; do&lt;br /&gt;
:* choose &amp;lt;math&amp;gt;y \in (0,1)&amp;lt;/math&amp;gt; uniformly at random;&lt;br /&gt;
:* update &amp;lt;math&amp;gt;x = x + y&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;count = count + 1&amp;lt;/math&amp;gt;;&lt;br /&gt;
:return &amp;lt;math&amp;gt;count&amp;lt;/math&amp;gt;;&lt;br /&gt;
}}&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Random semicircle&amp;lt;/strong&amp;gt;] We sample &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; points within a circle &amp;lt;math&amp;gt;C=\{(x,y) \in \mathbb{R}^2 \mid x^2+y^2 \le 1\}&amp;lt;/math&amp;gt; independently and uniformly at random (i.e., the density function &amp;lt;math&amp;gt;f(x,y) \propto 1_{(x,y) \in C}&amp;lt;/math&amp;gt;). Find out the probability that they all lie within some semicircle of the original circle &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;. (Hint: you may apply the technique of change of variables, see [https://en.wikipedia.org/wiki/Random_variable#Functions_of_random_variables function of random variables] or Chapter 4.7 in [GS])&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Stochastic domination&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X, Y&amp;lt;/math&amp;gt; be continuous random variables. Show that &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; dominates &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; stochastically if and only if &amp;lt;math&amp;gt;\mathbb{E}[f(X)]\geq \mathbb{E}[f(Y)]&amp;lt;/math&amp;gt; for any non-decreasing function &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; for which the expectations exist.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Concentration of measure, 10 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Tossing coins&amp;lt;/strong&amp;gt;] We repeatedly toss a fair coin (with an equal probability of heads and tails). Let the random variable &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of throws required to obtain a total of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; heads. Show that &amp;lt;math&amp;gt;\textbf{Pr}[X &amp;gt; 2n + 2\sqrt{n\log n}]\leq O(1/n)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Chernoff vs Chebyshev&amp;lt;/strong&amp;gt;] We have a standard six-sided die. Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of times a 6 occurs in &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; throws off the die. Compare the best upper bounds on &amp;lt;math&amp;gt;\textbf{Pr}[X\geq n/4]&amp;lt;/math&amp;gt; that you can obtain using Chebyshev&#039;s inequality and Chernoff bounds.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;&amp;lt;strong&amp;gt;-th moment bound]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a random variable with expectation &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; such that moment generating function &amp;lt;math&amp;gt;\mathbf{E}[\exp(t|X|)]&amp;lt;/math&amp;gt; is finite for some &amp;lt;math&amp;gt; t &amp;gt; 0 &amp;lt;/math&amp;gt;. We can use the following two kinds of tail inequalities for &amp;lt;math&amp;gt; X &amp;lt;/math&amp;gt;: &lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&amp;lt;nowiki&amp;gt;     &amp;lt;/nowiki&amp;gt;&#039;&#039;&#039;&#039;&#039;Chernoff Bound&#039;&#039;&#039;&#039;&#039;&lt;br /&gt;
:&amp;lt;math&amp;gt;&lt;br /&gt;
\begin{align}&lt;br /&gt;
\mathbf{Pr}[|X| \geq \delta] \leq \min_{t \geq 0} \frac{\mathbf{E}[e^{t|X|}]}{e^{t\delta}}&lt;br /&gt;
\end{align}&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&#039;&#039;&#039;&#039;&#039;&amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;th-Moment Bound&#039;&#039;&#039;&#039;&#039;&lt;br /&gt;
:&amp;lt;math&amp;gt;&lt;br /&gt;
\begin{align}&lt;br /&gt;
\mathbf{Pr}[|X| \geq \delta] \leq \frac{\mathbf{E}[|X|^k]}{\delta^k}&lt;br /&gt;
\end{align}&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
# Show that for each &amp;lt;math&amp;gt;\delta&amp;lt;/math&amp;gt;, there exists a choice of &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; such that the &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;th-moment bound is no weaker than the Chernoff bound. (Hint: Use the probabilistic method.)&lt;br /&gt;
# Why would we still prefer the Chernoff bound to the (seemingly) stronger &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;-th moment bound?&lt;br /&gt;
&lt;br /&gt;
* [&#039;&#039;&#039;Cut size in random graph&#039;&#039;&#039;] Show that with probability at least &amp;lt;math&amp;gt;2/3&amp;lt;/math&amp;gt;, the size of the max-cut in [[wikipedia:Erdős–Rényi_model#Definition|Erdős–Rényi random graph]] &amp;lt;math&amp;gt;G(n,1/2)&amp;lt;/math&amp;gt; is at most &amp;lt;math&amp;gt;n^2/8 + O(n^{1.5})&amp;lt;/math&amp;gt;. In the &amp;lt;math&amp;gt;G(n,1/2)&amp;lt;/math&amp;gt; model, each edge is included in the graph with probability &amp;lt;math&amp;gt;1/2&amp;lt;/math&amp;gt;, independently of every other edge.&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Modes of Convergence and Characteristic Function, 15 points) (&amp;lt;strong&amp;gt;Bonus problem&amp;lt;/strong&amp;gt;)==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Connection of convergence modes (I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;(X_n)_{n \ge 1}, (Y_n)_{n \ge 1}, X, Y&amp;lt;/math&amp;gt; be random variables and &amp;lt;math&amp;gt;c\in\mathbb{R}&amp;lt;/math&amp;gt; be a real number. &lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; Suppose &amp;lt;math&amp;gt;X_n \overset{D}{\to} X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_n \overset{D}{\to} c&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;X_nY_n \overset{D}{\to} cX&amp;lt;/math&amp;gt;. &lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Construct an example such that &amp;lt;math&amp;gt;X_n \overset{D}{\to} X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_n \overset{D}{\to} Y&amp;lt;/math&amp;gt; but &amp;lt;math&amp;gt;X_nY_n&amp;lt;/math&amp;gt; does not converge to &amp;lt;math&amp;gt;XY&amp;lt;/math&amp;gt; in distribution.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Connection of convergence modes (II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;(X_n)_{n \ge 1}, X&amp;lt;/math&amp;gt; be random variables. Prove that &amp;lt;math&amp;gt;X_n \overset{P}{\to} X&amp;lt;/math&amp;gt; if and only if for every subsequence &amp;lt;math&amp;gt;X_{n(m)}&amp;lt;/math&amp;gt;, there exists a further subsequence &amp;lt;math&amp;gt;Y_k = X_{n(m_k)}&amp;lt;/math&amp;gt; that converges almost surely to &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;. (Hint: you may use the first Borel-Cantelli lemma.)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Characteristic Function&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; is a discrete random variable only taking integer values and &amp;lt;math&amp;gt;\mathbf{E}[X]=0&amp;lt;/math&amp;gt;. Suppose furthermore that there is no infinite subprogression &amp;lt;math&amp;gt;a+q\mathbb{Z}&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;\mathbb{Z}&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;q&amp;gt;1&amp;lt;/math&amp;gt; for which &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; takes values almost surely in &amp;lt;math&amp;gt;a+q \mathbb{Z}&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;|\phi_X(t)|&amp;lt;1&amp;lt;/math&amp;gt; for &amp;lt;math&amp;gt;0 &amp;lt; t \le \pi&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\phi_X(t)=\mathbf{E}[\mathrm{e}^{itX}]&amp;lt;/math&amp;gt; is the characteristic function of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 4 (LLN and CLT, 15 points + 10 points) ==&lt;br /&gt;
&amp;lt;strong&amp;gt;In this problem, you may apply the results of Laws of Large Numbers (LLN) and the Central Limit Theorem (CLT) to solve the problems.&amp;lt;/strong&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;St. Petersburg paradox&amp;lt;/strong&amp;gt;] Consider the well-known game involving a fair coin. In this game, if it takes &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; tosses to obtain a head, you will win &amp;lt;math&amp;gt;2^k&amp;lt;/math&amp;gt; dollars as the reward. Despite the game&#039;s expected reward being infinite, people tend to offer relatively modest amounts to participate. The following provides a mathematical explanation for this phenomenon.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
For each &amp;lt;math&amp;gt;n \ge 1&amp;lt;/math&amp;gt;, let &amp;lt;math&amp;gt;X_{n,1}, X_{n,2},\ldots, X_{n,k}&amp;lt;/math&amp;gt; be independent random variables. Furthermore, let &amp;lt;math&amp;gt;b_n &amp;gt; 0&amp;lt;/math&amp;gt; be real numbers with &amp;lt;math&amp;gt;b_n \to \infty&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widetilde{X}_{n,k} = X_{n,k} \mathbf{1}_{|X_{n,k}| \le b_n}&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;1 \le k \le n&amp;lt;/math&amp;gt;. If &amp;lt;math&amp;gt;\sum_{k=1}^n \mathbf{Pr}(|X_{n,k}| &amp;gt; b_n) \to 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;b_n^{-2} \sum_{k=1}^n \mathbf{E}[\widetilde{X}_{n,k}^2] \to 0&amp;lt;/math&amp;gt; when &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, then &amp;lt;math&amp;gt;(S_n-a_n)/b_n \overset{P}{\to} 0 &amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;S_n = \sum_{k=1}^n X_{n,k}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;a_n = \sum_{k=1}^n \mathbf{E}[\widetilde{X}_{n,k}]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;S_n&amp;lt;/math&amp;gt; be the total winnings after playing &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds of the game. Prove that &amp;lt;math&amp;gt;\frac{S_n}{n \log_2 n} \overset{P}{\to} 1&amp;lt;/math&amp;gt;. (Therefore, a fair price to play this game &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; times is roughly &amp;lt;math&amp;gt;n \log_2 n&amp;lt;/math&amp;gt; dollars)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; (&amp;lt;strong&amp;gt;Bonus problem, 5 points&amp;lt;/strong&amp;gt;)&lt;br /&gt;
Let &amp;lt;math&amp;gt;S_n&amp;lt;/math&amp;gt; be the total winnings after playing &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds of the game. Prove that &amp;lt;math&amp;gt; \limsup_{n \to \infty} \frac{S_n}{n \log_2 n} = \infty&amp;lt;/math&amp;gt; almost surely. (Hint: You may use Borel-Cantelli lemmas)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; (&amp;lt;strong&amp;gt;Bonus problem, 5 points&amp;lt;/strong&amp;gt;)&lt;br /&gt;
Let &amp;lt;math&amp;gt;X_1,X_2,\ldots,X_n&amp;lt;/math&amp;gt; are i.i.d. copies of a random variable &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; with the standard Cauchy distribution (i.e., the probability density function of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; is &amp;lt;math&amp;gt;\frac{1}{\pi} \frac{1}{1+x^2}&amp;lt;/math&amp;gt;), show that &amp;lt;math&amp;gt;\frac{S_n}{n \log n}&amp;lt;/math&amp;gt; converges in probability to a constant &amp;lt;math&amp;gt;\frac{2}{\pi}&amp;lt;/math&amp;gt; but is almost surely unbounded.  &lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Monte Carlo Integration&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; be a continuous function on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;U_1,U_2,\ldots,U_n \sim U(0,1)&amp;lt;/math&amp;gt; be independent random variables. Show that &amp;lt;math&amp;gt;I = \frac{1}{n}\sum_{i=1}^n f(U_i) \to \int_0^1 f(x) \mathrm{d}x&amp;lt;/math&amp;gt; in probability. (Remark: This holds so long as &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; is a measurable function on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;\int_0^1 |f(x)| \mathrm{d}x &amp;lt; \infty&amp;lt;/math&amp;gt;)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Square root&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X_1,X_2,\ldots&amp;lt;/math&amp;gt; be i.i.d. random variables with &amp;lt;math&amp;gt;X_i \ge 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{E}[X_1] = 1&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{Var}[X_1] = \sigma^2 \in (0,+\infty)&amp;lt;/math&amp;gt;. Show &amp;lt;math&amp;gt;\sqrt{S_n} - \sqrt{n} \overset{D}{\to} N(0,1)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;S_n = \sum_{i=1}^n X_i&amp;lt;/math&amp;gt;. (Hint: You may use the statement in Problem 3.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2024)/Problem_Set_4&amp;diff=12447</id>
		<title>概率论与数理统计 (Spring 2024)/Problem Set 4</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2024)/Problem_Set_4&amp;diff=12447"/>
		<updated>2024-05-24T15:09:05Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: /* Problem 4 (LLN and CLT, 15 points + 5 points) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
*Bonus problem为附加题（选做）。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 4==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Continuous Random Variables, 30 points)==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Density function&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Determine the value of &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;f(x) = C\exp(-x-e^{-x}), x\in \mathbb{R}&amp;lt;/math&amp;gt; is a probability density function (PDF) for a continuous random variable.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Independence&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be independent and identically distributed continuous random variables with cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; and probability density function (PDF) &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt;. Find out the density functions of &amp;lt;math&amp;gt;V = \max\{X,Y\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;U = \min\{X,Y\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Correlation&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be uniformly distributed on &amp;lt;math&amp;gt;(-1,1)&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_k = \cos(k \pi X)&amp;lt;/math&amp;gt; for &amp;lt;math&amp;gt;k=1,2,\ldots,n&amp;lt;/math&amp;gt;. Are the random variables &amp;lt;math&amp;gt;Y_1, Y_2, \ldots, Y_n&amp;lt;/math&amp;gt; correlated? independent? You should prove your claim rigorously.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Expectation of random variables (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a continuous random variable with mean &amp;lt;math&amp;gt;\mu&amp;lt;/math&amp;gt; and cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X \ge 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\int_{0}^a F(x) dx = \int_{a}^{\infty} [1-F(x)] dx&amp;lt;/math&amp;gt; if and only if &amp;lt;math&amp;gt;a = \mu&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; has finite variance. Show that &amp;lt;math&amp;gt;g(a) = \mathbb{E}((X-a)^2)&amp;lt;/math&amp;gt; achieves the minimum when &amp;lt;math&amp;gt;a = \mu&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Expectation of random variables (II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X, Y&amp;lt;/math&amp;gt; be two independent and identically distributed continuous random variables with cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;. Furthermore, &amp;lt;math&amp;gt;X,Y \ge 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbb{E}[|X-Y|] = 2 \left(\mathbb{E}[X] - \int_{0}^{\infty} (1-F(x))^2 dx\right)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Conditional distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be two random variables. The joint density of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; is given by &amp;lt;math&amp;gt;f(x,y) = c(x^2 - y^2)e^{-x}&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;0\leq x &amp;lt;\infty&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-x\leq y \leq x&amp;lt;/math&amp;gt;. Here, &amp;lt;math&amp;gt;c\in \mathbb{R}_+&amp;lt;/math&amp;gt; is a constant. Find out the conditional distribution of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;, given &amp;lt;math&amp;gt;X = x&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Uniform Distribution (I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;P_i = (X_i,Y_i), 1\leq i\leq n&amp;lt;/math&amp;gt;, be independent, uniformly distributed points in the unit square &amp;lt;math&amp;gt;[0,1]^2&amp;lt;/math&amp;gt;. A point &amp;lt;math&amp;gt;P_i&amp;lt;/math&amp;gt; is called &amp;quot;peripheral&amp;quot; if, for all &amp;lt;math&amp;gt;r = 1,2,\cdots,n&amp;lt;/math&amp;gt;, either &amp;lt;math&amp;gt;X_r \leq X_i&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;Y_r \leq Y_i&amp;lt;/math&amp;gt;, or both. Find out the expected number of peripheral points.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Uniform Distribution (II)&amp;lt;/strong&amp;gt;] Derive the moment generating function of the standard uniform distribution, i.e., uniform distribution on &amp;lt;math&amp;gt;(0,1)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Exponential distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; have an exponential distribution. Show that &amp;lt;math&amp;gt;\textbf{Pr}[X&amp;gt;s+x|X&amp;gt;s] = \textbf{Pr}[X&amp;gt;x]&amp;lt;/math&amp;gt;, for &amp;lt;math&amp;gt;x,s\geq 0&amp;lt;/math&amp;gt;. This is the memoryless property. Show that the exponential distribution is the only continuous distribution with this property.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Normal distribution(I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X,Y\sim N(0,1)&amp;lt;/math&amp;gt; be two independent and identically distributed normal random variables. Let &amp;lt;math&amp;gt;Z = X-Y&amp;lt;/math&amp;gt;. Find the density function of &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;|Z|&amp;lt;/math&amp;gt; respectively.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Normal distribution(II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; have the &amp;lt;math&amp;gt;N(0,1)&amp;lt;/math&amp;gt; distribution and let &amp;lt;math&amp;gt;a&amp;gt;0&amp;lt;/math&amp;gt;. Show that the random variable &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; given by&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{equation*}&lt;br /&gt;
Y = \begin{cases}&lt;br /&gt;
X, &amp;amp; |X|&amp;lt; a \\&lt;br /&gt;
-X, &amp;amp; |X|\geq a&lt;br /&gt;
\end{cases}&lt;br /&gt;
\end{equation*}&amp;lt;/math&amp;gt;&lt;br /&gt;
has the &amp;lt;math&amp;gt;N(0,1)&amp;lt;/math&amp;gt; distribution, and find an expression for &amp;lt;math&amp;gt;\rho(a) = \textbf{Cov}(X,Y)&amp;lt;/math&amp;gt; in terms of the density function &amp;lt;math&amp;gt;\phi&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random process&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Given a real number &amp;lt;math&amp;gt;U&amp;lt;1&amp;lt;/math&amp;gt; as input of the following process, find out the expected returning value.&lt;br /&gt;
{{Theorem|&#039;&#039;Process&#039;&#039;|&lt;br /&gt;
:&#039;&#039;&#039;Input:&#039;&#039;&#039;  real numbers &amp;lt;math&amp;gt;U &amp;lt; 1&amp;lt;/math&amp;gt;;&lt;br /&gt;
----&lt;br /&gt;
:initialize &amp;lt;math&amp;gt;x = 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;count = 0&amp;lt;/math&amp;gt;;&lt;br /&gt;
:while &amp;lt;math&amp;gt; x &amp;lt; U &amp;lt;/math&amp;gt; do&lt;br /&gt;
:* choose &amp;lt;math&amp;gt;y \in (0,1)&amp;lt;/math&amp;gt; uniformly at random;&lt;br /&gt;
:* update &amp;lt;math&amp;gt;x = x + y&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;count = count + 1&amp;lt;/math&amp;gt;;&lt;br /&gt;
:return &amp;lt;math&amp;gt;count&amp;lt;/math&amp;gt;;&lt;br /&gt;
}}&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Random semicircle&amp;lt;/strong&amp;gt;] We sample &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; points within a circle &amp;lt;math&amp;gt;C=\{(x,y) \in \mathbb{R}^2 \mid x^2+y^2 \le 1\}&amp;lt;/math&amp;gt; independently and uniformly at random (i.e., the density function &amp;lt;math&amp;gt;f(x,y) \propto 1_{(x,y) \in C}&amp;lt;/math&amp;gt;). Find out the probability that they all lie within some semicircle of the original circle &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;. (Hint: you may apply the technique of change of variables, see [https://en.wikipedia.org/wiki/Random_variable#Functions_of_random_variables function of random variables] or Chapter 4.7 in [GS])&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Stochastic domination&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X, Y&amp;lt;/math&amp;gt; be continuous random variables. Show that &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; dominates &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; stochastically if and only if &amp;lt;math&amp;gt;\mathbb{E}[f(X)]\geq \mathbb{E}[f(Y)]&amp;lt;/math&amp;gt; for any non-decreasing function &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; for which the expectations exist.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Concentration of measure, 10 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Tossing coins&amp;lt;/strong&amp;gt;] We repeatedly toss a fair coin (with an equal probability of heads and tails). Let the random variable &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of throws required to obtain a total of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; heads. Show that &amp;lt;math&amp;gt;\textbf{Pr}[X &amp;gt; 2n + 2\sqrt{n\log n}]\leq O(1/n)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Chernoff vs Chebyshev&amp;lt;/strong&amp;gt;] We have a standard six-sided die. Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of times a 6 occurs in &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; throws off the die. Compare the best upper bounds on &amp;lt;math&amp;gt;\textbf{Pr}[X\geq n/4]&amp;lt;/math&amp;gt; that you can obtain using Chebyshev&#039;s inequality and Chernoff bounds.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;&amp;lt;strong&amp;gt;-th moment bound]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a random variable with expectation &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; such that moment generating function &amp;lt;math&amp;gt;\mathbf{E}[\exp(t|X|)]&amp;lt;/math&amp;gt; is finite for some &amp;lt;math&amp;gt; t &amp;gt; 0 &amp;lt;/math&amp;gt;. We can use the following two kinds of tail inequalities for &amp;lt;math&amp;gt; X &amp;lt;/math&amp;gt;: &lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&amp;lt;nowiki&amp;gt;     &amp;lt;/nowiki&amp;gt;&#039;&#039;&#039;&#039;&#039;Chernoff Bound&#039;&#039;&#039;&#039;&#039;&lt;br /&gt;
:&amp;lt;math&amp;gt;&lt;br /&gt;
\begin{align}&lt;br /&gt;
\mathbf{Pr}[|X| \geq \delta] \leq \min_{t \geq 0} \frac{\mathbf{E}[e^{t|X|}]}{e^{t\delta}}&lt;br /&gt;
\end{align}&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&#039;&#039;&#039;&#039;&#039;&amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;th-Moment Bound&#039;&#039;&#039;&#039;&#039;&lt;br /&gt;
:&amp;lt;math&amp;gt;&lt;br /&gt;
\begin{align}&lt;br /&gt;
\mathbf{Pr}[|X| \geq \delta] \leq \frac{\mathbf{E}[|X|^k]}{\delta^k}&lt;br /&gt;
\end{align}&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
# Show that for each &amp;lt;math&amp;gt;\delta&amp;lt;/math&amp;gt;, there exists a choice of &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; such that the &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;th-moment bound is no weaker than the Chernoff bound. (Hint: Use the probabilistic method.)&lt;br /&gt;
# Why would we still prefer the Chernoff bound to the (seemingly) stronger &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;-th moment bound?&lt;br /&gt;
&lt;br /&gt;
* [&#039;&#039;&#039;Cut size in random graph&#039;&#039;&#039;] Show that with probability at least &amp;lt;math&amp;gt;2/3&amp;lt;/math&amp;gt;, the size of the max-cut in [[wikipedia:Erdős–Rényi_model#Definition|Erdős–Rényi random graph]] &amp;lt;math&amp;gt;G(n,1/2)&amp;lt;/math&amp;gt; is at most &amp;lt;math&amp;gt;n^2/8 + O(n^{1.5})&amp;lt;/math&amp;gt;. In the &amp;lt;math&amp;gt;G(n,1/2)&amp;lt;/math&amp;gt; model, each edge is included in the graph with probability &amp;lt;math&amp;gt;1/2&amp;lt;/math&amp;gt;, independently of every other edge.&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Modes of Convergence and Characteristic Function, 15 points) (&amp;lt;strong&amp;gt;Bonus problem&amp;lt;/strong&amp;gt;)==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Connection of convergence modes (I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;(X_n)_{n \ge 1}, (Y_n)_{n \ge 1}, X, Y&amp;lt;/math&amp;gt; be random variables and &amp;lt;math&amp;gt;c\in\mathbb{R}&amp;lt;/math&amp;gt; be a real number. &lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; Suppose &amp;lt;math&amp;gt;X_n \overset{D}{\to} X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_n \overset{D}{\to} c&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;X_nY_n \overset{D}{\to} cX&amp;lt;/math&amp;gt;. &lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Construct an example such that &amp;lt;math&amp;gt;X_n \overset{D}{\to} X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_n \overset{D}{\to} Y&amp;lt;/math&amp;gt; but &amp;lt;math&amp;gt;X_nY_n&amp;lt;/math&amp;gt; does not converge to &amp;lt;math&amp;gt;XY&amp;lt;/math&amp;gt; in distribution.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Connection of convergence modes (II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;(X_n)_{n \ge 1}, X&amp;lt;/math&amp;gt; be random variables. Prove that &amp;lt;math&amp;gt;X_n \overset{P}{\to} X&amp;lt;/math&amp;gt; if and only if for every subsequence &amp;lt;math&amp;gt;X_{n(m)}&amp;lt;/math&amp;gt;, there exists a further subsequence &amp;lt;math&amp;gt;Y_k = X_{n(m_k)}&amp;lt;/math&amp;gt; that converges almost surely to &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;. (Hint: you may use the first Borel-Cantelli lemma.)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Characteristic Function&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; is a discrete random variable only taking integer values and &amp;lt;math&amp;gt;\mathbf{E}[X]=0&amp;lt;/math&amp;gt;. Suppose furthermore that there is no infinite subprogression &amp;lt;math&amp;gt;a+q\mathbb{Z}&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;\mathbb{Z}&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;q&amp;gt;1&amp;lt;/math&amp;gt; for which &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; takes values almost surely in &amp;lt;math&amp;gt;a+q \mathbb{Z}&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;|\phi_X(t)|&amp;lt;1&amp;lt;/math&amp;gt; for &amp;lt;math&amp;gt;0 &amp;lt; t \le \pi&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\phi_X(t)=\mathbf{E}[\mathrm{e}^{itX}]&amp;lt;/math&amp;gt; is the characteristic function of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 4 (LLN and CLT, 15 points + 5 points) ==&lt;br /&gt;
&amp;lt;strong&amp;gt;In this problem, you may apply the results of Laws of Large Numbers (LLN) and the Central Limit Theorem (CLT) to solve the problems.&amp;lt;/strong&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;St. Petersburg paradox&amp;lt;/strong&amp;gt;] Consider the well-known game involving a fair coin. In this game, if it takes &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; tosses to obtain a head, you will win &amp;lt;math&amp;gt;2^k&amp;lt;/math&amp;gt; dollars as the reward. Despite the game&#039;s expected reward being infinite, people tend to offer relatively modest amounts to participate. The following provides a mathematical explanation for this phenomenon.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
For each &amp;lt;math&amp;gt;n \ge 1&amp;lt;/math&amp;gt;, let &amp;lt;math&amp;gt;X_{n,1}, X_{n,2},\ldots, X_{n,k}&amp;lt;/math&amp;gt; be independent random variables. Furthermore, let &amp;lt;math&amp;gt;b_n &amp;gt; 0&amp;lt;/math&amp;gt; be real numbers with &amp;lt;math&amp;gt;b_n \to \infty&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widetilde{X}_{n,k} = X_{n,k} \mathbf{1}_{|X_{n,k}| \le b_n}&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;1 \le k \le n&amp;lt;/math&amp;gt;. If &amp;lt;math&amp;gt;\sum_{k=1}^n \mathbf{Pr}(|X_{n,k}| &amp;gt; b_n) \to 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;b_n^{-2} \sum_{k=1}^n \mathbf{E}[\widetilde{X}_{n,k}^2] \to 0&amp;lt;/math&amp;gt; when &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, then &amp;lt;math&amp;gt;(S_n-a_n)/b_n \overset{P}{\to} 0 &amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;S_n = \sum_{k=1}^n X_{n,k}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;a_n = \sum_{k=1}^n \mathbf{E}[\widetilde{X}_{n,k}]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;S_n&amp;lt;/math&amp;gt; be the total winnings after playing &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds of the game. Prove that &amp;lt;math&amp;gt;\frac{S_n}{n \log_2 n} \overset{P}{\to} 1&amp;lt;/math&amp;gt;. (Therefore, a fair price to play this game &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; times is roughly &amp;lt;math&amp;gt;n \log_2 n&amp;lt;/math&amp;gt; dollars)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; (&amp;lt;strong&amp;gt;Bonus problem, 5 points&amp;lt;/strong&amp;gt;)&lt;br /&gt;
Let &amp;lt;math&amp;gt;S_n&amp;lt;/math&amp;gt; be the total winnings after playing &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds of the game. Prove that &amp;lt;math&amp;gt; \limsup_{n \to \infty} \frac{S_n}{n \log_2 n} = \infty&amp;lt;/math&amp;gt; almost surely. (Hint: You may use Borel-Cantelli lemmas)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Monte Carlo Integration&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; be a continuous function on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;U_1,U_2,\ldots,U_n \sim U(0,1)&amp;lt;/math&amp;gt; be independent random variables. Show that &amp;lt;math&amp;gt;I = \frac{1}{n}\sum_{i=1}^n f(U_i) \to \int_0^1 f(x) \mathrm{d}x&amp;lt;/math&amp;gt; in probability. (Remark: This holds so long as &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; is a measurable function on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;\int_0^1 |f(x)| \mathrm{d}x &amp;lt; \infty&amp;lt;/math&amp;gt;)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Square root&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X_1,X_2,\ldots&amp;lt;/math&amp;gt; be i.i.d. random variables with &amp;lt;math&amp;gt;X_i \ge 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{E}[X_1] = 1&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{Var}[X_1] = \sigma^2 \in (0,+\infty)&amp;lt;/math&amp;gt;. Show &amp;lt;math&amp;gt;\sqrt{S_n} - \sqrt{n} \overset{D}{\to} N(0,1)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;S_n = \sum_{i=1}^n X_i&amp;lt;/math&amp;gt;. (Hint: You may use the statement in Problem 3.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2024)/Problem_Set_4&amp;diff=12446</id>
		<title>概率论与数理统计 (Spring 2024)/Problem Set 4</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2024)/Problem_Set_4&amp;diff=12446"/>
		<updated>2024-05-24T15:07:52Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: /* Problem 3 (Modes of Convergence and Characteristic Function, 20 points) (Bonus problem) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
*Bonus problem为附加题（选做）。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 4==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Continuous Random Variables, 30 points)==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Density function&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Determine the value of &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;f(x) = C\exp(-x-e^{-x}), x\in \mathbb{R}&amp;lt;/math&amp;gt; is a probability density function (PDF) for a continuous random variable.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Independence&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be independent and identically distributed continuous random variables with cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; and probability density function (PDF) &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt;. Find out the density functions of &amp;lt;math&amp;gt;V = \max\{X,Y\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;U = \min\{X,Y\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Correlation&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be uniformly distributed on &amp;lt;math&amp;gt;(-1,1)&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_k = \cos(k \pi X)&amp;lt;/math&amp;gt; for &amp;lt;math&amp;gt;k=1,2,\ldots,n&amp;lt;/math&amp;gt;. Are the random variables &amp;lt;math&amp;gt;Y_1, Y_2, \ldots, Y_n&amp;lt;/math&amp;gt; correlated? independent? You should prove your claim rigorously.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Expectation of random variables (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a continuous random variable with mean &amp;lt;math&amp;gt;\mu&amp;lt;/math&amp;gt; and cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X \ge 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\int_{0}^a F(x) dx = \int_{a}^{\infty} [1-F(x)] dx&amp;lt;/math&amp;gt; if and only if &amp;lt;math&amp;gt;a = \mu&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; has finite variance. Show that &amp;lt;math&amp;gt;g(a) = \mathbb{E}((X-a)^2)&amp;lt;/math&amp;gt; achieves the minimum when &amp;lt;math&amp;gt;a = \mu&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Expectation of random variables (II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X, Y&amp;lt;/math&amp;gt; be two independent and identically distributed continuous random variables with cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;. Furthermore, &amp;lt;math&amp;gt;X,Y \ge 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbb{E}[|X-Y|] = 2 \left(\mathbb{E}[X] - \int_{0}^{\infty} (1-F(x))^2 dx\right)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Conditional distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be two random variables. The joint density of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; is given by &amp;lt;math&amp;gt;f(x,y) = c(x^2 - y^2)e^{-x}&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;0\leq x &amp;lt;\infty&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-x\leq y \leq x&amp;lt;/math&amp;gt;. Here, &amp;lt;math&amp;gt;c\in \mathbb{R}_+&amp;lt;/math&amp;gt; is a constant. Find out the conditional distribution of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;, given &amp;lt;math&amp;gt;X = x&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Uniform Distribution (I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;P_i = (X_i,Y_i), 1\leq i\leq n&amp;lt;/math&amp;gt;, be independent, uniformly distributed points in the unit square &amp;lt;math&amp;gt;[0,1]^2&amp;lt;/math&amp;gt;. A point &amp;lt;math&amp;gt;P_i&amp;lt;/math&amp;gt; is called &amp;quot;peripheral&amp;quot; if, for all &amp;lt;math&amp;gt;r = 1,2,\cdots,n&amp;lt;/math&amp;gt;, either &amp;lt;math&amp;gt;X_r \leq X_i&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;Y_r \leq Y_i&amp;lt;/math&amp;gt;, or both. Find out the expected number of peripheral points.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Uniform Distribution (II)&amp;lt;/strong&amp;gt;] Derive the moment generating function of the standard uniform distribution, i.e., uniform distribution on &amp;lt;math&amp;gt;(0,1)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Exponential distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; have an exponential distribution. Show that &amp;lt;math&amp;gt;\textbf{Pr}[X&amp;gt;s+x|X&amp;gt;s] = \textbf{Pr}[X&amp;gt;x]&amp;lt;/math&amp;gt;, for &amp;lt;math&amp;gt;x,s\geq 0&amp;lt;/math&amp;gt;. This is the memoryless property. Show that the exponential distribution is the only continuous distribution with this property.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Normal distribution(I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X,Y\sim N(0,1)&amp;lt;/math&amp;gt; be two independent and identically distributed normal random variables. Let &amp;lt;math&amp;gt;Z = X-Y&amp;lt;/math&amp;gt;. Find the density function of &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;|Z|&amp;lt;/math&amp;gt; respectively.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Normal distribution(II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; have the &amp;lt;math&amp;gt;N(0,1)&amp;lt;/math&amp;gt; distribution and let &amp;lt;math&amp;gt;a&amp;gt;0&amp;lt;/math&amp;gt;. Show that the random variable &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; given by&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{equation*}&lt;br /&gt;
Y = \begin{cases}&lt;br /&gt;
X, &amp;amp; |X|&amp;lt; a \\&lt;br /&gt;
-X, &amp;amp; |X|\geq a&lt;br /&gt;
\end{cases}&lt;br /&gt;
\end{equation*}&amp;lt;/math&amp;gt;&lt;br /&gt;
has the &amp;lt;math&amp;gt;N(0,1)&amp;lt;/math&amp;gt; distribution, and find an expression for &amp;lt;math&amp;gt;\rho(a) = \textbf{Cov}(X,Y)&amp;lt;/math&amp;gt; in terms of the density function &amp;lt;math&amp;gt;\phi&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random process&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Given a real number &amp;lt;math&amp;gt;U&amp;lt;1&amp;lt;/math&amp;gt; as input of the following process, find out the expected returning value.&lt;br /&gt;
{{Theorem|&#039;&#039;Process&#039;&#039;|&lt;br /&gt;
:&#039;&#039;&#039;Input:&#039;&#039;&#039;  real numbers &amp;lt;math&amp;gt;U &amp;lt; 1&amp;lt;/math&amp;gt;;&lt;br /&gt;
----&lt;br /&gt;
:initialize &amp;lt;math&amp;gt;x = 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;count = 0&amp;lt;/math&amp;gt;;&lt;br /&gt;
:while &amp;lt;math&amp;gt; x &amp;lt; U &amp;lt;/math&amp;gt; do&lt;br /&gt;
:* choose &amp;lt;math&amp;gt;y \in (0,1)&amp;lt;/math&amp;gt; uniformly at random;&lt;br /&gt;
:* update &amp;lt;math&amp;gt;x = x + y&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;count = count + 1&amp;lt;/math&amp;gt;;&lt;br /&gt;
:return &amp;lt;math&amp;gt;count&amp;lt;/math&amp;gt;;&lt;br /&gt;
}}&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Random semicircle&amp;lt;/strong&amp;gt;] We sample &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; points within a circle &amp;lt;math&amp;gt;C=\{(x,y) \in \mathbb{R}^2 \mid x^2+y^2 \le 1\}&amp;lt;/math&amp;gt; independently and uniformly at random (i.e., the density function &amp;lt;math&amp;gt;f(x,y) \propto 1_{(x,y) \in C}&amp;lt;/math&amp;gt;). Find out the probability that they all lie within some semicircle of the original circle &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;. (Hint: you may apply the technique of change of variables, see [https://en.wikipedia.org/wiki/Random_variable#Functions_of_random_variables function of random variables] or Chapter 4.7 in [GS])&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Stochastic domination&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X, Y&amp;lt;/math&amp;gt; be continuous random variables. Show that &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; dominates &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; stochastically if and only if &amp;lt;math&amp;gt;\mathbb{E}[f(X)]\geq \mathbb{E}[f(Y)]&amp;lt;/math&amp;gt; for any non-decreasing function &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; for which the expectations exist.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Concentration of measure, 10 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Tossing coins&amp;lt;/strong&amp;gt;] We repeatedly toss a fair coin (with an equal probability of heads and tails). Let the random variable &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of throws required to obtain a total of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; heads. Show that &amp;lt;math&amp;gt;\textbf{Pr}[X &amp;gt; 2n + 2\sqrt{n\log n}]\leq O(1/n)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Chernoff vs Chebyshev&amp;lt;/strong&amp;gt;] We have a standard six-sided die. Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of times a 6 occurs in &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; throws off the die. Compare the best upper bounds on &amp;lt;math&amp;gt;\textbf{Pr}[X\geq n/4]&amp;lt;/math&amp;gt; that you can obtain using Chebyshev&#039;s inequality and Chernoff bounds.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;&amp;lt;strong&amp;gt;-th moment bound]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a random variable with expectation &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; such that moment generating function &amp;lt;math&amp;gt;\mathbf{E}[\exp(t|X|)]&amp;lt;/math&amp;gt; is finite for some &amp;lt;math&amp;gt; t &amp;gt; 0 &amp;lt;/math&amp;gt;. We can use the following two kinds of tail inequalities for &amp;lt;math&amp;gt; X &amp;lt;/math&amp;gt;: &lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&amp;lt;nowiki&amp;gt;     &amp;lt;/nowiki&amp;gt;&#039;&#039;&#039;&#039;&#039;Chernoff Bound&#039;&#039;&#039;&#039;&#039;&lt;br /&gt;
:&amp;lt;math&amp;gt;&lt;br /&gt;
\begin{align}&lt;br /&gt;
\mathbf{Pr}[|X| \geq \delta] \leq \min_{t \geq 0} \frac{\mathbf{E}[e^{t|X|}]}{e^{t\delta}}&lt;br /&gt;
\end{align}&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&#039;&#039;&#039;&#039;&#039;&amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;th-Moment Bound&#039;&#039;&#039;&#039;&#039;&lt;br /&gt;
:&amp;lt;math&amp;gt;&lt;br /&gt;
\begin{align}&lt;br /&gt;
\mathbf{Pr}[|X| \geq \delta] \leq \frac{\mathbf{E}[|X|^k]}{\delta^k}&lt;br /&gt;
\end{align}&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
# Show that for each &amp;lt;math&amp;gt;\delta&amp;lt;/math&amp;gt;, there exists a choice of &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; such that the &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;th-moment bound is no weaker than the Chernoff bound. (Hint: Use the probabilistic method.)&lt;br /&gt;
# Why would we still prefer the Chernoff bound to the (seemingly) stronger &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;-th moment bound?&lt;br /&gt;
&lt;br /&gt;
* [&#039;&#039;&#039;Cut size in random graph&#039;&#039;&#039;] Show that with probability at least &amp;lt;math&amp;gt;2/3&amp;lt;/math&amp;gt;, the size of the max-cut in [[wikipedia:Erdős–Rényi_model#Definition|Erdős–Rényi random graph]] &amp;lt;math&amp;gt;G(n,1/2)&amp;lt;/math&amp;gt; is at most &amp;lt;math&amp;gt;n^2/8 + O(n^{1.5})&amp;lt;/math&amp;gt;. In the &amp;lt;math&amp;gt;G(n,1/2)&amp;lt;/math&amp;gt; model, each edge is included in the graph with probability &amp;lt;math&amp;gt;1/2&amp;lt;/math&amp;gt;, independently of every other edge.&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Modes of Convergence and Characteristic Function, 15 points) (&amp;lt;strong&amp;gt;Bonus problem&amp;lt;/strong&amp;gt;)==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Connection of convergence modes (I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;(X_n)_{n \ge 1}, (Y_n)_{n \ge 1}, X, Y&amp;lt;/math&amp;gt; be random variables and &amp;lt;math&amp;gt;c\in\mathbb{R}&amp;lt;/math&amp;gt; be a real number. &lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; Suppose &amp;lt;math&amp;gt;X_n \overset{D}{\to} X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_n \overset{D}{\to} c&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;X_nY_n \overset{D}{\to} cX&amp;lt;/math&amp;gt;. &lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Construct an example such that &amp;lt;math&amp;gt;X_n \overset{D}{\to} X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_n \overset{D}{\to} Y&amp;lt;/math&amp;gt; but &amp;lt;math&amp;gt;X_nY_n&amp;lt;/math&amp;gt; does not converge to &amp;lt;math&amp;gt;XY&amp;lt;/math&amp;gt; in distribution.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Connection of convergence modes (II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;(X_n)_{n \ge 1}, X&amp;lt;/math&amp;gt; be random variables. Prove that &amp;lt;math&amp;gt;X_n \overset{P}{\to} X&amp;lt;/math&amp;gt; if and only if for every subsequence &amp;lt;math&amp;gt;X_{n(m)}&amp;lt;/math&amp;gt;, there exists a further subsequence &amp;lt;math&amp;gt;Y_k = X_{n(m_k)}&amp;lt;/math&amp;gt; that converges almost surely to &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;. (Hint: you may use the first Borel-Cantelli lemma.)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Characteristic Function&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; is a discrete random variable only taking integer values and &amp;lt;math&amp;gt;\mathbf{E}[X]=0&amp;lt;/math&amp;gt;. Suppose furthermore that there is no infinite subprogression &amp;lt;math&amp;gt;a+q\mathbb{Z}&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;\mathbb{Z}&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;q&amp;gt;1&amp;lt;/math&amp;gt; for which &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; takes values almost surely in &amp;lt;math&amp;gt;a+q \mathbb{Z}&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;|\phi_X(t)|&amp;lt;1&amp;lt;/math&amp;gt; for &amp;lt;math&amp;gt;0 &amp;lt; t \le \pi&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\phi_X(t)=\mathbf{E}[\mathrm{e}^{itX}]&amp;lt;/math&amp;gt; is the characteristic function of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 4 (LLN and CLT, 15 points + 5 points) ==&lt;br /&gt;
&amp;lt;strong&amp;gt;In this problem, you may apply the results of Laws of Large Numbers (LLN) and the Central Limit Theorem (CLT) to solve the problems.&amp;lt;/strong&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;St. Petersburg paradox&amp;lt;/strong&amp;gt;] Consider the well-known game involving a fair coin. In this game, if it takes &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; tosses to obtain a head, you will win &amp;lt;math&amp;gt;2^k&amp;lt;/math&amp;gt; dollars as the reward. Despite the game&#039;s expected reward being infinite, people tend to offer relatively modest amounts to participate. The following provides a mathematical explanation for this phenomenon.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
For each &amp;lt;math&amp;gt;n \ge 1&amp;lt;/math&amp;gt;, let &amp;lt;math&amp;gt;X_{n,1}, X_{n,2},\ldots, X_{n,k}&amp;lt;/math&amp;gt; be independent random variables. Furthermore, let &amp;lt;math&amp;gt;b_n &amp;gt; 0&amp;lt;/math&amp;gt; be real numbers with &amp;lt;math&amp;gt;b_n \to \infty&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widetilde{X}_{n,k} = X_{n,k} \mathbf{1}_{|X_{n,k}| \le b_n}&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;1 \le k \le n&amp;lt;/math&amp;gt;. If &amp;lt;math&amp;gt;\sum_{k=1}^n \mathbf{Pr}(|X_{n,k}| &amp;gt; b_n) \to 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;b_n^{-2} \sum_{k=1}^n \mathbf{E}[\widetilde{X}_{n,k}^2] \to 0&amp;lt;/math&amp;gt; when &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, then &amp;lt;math&amp;gt;(S_n-a_n)/b_n \overset{P}{\to} 0 &amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;S_n = \sum_{k=1}^n X_{n,k}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;a_n = \sum_{k=1}^n \mathbf{E}[\widetilde{X}_{n,k}]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;S_n&amp;lt;/math&amp;gt; be the total winnings after playing &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds of the game. Prove that &amp;lt;math&amp;gt;\frac{S_n}{n \log_2 n} \overset{P}{\to} 1&amp;lt;/math&amp;gt;. (Therefore, a fair price to play this game &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; times is roughly &amp;lt;math&amp;gt;n \log_2 n&amp;lt;/math&amp;gt; dollars)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; (&amp;lt;strong&amp;gt;Bonus problem, 5 points&amp;lt;/strong&amp;gt;)&lt;br /&gt;
Let &amp;lt;math&amp;gt;S_n&amp;lt;/math&amp;gt; be the total winnings after playing &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds of the game. Prove that &amp;lt;math&amp;gt; \limsup_{n \to \infty} \frac{S_n}{n \log_2 n} = \infty&amp;lt;/math&amp;gt; almost surely. (Hint: You may use Borel-Cantelli lemmas)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Monte Carlo Integration&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; be a continuous function on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;U_1,U_2,\ldots,U_n \sim U(0,1)&amp;lt;/math&amp;gt; be independent random variables. Show that &amp;lt;math&amp;gt;I = \frac{1}{n}\sum_{i=1}^n f(U_i) \to \int_0^1 f(x) \mathrm{d}x&amp;lt;/math&amp;gt; in probability. (Remark: This holds so long as &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; is a measurable function on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;\int_0^1 |f(x)| \mathrm{d}x&amp;lt;/math&amp;gt;)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Square root&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X_1,X_2,\ldots&amp;lt;/math&amp;gt; be i.i.d. random variables with &amp;lt;math&amp;gt;X_i \ge 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{E}[X_1] = 1&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{Var}[X_1] = \sigma^2 \in (0,+\infty)&amp;lt;/math&amp;gt;. Show &amp;lt;math&amp;gt;\sqrt{S_n} - \sqrt{n} \overset{D}{\to} N(0,1)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;S_n = \sum_{i=1}^n X_i&amp;lt;/math&amp;gt;. (Hint: You may use the statement in Problem 3.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2024)/Problem_Set_4&amp;diff=12445</id>
		<title>概率论与数理统计 (Spring 2024)/Problem Set 4</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2024)/Problem_Set_4&amp;diff=12445"/>
		<updated>2024-05-24T13:31:17Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: /* Problem 3 (Modes of Convergence and Characteristic Function, 20 points) (Bonus problem) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
*Bonus problem为附加题（选做）。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 4==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Continuous Random Variables, 30 points)==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Density function&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Determine the value of &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;f(x) = C\exp(-x-e^{-x}), x\in \mathbb{R}&amp;lt;/math&amp;gt; is a probability density function (PDF) for a continuous random variable.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Independence&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be independent and identically distributed continuous random variables with cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; and probability density function (PDF) &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt;. Find out the density functions of &amp;lt;math&amp;gt;V = \max\{X,Y\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;U = \min\{X,Y\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Correlation&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be uniformly distributed on &amp;lt;math&amp;gt;(-1,1)&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_k = \cos(k \pi X)&amp;lt;/math&amp;gt; for &amp;lt;math&amp;gt;k=1,2,\ldots,n&amp;lt;/math&amp;gt;. Are the random variables &amp;lt;math&amp;gt;Y_1, Y_2, \ldots, Y_n&amp;lt;/math&amp;gt; correlated? independent? You should prove your claim rigorously.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Expectation of random variables (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a continuous random variable with mean &amp;lt;math&amp;gt;\mu&amp;lt;/math&amp;gt; and cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X \ge 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\int_{0}^a F(x) dx = \int_{a}^{\infty} [1-F(x)] dx&amp;lt;/math&amp;gt; if and only if &amp;lt;math&amp;gt;a = \mu&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; has finite variance. Show that &amp;lt;math&amp;gt;g(a) = \mathbb{E}((X-a)^2)&amp;lt;/math&amp;gt; achieves the minimum when &amp;lt;math&amp;gt;a = \mu&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Expectation of random variables (II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X, Y&amp;lt;/math&amp;gt; be two independent and identically distributed continuous random variables with cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;. Furthermore, &amp;lt;math&amp;gt;X,Y \ge 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbb{E}[|X-Y|] = 2 \left(\mathbb{E}[X] - \int_{0}^{\infty} (1-F(x))^2 dx\right)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Conditional distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be two random variables. The joint density of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; is given by &amp;lt;math&amp;gt;f(x,y) = c(x^2 - y^2)e^{-x}&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;0\leq x &amp;lt;\infty&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-x\leq y \leq x&amp;lt;/math&amp;gt;. Here, &amp;lt;math&amp;gt;c\in \mathbb{R}_+&amp;lt;/math&amp;gt; is a constant. Find out the conditional distribution of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;, given &amp;lt;math&amp;gt;X = x&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Uniform Distribution (I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;P_i = (X_i,Y_i), 1\leq i\leq n&amp;lt;/math&amp;gt;, be independent, uniformly distributed points in the unit square &amp;lt;math&amp;gt;[0,1]^2&amp;lt;/math&amp;gt;. A point &amp;lt;math&amp;gt;P_i&amp;lt;/math&amp;gt; is called &amp;quot;peripheral&amp;quot; if, for all &amp;lt;math&amp;gt;r = 1,2,\cdots,n&amp;lt;/math&amp;gt;, either &amp;lt;math&amp;gt;X_r \leq X_i&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;Y_r \leq Y_i&amp;lt;/math&amp;gt;, or both. Find out the expected number of peripheral points.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Uniform Distribution (II)&amp;lt;/strong&amp;gt;] Derive the moment generating function of the standard uniform distribution, i.e., uniform distribution on &amp;lt;math&amp;gt;(0,1)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Exponential distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; have an exponential distribution. Show that &amp;lt;math&amp;gt;\textbf{Pr}[X&amp;gt;s+x|X&amp;gt;s] = \textbf{Pr}[X&amp;gt;x]&amp;lt;/math&amp;gt;, for &amp;lt;math&amp;gt;x,s\geq 0&amp;lt;/math&amp;gt;. This is the memoryless property. Show that the exponential distribution is the only continuous distribution with this property.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Normal distribution(I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X,Y\sim N(0,1)&amp;lt;/math&amp;gt; be two independent and identically distributed normal random variables. Let &amp;lt;math&amp;gt;Z = X-Y&amp;lt;/math&amp;gt;. Find the density function of &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;|Z|&amp;lt;/math&amp;gt; respectively.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Normal distribution(II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; have the &amp;lt;math&amp;gt;N(0,1)&amp;lt;/math&amp;gt; distribution and let &amp;lt;math&amp;gt;a&amp;gt;0&amp;lt;/math&amp;gt;. Show that the random variable &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; given by&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{equation*}&lt;br /&gt;
Y = \begin{cases}&lt;br /&gt;
X, &amp;amp; |X|&amp;lt; a \\&lt;br /&gt;
-X, &amp;amp; |X|\geq a&lt;br /&gt;
\end{cases}&lt;br /&gt;
\end{equation*}&amp;lt;/math&amp;gt;&lt;br /&gt;
has the &amp;lt;math&amp;gt;N(0,1)&amp;lt;/math&amp;gt; distribution, and find an expression for &amp;lt;math&amp;gt;\rho(a) = \textbf{Cov}(X,Y)&amp;lt;/math&amp;gt; in terms of the density function &amp;lt;math&amp;gt;\phi&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random process&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Given a real number &amp;lt;math&amp;gt;U&amp;lt;1&amp;lt;/math&amp;gt; as input of the following process, find out the expected returning value.&lt;br /&gt;
{{Theorem|&#039;&#039;Process&#039;&#039;|&lt;br /&gt;
:&#039;&#039;&#039;Input:&#039;&#039;&#039;  real numbers &amp;lt;math&amp;gt;U &amp;lt; 1&amp;lt;/math&amp;gt;;&lt;br /&gt;
----&lt;br /&gt;
:initialize &amp;lt;math&amp;gt;x = 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;count = 0&amp;lt;/math&amp;gt;;&lt;br /&gt;
:while &amp;lt;math&amp;gt; x &amp;lt; U &amp;lt;/math&amp;gt; do&lt;br /&gt;
:* choose &amp;lt;math&amp;gt;y \in (0,1)&amp;lt;/math&amp;gt; uniformly at random;&lt;br /&gt;
:* update &amp;lt;math&amp;gt;x = x + y&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;count = count + 1&amp;lt;/math&amp;gt;;&lt;br /&gt;
:return &amp;lt;math&amp;gt;count&amp;lt;/math&amp;gt;;&lt;br /&gt;
}}&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Random semicircle&amp;lt;/strong&amp;gt;] We sample &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; points within a circle &amp;lt;math&amp;gt;C=\{(x,y) \in \mathbb{R}^2 \mid x^2+y^2 \le 1\}&amp;lt;/math&amp;gt; independently and uniformly at random (i.e., the density function &amp;lt;math&amp;gt;f(x,y) \propto 1_{(x,y) \in C}&amp;lt;/math&amp;gt;). Find out the probability that they all lie within some semicircle of the original circle &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;. (Hint: you may apply the technique of change of variables, see [https://en.wikipedia.org/wiki/Random_variable#Functions_of_random_variables function of random variables] or Chapter 4.7 in [GS])&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Stochastic domination&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X, Y&amp;lt;/math&amp;gt; be continuous random variables. Show that &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; dominates &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; stochastically if and only if &amp;lt;math&amp;gt;\mathbb{E}[f(X)]\geq \mathbb{E}[f(Y)]&amp;lt;/math&amp;gt; for any non-decreasing function &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; for which the expectations exist.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Concentration of measure, 10 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Tossing coins&amp;lt;/strong&amp;gt;] We repeatedly toss a fair coin (with an equal probability of heads and tails). Let the random variable &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of throws required to obtain a total of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; heads. Show that &amp;lt;math&amp;gt;\textbf{Pr}[X &amp;gt; 2n + 2\sqrt{n\log n}]\leq O(1/n)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Chernoff vs Chebyshev&amp;lt;/strong&amp;gt;] We have a standard six-sided die. Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of times a 6 occurs in &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; throws off the die. Compare the best upper bounds on &amp;lt;math&amp;gt;\textbf{Pr}[X\geq n/4]&amp;lt;/math&amp;gt; that you can obtain using Chebyshev&#039;s inequality and Chernoff bounds.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;&amp;lt;strong&amp;gt;-th moment bound]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a random variable with expectation &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; such that moment generating function &amp;lt;math&amp;gt;\mathbf{E}[\exp(t|X|)]&amp;lt;/math&amp;gt; is finite for some &amp;lt;math&amp;gt; t &amp;gt; 0 &amp;lt;/math&amp;gt;. We can use the following two kinds of tail inequalities for &amp;lt;math&amp;gt; X &amp;lt;/math&amp;gt;: &lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&amp;lt;nowiki&amp;gt;     &amp;lt;/nowiki&amp;gt;&#039;&#039;&#039;&#039;&#039;Chernoff Bound&#039;&#039;&#039;&#039;&#039;&lt;br /&gt;
:&amp;lt;math&amp;gt;&lt;br /&gt;
\begin{align}&lt;br /&gt;
\mathbf{Pr}[|X| \geq \delta] \leq \min_{t \geq 0} \frac{\mathbf{E}[e^{t|X|}]}{e^{t\delta}}&lt;br /&gt;
\end{align}&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&#039;&#039;&#039;&#039;&#039;&amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;th-Moment Bound&#039;&#039;&#039;&#039;&#039;&lt;br /&gt;
:&amp;lt;math&amp;gt;&lt;br /&gt;
\begin{align}&lt;br /&gt;
\mathbf{Pr}[|X| \geq \delta] \leq \frac{\mathbf{E}[|X|^k]}{\delta^k}&lt;br /&gt;
\end{align}&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
# Show that for each &amp;lt;math&amp;gt;\delta&amp;lt;/math&amp;gt;, there exists a choice of &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; such that the &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;th-moment bound is no weaker than the Chernoff bound. (Hint: Use the probabilistic method.)&lt;br /&gt;
# Why would we still prefer the Chernoff bound to the (seemingly) stronger &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;-th moment bound?&lt;br /&gt;
&lt;br /&gt;
* [&#039;&#039;&#039;Cut size in random graph&#039;&#039;&#039;] Show that with probability at least &amp;lt;math&amp;gt;2/3&amp;lt;/math&amp;gt;, the size of the max-cut in [[wikipedia:Erdős–Rényi_model#Definition|Erdős–Rényi random graph]] &amp;lt;math&amp;gt;G(n,1/2)&amp;lt;/math&amp;gt; is at most &amp;lt;math&amp;gt;n^2/8 + O(n^{1.5})&amp;lt;/math&amp;gt;. In the &amp;lt;math&amp;gt;G(n,1/2)&amp;lt;/math&amp;gt; model, each edge is included in the graph with probability &amp;lt;math&amp;gt;1/2&amp;lt;/math&amp;gt;, independently of every other edge.&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Modes of Convergence and Characteristic Function, 20 points) (&amp;lt;strong&amp;gt;Bonus problem&amp;lt;/strong&amp;gt;)==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Connection of convergence modes (I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;(X_n)_{n \ge 1}, (Y_n)_{n \ge 1}, X, Y&amp;lt;/math&amp;gt; be random variables and &amp;lt;math&amp;gt;c\in\mathbb{R}&amp;lt;/math&amp;gt; be a real number. &lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; Suppose &amp;lt;math&amp;gt;X_n \overset{D}{\to} X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_n \overset{D}{\to} c&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;X_nY_n \overset{D}{\to} cX&amp;lt;/math&amp;gt;. &lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Construct an example such that &amp;lt;math&amp;gt;X_n \overset{D}{\to} X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_n \overset{D}{\to} Y&amp;lt;/math&amp;gt; but &amp;lt;math&amp;gt;X_nY_n&amp;lt;/math&amp;gt; does not converge to &amp;lt;math&amp;gt;XY&amp;lt;/math&amp;gt; in distribution.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Connection of convergence modes (II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;(X_n)_{n \ge 1}, X&amp;lt;/math&amp;gt; be random variables. Prove that &amp;lt;math&amp;gt;X_n \overset{P}{\to} X&amp;lt;/math&amp;gt; if and only if for every subsequence &amp;lt;math&amp;gt;X_{n(m)}&amp;lt;/math&amp;gt;, there exists a further subsequence &amp;lt;math&amp;gt;Y_k = X_{n(m_k)}&amp;lt;/math&amp;gt; that converges almost surely to &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;. (Hint: you may use the first Borel-Cantelli lemma.)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Extension of Borel-Cantelli Lemma&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;(A_n)_{n \ge 1}&amp;lt;/math&amp;gt; be events. Suppose &amp;lt;math&amp;gt;\sum_{n \ge 1} \mathbf{Pr}(A_n)=+\infty&amp;lt;/math&amp;gt;. Show that &lt;br /&gt;
&amp;lt;math&amp;gt;\mathbf{Pr}(A_n \text{ i.o.}) \ge \limsup_{n \to \infty} \frac{ \left(\sum_{k=1}^n\mathbf{Pr}(A_k)\right)^2 }{\sum_{1\le j,k \le n} \mathbf{Pr}(A_j \cap A_k)}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Characteristic Function&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; is a discrete random variable only taking integer values and &amp;lt;math&amp;gt;\mathbf{E}[X]=0&amp;lt;/math&amp;gt;. Suppose furthermore that there is no infinite subprogression &amp;lt;math&amp;gt;a+q\mathbb{Z}&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;\mathbb{Z}&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;q&amp;gt;1&amp;lt;/math&amp;gt; for which &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; takes values almost surely in &amp;lt;math&amp;gt;a+q \mathbb{Z}&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;|\phi_X(t)|&amp;lt;1&amp;lt;/math&amp;gt; for &amp;lt;math&amp;gt;0 &amp;lt; t \le \pi&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\phi_X(t)=\mathbf{E}[\mathrm{e}^{itX}]&amp;lt;/math&amp;gt; is the characteristic function of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 4 (LLN and CLT, 15 points + 5 points) ==&lt;br /&gt;
&amp;lt;strong&amp;gt;In this problem, you may apply the results of Laws of Large Numbers (LLN) and the Central Limit Theorem (CLT) to solve the problems.&amp;lt;/strong&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;St. Petersburg paradox&amp;lt;/strong&amp;gt;] Consider the well-known game involving a fair coin. In this game, if it takes &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; tosses to obtain a head, you will win &amp;lt;math&amp;gt;2^k&amp;lt;/math&amp;gt; dollars as the reward. Despite the game&#039;s expected reward being infinite, people tend to offer relatively modest amounts to participate. The following provides a mathematical explanation for this phenomenon.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
For each &amp;lt;math&amp;gt;n \ge 1&amp;lt;/math&amp;gt;, let &amp;lt;math&amp;gt;X_{n,1}, X_{n,2},\ldots, X_{n,k}&amp;lt;/math&amp;gt; be independent random variables. Furthermore, let &amp;lt;math&amp;gt;b_n &amp;gt; 0&amp;lt;/math&amp;gt; be real numbers with &amp;lt;math&amp;gt;b_n \to \infty&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widetilde{X}_{n,k} = X_{n,k} \mathbf{1}_{|X_{n,k}| \le b_n}&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;1 \le k \le n&amp;lt;/math&amp;gt;. If &amp;lt;math&amp;gt;\sum_{k=1}^n \mathbf{Pr}(|X_{n,k}| &amp;gt; b_n) \to 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;b_n^{-2} \sum_{k=1}^n \mathbf{E}[\widetilde{X}_{n,k}^2] \to 0&amp;lt;/math&amp;gt; when &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, then &amp;lt;math&amp;gt;(S_n-a_n)/b_n \overset{P}{\to} 0 &amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;S_n = \sum_{k=1}^n X_{n,k}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;a_n = \sum_{k=1}^n \mathbf{E}[\widetilde{X}_{n,k}]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;S_n&amp;lt;/math&amp;gt; be the total winnings after playing &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds of the game. Prove that &amp;lt;math&amp;gt;\frac{S_n}{n \log_2 n} \overset{P}{\to} 1&amp;lt;/math&amp;gt;. (Therefore, a fair price to play this game &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; times is roughly &amp;lt;math&amp;gt;n \log_2 n&amp;lt;/math&amp;gt; dollars)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; (&amp;lt;strong&amp;gt;Bonus problem, 5 points&amp;lt;/strong&amp;gt;)&lt;br /&gt;
Let &amp;lt;math&amp;gt;S_n&amp;lt;/math&amp;gt; be the total winnings after playing &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds of the game. Prove that &amp;lt;math&amp;gt; \limsup_{n \to \infty} \frac{S_n}{n \log_2 n} = \infty&amp;lt;/math&amp;gt; almost surely. (Hint: You may use Borel-Cantelli lemmas)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Monte Carlo Integration&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; be a continuous function on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;U_1,U_2,\ldots,U_n \sim U(0,1)&amp;lt;/math&amp;gt; be independent random variables. Show that &amp;lt;math&amp;gt;I = \frac{1}{n}\sum_{i=1}^n f(U_i) \to \int_0^1 f(x) \mathrm{d}x&amp;lt;/math&amp;gt; in probability. (Remark: This holds so long as &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; is a measurable function on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;\int_0^1 |f(x)| \mathrm{d}x&amp;lt;/math&amp;gt;)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Square root&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X_1,X_2,\ldots&amp;lt;/math&amp;gt; be i.i.d. random variables with &amp;lt;math&amp;gt;X_i \ge 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{E}[X_1] = 1&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{Var}[X_1] = \sigma^2 \in (0,+\infty)&amp;lt;/math&amp;gt;. Show &amp;lt;math&amp;gt;\sqrt{S_n} - \sqrt{n} \overset{D}{\to} N(0,1)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;S_n = \sum_{i=1}^n X_i&amp;lt;/math&amp;gt;. (Hint: You may use the statement in Problem 3.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2024)/Problem_Set_4&amp;diff=12444</id>
		<title>概率论与数理统计 (Spring 2024)/Problem Set 4</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2024)/Problem_Set_4&amp;diff=12444"/>
		<updated>2024-05-24T13:27:55Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: /* Problem 3 (Modes of Convergence, 15 points) (Bonus problem) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
*Bonus problem为附加题（选做）。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 4==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Continuous Random Variables, 30 points)==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Density function&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Determine the value of &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;f(x) = C\exp(-x-e^{-x}), x\in \mathbb{R}&amp;lt;/math&amp;gt; is a probability density function (PDF) for a continuous random variable.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Independence&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be independent and identically distributed continuous random variables with cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; and probability density function (PDF) &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt;. Find out the density functions of &amp;lt;math&amp;gt;V = \max\{X,Y\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;U = \min\{X,Y\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Correlation&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be uniformly distributed on &amp;lt;math&amp;gt;(-1,1)&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_k = \cos(k \pi X)&amp;lt;/math&amp;gt; for &amp;lt;math&amp;gt;k=1,2,\ldots,n&amp;lt;/math&amp;gt;. Are the random variables &amp;lt;math&amp;gt;Y_1, Y_2, \ldots, Y_n&amp;lt;/math&amp;gt; correlated? independent? You should prove your claim rigorously.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Expectation of random variables (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a continuous random variable with mean &amp;lt;math&amp;gt;\mu&amp;lt;/math&amp;gt; and cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X \ge 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\int_{0}^a F(x) dx = \int_{a}^{\infty} [1-F(x)] dx&amp;lt;/math&amp;gt; if and only if &amp;lt;math&amp;gt;a = \mu&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; has finite variance. Show that &amp;lt;math&amp;gt;g(a) = \mathbb{E}((X-a)^2)&amp;lt;/math&amp;gt; achieves the minimum when &amp;lt;math&amp;gt;a = \mu&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Expectation of random variables (II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X, Y&amp;lt;/math&amp;gt; be two independent and identically distributed continuous random variables with cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;. Furthermore, &amp;lt;math&amp;gt;X,Y \ge 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbb{E}[|X-Y|] = 2 \left(\mathbb{E}[X] - \int_{0}^{\infty} (1-F(x))^2 dx\right)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Conditional distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be two random variables. The joint density of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; is given by &amp;lt;math&amp;gt;f(x,y) = c(x^2 - y^2)e^{-x}&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;0\leq x &amp;lt;\infty&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-x\leq y \leq x&amp;lt;/math&amp;gt;. Here, &amp;lt;math&amp;gt;c\in \mathbb{R}_+&amp;lt;/math&amp;gt; is a constant. Find out the conditional distribution of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;, given &amp;lt;math&amp;gt;X = x&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Uniform Distribution (I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;P_i = (X_i,Y_i), 1\leq i\leq n&amp;lt;/math&amp;gt;, be independent, uniformly distributed points in the unit square &amp;lt;math&amp;gt;[0,1]^2&amp;lt;/math&amp;gt;. A point &amp;lt;math&amp;gt;P_i&amp;lt;/math&amp;gt; is called &amp;quot;peripheral&amp;quot; if, for all &amp;lt;math&amp;gt;r = 1,2,\cdots,n&amp;lt;/math&amp;gt;, either &amp;lt;math&amp;gt;X_r \leq X_i&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;Y_r \leq Y_i&amp;lt;/math&amp;gt;, or both. Find out the expected number of peripheral points.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Uniform Distribution (II)&amp;lt;/strong&amp;gt;] Derive the moment generating function of the standard uniform distribution, i.e., uniform distribution on &amp;lt;math&amp;gt;(0,1)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Exponential distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; have an exponential distribution. Show that &amp;lt;math&amp;gt;\textbf{Pr}[X&amp;gt;s+x|X&amp;gt;s] = \textbf{Pr}[X&amp;gt;x]&amp;lt;/math&amp;gt;, for &amp;lt;math&amp;gt;x,s\geq 0&amp;lt;/math&amp;gt;. This is the memoryless property. Show that the exponential distribution is the only continuous distribution with this property.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Normal distribution(I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X,Y\sim N(0,1)&amp;lt;/math&amp;gt; be two independent and identically distributed normal random variables. Let &amp;lt;math&amp;gt;Z = X-Y&amp;lt;/math&amp;gt;. Find the density function of &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;|Z|&amp;lt;/math&amp;gt; respectively.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Normal distribution(II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; have the &amp;lt;math&amp;gt;N(0,1)&amp;lt;/math&amp;gt; distribution and let &amp;lt;math&amp;gt;a&amp;gt;0&amp;lt;/math&amp;gt;. Show that the random variable &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; given by&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{equation*}&lt;br /&gt;
Y = \begin{cases}&lt;br /&gt;
X, &amp;amp; |X|&amp;lt; a \\&lt;br /&gt;
-X, &amp;amp; |X|\geq a&lt;br /&gt;
\end{cases}&lt;br /&gt;
\end{equation*}&amp;lt;/math&amp;gt;&lt;br /&gt;
has the &amp;lt;math&amp;gt;N(0,1)&amp;lt;/math&amp;gt; distribution, and find an expression for &amp;lt;math&amp;gt;\rho(a) = \textbf{Cov}(X,Y)&amp;lt;/math&amp;gt; in terms of the density function &amp;lt;math&amp;gt;\phi&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random process&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Given a real number &amp;lt;math&amp;gt;U&amp;lt;1&amp;lt;/math&amp;gt; as input of the following process, find out the expected returning value.&lt;br /&gt;
{{Theorem|&#039;&#039;Process&#039;&#039;|&lt;br /&gt;
:&#039;&#039;&#039;Input:&#039;&#039;&#039;  real numbers &amp;lt;math&amp;gt;U &amp;lt; 1&amp;lt;/math&amp;gt;;&lt;br /&gt;
----&lt;br /&gt;
:initialize &amp;lt;math&amp;gt;x = 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;count = 0&amp;lt;/math&amp;gt;;&lt;br /&gt;
:while &amp;lt;math&amp;gt; x &amp;lt; U &amp;lt;/math&amp;gt; do&lt;br /&gt;
:* choose &amp;lt;math&amp;gt;y \in (0,1)&amp;lt;/math&amp;gt; uniformly at random;&lt;br /&gt;
:* update &amp;lt;math&amp;gt;x = x + y&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;count = count + 1&amp;lt;/math&amp;gt;;&lt;br /&gt;
:return &amp;lt;math&amp;gt;count&amp;lt;/math&amp;gt;;&lt;br /&gt;
}}&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Random semicircle&amp;lt;/strong&amp;gt;] We sample &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; points within a circle &amp;lt;math&amp;gt;C=\{(x,y) \in \mathbb{R}^2 \mid x^2+y^2 \le 1\}&amp;lt;/math&amp;gt; independently and uniformly at random (i.e., the density function &amp;lt;math&amp;gt;f(x,y) \propto 1_{(x,y) \in C}&amp;lt;/math&amp;gt;). Find out the probability that they all lie within some semicircle of the original circle &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;. (Hint: you may apply the technique of change of variables, see [https://en.wikipedia.org/wiki/Random_variable#Functions_of_random_variables function of random variables] or Chapter 4.7 in [GS])&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Stochastic domination&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X, Y&amp;lt;/math&amp;gt; be continuous random variables. Show that &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; dominates &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; stochastically if and only if &amp;lt;math&amp;gt;\mathbb{E}[f(X)]\geq \mathbb{E}[f(Y)]&amp;lt;/math&amp;gt; for any non-decreasing function &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; for which the expectations exist.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Concentration of measure, 10 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Tossing coins&amp;lt;/strong&amp;gt;] We repeatedly toss a fair coin (with an equal probability of heads and tails). Let the random variable &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of throws required to obtain a total of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; heads. Show that &amp;lt;math&amp;gt;\textbf{Pr}[X &amp;gt; 2n + 2\sqrt{n\log n}]\leq O(1/n)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Chernoff vs Chebyshev&amp;lt;/strong&amp;gt;] We have a standard six-sided die. Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of times a 6 occurs in &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; throws off the die. Compare the best upper bounds on &amp;lt;math&amp;gt;\textbf{Pr}[X\geq n/4]&amp;lt;/math&amp;gt; that you can obtain using Chebyshev&#039;s inequality and Chernoff bounds.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;&amp;lt;strong&amp;gt;-th moment bound]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a random variable with expectation &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; such that moment generating function &amp;lt;math&amp;gt;\mathbf{E}[\exp(t|X|)]&amp;lt;/math&amp;gt; is finite for some &amp;lt;math&amp;gt; t &amp;gt; 0 &amp;lt;/math&amp;gt;. We can use the following two kinds of tail inequalities for &amp;lt;math&amp;gt; X &amp;lt;/math&amp;gt;: &lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&amp;lt;nowiki&amp;gt;     &amp;lt;/nowiki&amp;gt;&#039;&#039;&#039;&#039;&#039;Chernoff Bound&#039;&#039;&#039;&#039;&#039;&lt;br /&gt;
:&amp;lt;math&amp;gt;&lt;br /&gt;
\begin{align}&lt;br /&gt;
\mathbf{Pr}[|X| \geq \delta] \leq \min_{t \geq 0} \frac{\mathbf{E}[e^{t|X|}]}{e^{t\delta}}&lt;br /&gt;
\end{align}&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&#039;&#039;&#039;&#039;&#039;&amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;th-Moment Bound&#039;&#039;&#039;&#039;&#039;&lt;br /&gt;
:&amp;lt;math&amp;gt;&lt;br /&gt;
\begin{align}&lt;br /&gt;
\mathbf{Pr}[|X| \geq \delta] \leq \frac{\mathbf{E}[|X|^k]}{\delta^k}&lt;br /&gt;
\end{align}&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
# Show that for each &amp;lt;math&amp;gt;\delta&amp;lt;/math&amp;gt;, there exists a choice of &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; such that the &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;th-moment bound is no weaker than the Chernoff bound. (Hint: Use the probabilistic method.)&lt;br /&gt;
# Why would we still prefer the Chernoff bound to the (seemingly) stronger &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;-th moment bound?&lt;br /&gt;
&lt;br /&gt;
* [&#039;&#039;&#039;Cut size in random graph&#039;&#039;&#039;] Show that with probability at least &amp;lt;math&amp;gt;2/3&amp;lt;/math&amp;gt;, the size of the max-cut in [[wikipedia:Erdős–Rényi_model#Definition|Erdős–Rényi random graph]] &amp;lt;math&amp;gt;G(n,1/2)&amp;lt;/math&amp;gt; is at most &amp;lt;math&amp;gt;n^2/8 + O(n^{1.5})&amp;lt;/math&amp;gt;. In the &amp;lt;math&amp;gt;G(n,1/2)&amp;lt;/math&amp;gt; model, each edge is included in the graph with probability &amp;lt;math&amp;gt;1/2&amp;lt;/math&amp;gt;, independently of every other edge.&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Modes of Convergence and Characteristic Function, 20 points) (&amp;lt;strong&amp;gt;Bonus problem&amp;lt;/strong&amp;gt;)==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Connection of convergence modes (I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;(X_n)_{n \ge 1}, (Y_n)_{n \ge 1}, X, Y&amp;lt;/math&amp;gt; be random variables and &amp;lt;math&amp;gt;c\in\mathbb{R}&amp;lt;/math&amp;gt; be a real number. &lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; Suppose &amp;lt;math&amp;gt;X_n \overset{D}{\to} X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_n \overset{D}{\to} c&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;X_nY_n \overset{D}{\to} cX&amp;lt;/math&amp;gt;. &lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Construct an example such that &amp;lt;math&amp;gt;X_n \overset{D}{\to} X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_n \overset{D}{\to} Y&amp;lt;/math&amp;gt; but &amp;lt;math&amp;gt;X_nY_n&amp;lt;/math&amp;gt; does not converge to &amp;lt;math&amp;gt;XY&amp;lt;/math&amp;gt; in distribution.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Connection of convergence modes (II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;(X_n)_{n \ge 1}, X&amp;lt;/math&amp;gt; be random variables. Prove that &amp;lt;math&amp;gt;X_n \overset{P}{\to} X&amp;lt;/math&amp;gt; if and only if for every subsequence &amp;lt;math&amp;gt;X_{n(m)}&amp;lt;/math&amp;gt;, there exists a further subsequence &amp;lt;math&amp;gt;Y_k = X_{n(m_k)}&amp;lt;/math&amp;gt; that converges almost surely to &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;. (Hint: you may use the first Borel-Cantelli lemma.)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Extension of Borel-Cantelli Lemma&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;(A_n)_{n \ge 1}&amp;lt;/math&amp;gt; be events. Suppose &amp;lt;math&amp;gt;\sum_{n \ge 1} \mathbf{Pr}(A_n)=+\infty&amp;lt;/math&amp;gt;. Show that &lt;br /&gt;
&amp;lt;math&amp;gt;\mathbf{Pr}(A_n \text{ i.o.}) \ge \limsup_{n \to \infty} \frac{ \left(\sum_{k=1}^n\mathbf{Pr}(A_k)\right)^2 }{\sum_{1\le j,k \le n} \mathbf{Pr}(A_j \cap A_k)}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Characteristic Function&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; is a discrete random variable only taking integer values and &amp;lt;math&amp;gt;\mathbf{E}[X]=0&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;|\phi_X(t)|&amp;lt;1&amp;lt;/math&amp;gt; for &amp;lt;math&amp;gt;0 &amp;lt; t \le \pi&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\phi_X(t)=\mathbf{E}[\mathrm{e}^{itX}]&amp;lt;/math&amp;gt; is the characteristic function of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 4 (LLN and CLT, 15 points + 5 points) ==&lt;br /&gt;
&amp;lt;strong&amp;gt;In this problem, you may apply the results of Laws of Large Numbers (LLN) and the Central Limit Theorem (CLT) to solve the problems.&amp;lt;/strong&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;St. Petersburg paradox&amp;lt;/strong&amp;gt;] Consider the well-known game involving a fair coin. In this game, if it takes &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; tosses to obtain a head, you will win &amp;lt;math&amp;gt;2^k&amp;lt;/math&amp;gt; dollars as the reward. Despite the game&#039;s expected reward being infinite, people tend to offer relatively modest amounts to participate. The following provides a mathematical explanation for this phenomenon.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
For each &amp;lt;math&amp;gt;n \ge 1&amp;lt;/math&amp;gt;, let &amp;lt;math&amp;gt;X_{n,1}, X_{n,2},\ldots, X_{n,k}&amp;lt;/math&amp;gt; be independent random variables. Furthermore, let &amp;lt;math&amp;gt;b_n &amp;gt; 0&amp;lt;/math&amp;gt; be real numbers with &amp;lt;math&amp;gt;b_n \to \infty&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widetilde{X}_{n,k} = X_{n,k} \mathbf{1}_{|X_{n,k}| \le b_n}&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;1 \le k \le n&amp;lt;/math&amp;gt;. If &amp;lt;math&amp;gt;\sum_{k=1}^n \mathbf{Pr}(|X_{n,k}| &amp;gt; b_n) \to 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;b_n^{-2} \sum_{k=1}^n \mathbf{E}[\widetilde{X}_{n,k}^2] \to 0&amp;lt;/math&amp;gt; when &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, then &amp;lt;math&amp;gt;(S_n-a_n)/b_n \overset{P}{\to} 0 &amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;S_n = \sum_{k=1}^n X_{n,k}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;a_n = \sum_{k=1}^n \mathbf{E}[\widetilde{X}_{n,k}]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;S_n&amp;lt;/math&amp;gt; be the total winnings after playing &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds of the game. Prove that &amp;lt;math&amp;gt;\frac{S_n}{n \log_2 n} \overset{P}{\to} 1&amp;lt;/math&amp;gt;. (Therefore, a fair price to play this game &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; times is roughly &amp;lt;math&amp;gt;n \log_2 n&amp;lt;/math&amp;gt; dollars)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; (&amp;lt;strong&amp;gt;Bonus problem, 5 points&amp;lt;/strong&amp;gt;)&lt;br /&gt;
Let &amp;lt;math&amp;gt;S_n&amp;lt;/math&amp;gt; be the total winnings after playing &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds of the game. Prove that &amp;lt;math&amp;gt; \limsup_{n \to \infty} \frac{S_n}{n \log_2 n} = \infty&amp;lt;/math&amp;gt; almost surely. (Hint: You may use Borel-Cantelli lemmas)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Monte Carlo Integration&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; be a continuous function on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;U_1,U_2,\ldots,U_n \sim U(0,1)&amp;lt;/math&amp;gt; be independent random variables. Show that &amp;lt;math&amp;gt;I = \frac{1}{n}\sum_{i=1}^n f(U_i) \to \int_0^1 f(x) \mathrm{d}x&amp;lt;/math&amp;gt; in probability. (Remark: This holds so long as &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; is a measurable function on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;\int_0^1 |f(x)| \mathrm{d}x&amp;lt;/math&amp;gt;)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Square root&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X_1,X_2,\ldots&amp;lt;/math&amp;gt; be i.i.d. random variables with &amp;lt;math&amp;gt;X_i \ge 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{E}[X_1] = 1&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{Var}[X_1] = \sigma^2 \in (0,+\infty)&amp;lt;/math&amp;gt;. Show &amp;lt;math&amp;gt;\sqrt{S_n} - \sqrt{n} \overset{D}{\to} N(0,1)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;S_n = \sum_{i=1}^n X_i&amp;lt;/math&amp;gt;. (Hint: You may use the statement in Problem 3.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2024)/Problem_Set_4&amp;diff=12440</id>
		<title>概率论与数理统计 (Spring 2024)/Problem Set 4</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2024)/Problem_Set_4&amp;diff=12440"/>
		<updated>2024-05-24T13:09:23Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: /* Problem 4 (LLN and CLT, 15 points + 5 points) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
*Bonus problem为附加题（选做）。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 4==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Continuous Random Variables, 30 points)==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Density function&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Determine the value of &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;f(x) = C\exp(-x-e^{-x}), x\in \mathbb{R}&amp;lt;/math&amp;gt; is a probability density function (PDF) for a continuous random variable.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Independence&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be independent and identically distributed continuous random variables with cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; and probability density function (PDF) &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt;. Find out the density functions of &amp;lt;math&amp;gt;V = \max\{X,Y\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;U = \min\{X,Y\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Correlation&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be uniformly distributed on &amp;lt;math&amp;gt;(-1,1)&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_k = \cos(k \pi X)&amp;lt;/math&amp;gt; for &amp;lt;math&amp;gt;k=1,2,\ldots,n&amp;lt;/math&amp;gt;. Are the random variables &amp;lt;math&amp;gt;Y_1, Y_2, \ldots, Y_n&amp;lt;/math&amp;gt; correlated? independent? You should prove your claim rigorously.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Expectation of random variables (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a continuous random variable with mean &amp;lt;math&amp;gt;\mu&amp;lt;/math&amp;gt; and cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X \ge 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\int_{0}^a F(x) dx = \int_{a}^{\infty} [1-F(x)] dx&amp;lt;/math&amp;gt; if and only if &amp;lt;math&amp;gt;a = \mu&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; has finite variance. Show that &amp;lt;math&amp;gt;g(a) = \mathbb{E}((X-a)^2)&amp;lt;/math&amp;gt; achieves the minimum when &amp;lt;math&amp;gt;a = \mu&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Expectation of random variables (II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X, Y&amp;lt;/math&amp;gt; be two independent and identically distributed continuous random variables with cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;. Furthermore, &amp;lt;math&amp;gt;X,Y \ge 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbb{E}[|X-Y|] = 2 \left(\mathbb{E}[X] - \int_{0}^{\infty} (1-F(x))^2 dx\right)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Conditional distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be two random variables. The joint density of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; is given by &amp;lt;math&amp;gt;f(x,y) = c(x^2 - y^2)e^{-x}&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;0\leq x &amp;lt;\infty&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-x\leq y \leq x&amp;lt;/math&amp;gt;. Here, &amp;lt;math&amp;gt;c\in \mathbb{R}_+&amp;lt;/math&amp;gt; is a constant. Find out the conditional distribution of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;, given &amp;lt;math&amp;gt;X = x&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Uniform Distribution (I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;P_i = (X_i,Y_i), 1\leq i\leq n&amp;lt;/math&amp;gt;, be independent, uniformly distributed points in the unit square &amp;lt;math&amp;gt;[0,1]^2&amp;lt;/math&amp;gt;. A point &amp;lt;math&amp;gt;P_i&amp;lt;/math&amp;gt; is called &amp;quot;peripheral&amp;quot; if, for all &amp;lt;math&amp;gt;r = 1,2,\cdots,n&amp;lt;/math&amp;gt;, either &amp;lt;math&amp;gt;X_r \leq X_i&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;Y_r \leq Y_i&amp;lt;/math&amp;gt;, or both. Find out the expected number of peripheral points.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Uniform Distribution (II)&amp;lt;/strong&amp;gt;] Derive the moment generating function of the standard uniform distribution, i.e., uniform distribution on &amp;lt;math&amp;gt;(0,1)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Exponential distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; have an exponential distribution. Show that &amp;lt;math&amp;gt;\textbf{Pr}[X&amp;gt;s+x|X&amp;gt;s] = \textbf{Pr}[X&amp;gt;x]&amp;lt;/math&amp;gt;, for &amp;lt;math&amp;gt;x,s\geq 0&amp;lt;/math&amp;gt;. This is the memoryless property. Show that the exponential distribution is the only continuous distribution with this property.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Normal distribution(I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X,Y\sim N(0,1)&amp;lt;/math&amp;gt; be two independent and identically distributed normal random variables. Let &amp;lt;math&amp;gt;Z = X-Y&amp;lt;/math&amp;gt;. Find the density function of &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;|Z|&amp;lt;/math&amp;gt; respectively.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Normal distribution(II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; have the &amp;lt;math&amp;gt;N(0,1)&amp;lt;/math&amp;gt; distribution and let &amp;lt;math&amp;gt;a&amp;gt;0&amp;lt;/math&amp;gt;. Show that the random variable &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; given by&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{equation*}&lt;br /&gt;
Y = \begin{cases}&lt;br /&gt;
X, &amp;amp; |X|&amp;lt; a \\&lt;br /&gt;
-X, &amp;amp; |X|\geq a&lt;br /&gt;
\end{cases}&lt;br /&gt;
\end{equation*}&amp;lt;/math&amp;gt;&lt;br /&gt;
has the &amp;lt;math&amp;gt;N(0,1)&amp;lt;/math&amp;gt; distribution, and find an expression for &amp;lt;math&amp;gt;\rho(a) = \textbf{Cov}(X,Y)&amp;lt;/math&amp;gt; in terms of the density function &amp;lt;math&amp;gt;\phi&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random process &amp;lt;/strong&amp;gt;]&lt;br /&gt;
Given a real number &amp;lt;math&amp;gt;U&amp;lt;1&amp;lt;/math&amp;gt; as input of the following process, find out the expected returning value.&lt;br /&gt;
{{Theorem|&#039;&#039;Process&#039;&#039;|&lt;br /&gt;
:&#039;&#039;&#039;Input:&#039;&#039;&#039;  real numbers &amp;lt;math&amp;gt;U &amp;lt; 1&amp;lt;/math&amp;gt;;&lt;br /&gt;
----&lt;br /&gt;
:initialize &amp;lt;math&amp;gt;x = 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;count = 0&amp;lt;/math&amp;gt;;&lt;br /&gt;
:while &amp;lt;math&amp;gt; x &amp;lt; U &amp;lt;/math&amp;gt; do&lt;br /&gt;
:* choose &amp;lt;math&amp;gt;y \in (0,1)&amp;lt;/math&amp;gt; uniformly at random;&lt;br /&gt;
:* update &amp;lt;math&amp;gt;x = x + y&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;count = count + 1&amp;lt;/math&amp;gt;;&lt;br /&gt;
:return &amp;lt;math&amp;gt;count&amp;lt;/math&amp;gt;;&lt;br /&gt;
}}&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Random semicircle&amp;lt;/strong&amp;gt;] We sample &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; points within a circle &amp;lt;math&amp;gt;C=\{(x,y) \in \mathbb{R}^2 \mid x^2+y^2 \le 1\}&amp;lt;/math&amp;gt; independently and uniformly at random (i.e., the density function &amp;lt;math&amp;gt;f(x,y) \propto 1_{(x,y) \in C}&amp;lt;/math&amp;gt;). Find out the probability that they all lie within some semicircle of the original circle &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;. (Hint: you may apply the technique of change of variables, see [https://en.wikipedia.org/wiki/Random_variable#Functions_of_random_variables function of random variables] or Chapter 4.7 in [GS])&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Stochastic domination&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X, Y&amp;lt;/math&amp;gt; be continuous random variables. Show that &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; dominates &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; stochastically if and only if &amp;lt;math&amp;gt;\mathbb{E}[f(X)]\geq \mathbb{E}[f(Y)]&amp;lt;/math&amp;gt; for any non-decreasing function &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; for which the expectations exist.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Concentration of measure, 10 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Tossing coins&amp;lt;/strong&amp;gt;] We repeatedly toss a fair coin (with an equal probability of heads and tails). Let the random variable &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of throws required to obtain a total of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; heads. Show that &amp;lt;math&amp;gt;\textbf{Pr}[X &amp;gt; 2n + 2\sqrt{n\log n}]\leq O(1/n)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Chernoff vs Chebyshev&amp;lt;/strong&amp;gt;] We have a standard six-sided die. Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of times a 6 occurs in &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; throws off the die. Compare the best upper bounds on &amp;lt;math&amp;gt;\textbf{Pr}[X\geq n/4]&amp;lt;/math&amp;gt; that you can obtain using Chebyshev&#039;s inequality and Chernoff bounds.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;&amp;lt;strong&amp;gt;-th moment bound]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a random variable with expectation &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; such that moment generating function &amp;lt;math&amp;gt;\mathbf{E}[\exp(t|X|)]&amp;lt;/math&amp;gt; is finite for some &amp;lt;math&amp;gt; t &amp;gt; 0 &amp;lt;/math&amp;gt;. We can use the following two kinds of tail inequalities for &amp;lt;math&amp;gt; X &amp;lt;/math&amp;gt;: &lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&amp;lt;nowiki&amp;gt;     &amp;lt;/nowiki&amp;gt;&#039;&#039;&#039;&#039;&#039;Chernoff Bound&#039;&#039;&#039;&#039;&#039;&lt;br /&gt;
:&amp;lt;math&amp;gt;&lt;br /&gt;
\begin{align}&lt;br /&gt;
\mathbf{Pr}[|X| \geq \delta] \leq \min_{t \geq 0} \frac{\mathbf{E}[e^{t|X|}]}{e^{t\delta}}&lt;br /&gt;
\end{align}&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&#039;&#039;&#039;&#039;&#039;&amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;th-Moment Bound&#039;&#039;&#039;&#039;&#039;&lt;br /&gt;
:&amp;lt;math&amp;gt;&lt;br /&gt;
\begin{align}&lt;br /&gt;
\mathbf{Pr}[|X| \geq \delta] \leq \frac{\mathbf{E}[|X|^k]}{\delta^k}&lt;br /&gt;
\end{align}&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
# Show that for each &amp;lt;math&amp;gt;\delta&amp;lt;/math&amp;gt;, there exists a choice of &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; such that the &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;th-moment bound is no weaker than the Chernoff bound. (Hint: Use the probabilistic method.)&lt;br /&gt;
# Why would we still prefer the Chernoff bound to the (seemingly) stronger &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;-th moment bound?&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Chernoff bound meets graph theory&amp;lt;/strong&amp;gt;]&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; Show that with a probability approaching 1 (as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity), the Erdős–Rényi random graph &amp;lt;math&amp;gt;\textbf{G}(n,1/2)&amp;lt;/math&amp;gt; has the property that the maximum degree is &amp;lt;math&amp;gt;(\frac{n}{2} + O(\sqrt{n\log n}))&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; Show that with a probability approaching 1 (as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity), the Erdős–Rényi random graph &amp;lt;math&amp;gt;\textbf{G}(n,1/2)&amp;lt;/math&amp;gt; has the property that the diameter is exactly 2. The diameter of a graph &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; is the maximum distance between any pair of vertices.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Modes of Convergence, 15 points) (&amp;lt;strong&amp;gt;Bonus problem&amp;lt;/strong&amp;gt;)==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Connection of convergence modes (I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;(X_n)_{n \ge 1}, (Y_n)_{n \ge 1}, X, Y&amp;lt;/math&amp;gt; be random variables and &amp;lt;math&amp;gt;c\in\mathbb{R}&amp;lt;/math&amp;gt; be a real number. &lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; Suppose &amp;lt;math&amp;gt;X_n \overset{D}{\to} X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_n \overset{D}{\to} c&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;X_nY_n \overset{D}{\to} cX&amp;lt;/math&amp;gt;. &lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Construct an example such that &amp;lt;math&amp;gt;X_n \overset{D}{\to} X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_n \overset{D}{\to} Y&amp;lt;/math&amp;gt; but &amp;lt;math&amp;gt;X_nY_n&amp;lt;/math&amp;gt; does not converge to &amp;lt;math&amp;gt;XY&amp;lt;/math&amp;gt; in distribution.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Connection of convergence modes (II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;(X_n)_{n \ge 1}, X&amp;lt;/math&amp;gt; be random variables. Prove that &amp;lt;math&amp;gt;X_n \overset{P}{\to} X&amp;lt;/math&amp;gt; if and only if for every subsequence &amp;lt;math&amp;gt;X_{n(m)}&amp;lt;/math&amp;gt;, there exists a further subsequence &amp;lt;math&amp;gt;Y_k = X_{n(m_k)}&amp;lt;/math&amp;gt; that converges almost surely to &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;. (Hint: you may use the first Borel-Cantelli lemma.)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Extension of Borel-Cantelli Lemma&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;(A_n)_{n \ge 1}&amp;lt;/math&amp;gt; be events. Suppose &amp;lt;math&amp;gt;\sum_{n \ge 1} \mathbf{Pr}(A_n)=+\infty&amp;lt;/math&amp;gt;. Show that &lt;br /&gt;
&amp;lt;math&amp;gt;\mathbf{Pr}(A_n \text{ i.o.}) \ge \limsup_{n \to \infty} \frac{ \left(\sum_{k=1}^n\mathbf{Pr}(A_k)\right)^2 }{\sum_{1\le j,k \le n} \mathbf{Pr}(A_j \cap A_k)}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 4 (LLN and CLT, 15 points + 5 points) ==&lt;br /&gt;
&amp;lt;strong&amp;gt;In this problem, you may apply the results of Laws of Large Numbers (LLN) and the Central Limit Theorem (CLT) to solve the problems.&amp;lt;/strong&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;St. Petersburg paradox&amp;lt;/strong&amp;gt;] Consider the well-known game involving a fair coin. In this game, if it takes &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; tosses to obtain a head, you will win &amp;lt;math&amp;gt;2^k&amp;lt;/math&amp;gt; dollars as the reward. Despite the game&#039;s expected reward being infinite, people tend to offer relatively modest amounts to participate. The following provides a mathematical explanation for this phenomenon.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
For each &amp;lt;math&amp;gt;n \ge 1&amp;lt;/math&amp;gt;, let &amp;lt;math&amp;gt;X_{n,1}, X_{n,2},\ldots, X_{n,k}&amp;lt;/math&amp;gt; be independent random variables. Furthermore, let &amp;lt;math&amp;gt;b_n &amp;gt; 0&amp;lt;/math&amp;gt; be real numbers with &amp;lt;math&amp;gt;b_n \to \infty&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widetilde{X}_{n,k} = X_{n,k} \mathbf{1}_{|X_{n,k}| \le b_n}&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;1 \le k \le n&amp;lt;/math&amp;gt;. If &amp;lt;math&amp;gt;\sum_{k=1}^n \mathbf{Pr}(|X_{n,k}| &amp;gt; b_n) \to 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;b_n^{-2} \sum_{k=1}^n \mathbf{E}[\widetilde{X}_{n,k}^2] \to 0&amp;lt;/math&amp;gt; when &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, then &amp;lt;math&amp;gt;(S_n-a_n)/b_n \overset{P}{\to} 0 &amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;S_n = \sum_{k=1}^n X_{n,k}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;a_n = \sum_{k=1}^n \mathbf{E}[\widetilde{X}_{n,k}]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;S_n&amp;lt;/math&amp;gt; be the total winnings after playing &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds of the game. Prove that &amp;lt;math&amp;gt;\frac{S_n}{n \log_2 n} \overset{P}{\to} 1&amp;lt;/math&amp;gt;. (Therefore, a fair price to play this game &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; times is roughly &amp;lt;math&amp;gt;n \log_2 n&amp;lt;/math&amp;gt; dollars)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; (&amp;lt;strong&amp;gt;Bonus problem, 5 points&amp;lt;/strong&amp;gt;)&lt;br /&gt;
Let &amp;lt;math&amp;gt;S_n&amp;lt;/math&amp;gt; be the total winnings after playing &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds of the game. Prove that &amp;lt;math&amp;gt; \limsup_{n \to \infty} \frac{S_n}{n \log_2 n} = \infty&amp;lt;/math&amp;gt; almost surely. (Hint: You may use Borel-Cantelli lemmas)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Monte Carlo Integration&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; be a continuous function on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;U_1,U_2,\ldots,U_n \sim U(0,1)&amp;lt;/math&amp;gt; be independent random variables. Show that &amp;lt;math&amp;gt;I = \frac{1}{n}\sum_{i=1}^n f(U_i) \to \int_0^1 f(x) \mathrm{d}x&amp;lt;/math&amp;gt; in probability. (Remark: This holds so long as &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; is a measurable function on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;\int_0^1 |f(x)| \mathrm{d}x&amp;lt;/math&amp;gt;)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Square root&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X_1,X_2,\ldots&amp;lt;/math&amp;gt; be i.i.d. random variables with &amp;lt;math&amp;gt;X_i \ge 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{E}[X_1] = 1&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{Var}[X_1] = \sigma^2 \in (0,+\infty)&amp;lt;/math&amp;gt;. Show &amp;lt;math&amp;gt;\sqrt{S_n} - \sqrt{n} \overset{D}{\to} N(0,1)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;S_n = \sum_{i=1}^n X_i&amp;lt;/math&amp;gt;. (Hint: You may use the statement in Problem 3.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2024)/Problem_Set_4&amp;diff=12438</id>
		<title>概率论与数理统计 (Spring 2024)/Problem Set 4</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2024)/Problem_Set_4&amp;diff=12438"/>
		<updated>2024-05-24T13:08:58Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: /* Problem 4 (LLN and CLT, 15 points + 5 points) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
*Bonus problem为附加题（选做）。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 4==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Continuous Random Variables, 30 points)==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Density function&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Determine the value of &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;f(x) = C\exp(-x-e^{-x}), x\in \mathbb{R}&amp;lt;/math&amp;gt; is a probability density function (PDF) for a continuous random variable.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Independence&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be independent and identically distributed continuous random variables with cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; and probability density function (PDF) &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt;. Find out the density functions of &amp;lt;math&amp;gt;V = \max\{X,Y\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;U = \min\{X,Y\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Correlation&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be uniformly distributed on &amp;lt;math&amp;gt;(-1,1)&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_k = \cos(k \pi X)&amp;lt;/math&amp;gt; for &amp;lt;math&amp;gt;k=1,2,\ldots,n&amp;lt;/math&amp;gt;. Are the random variables &amp;lt;math&amp;gt;Y_1, Y_2, \ldots, Y_n&amp;lt;/math&amp;gt; correlated? independent? You should prove your claim rigorously.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Expectation of random variables (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a continuous random variable with mean &amp;lt;math&amp;gt;\mu&amp;lt;/math&amp;gt; and cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X \ge 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\int_{0}^a F(x) dx = \int_{a}^{\infty} [1-F(x)] dx&amp;lt;/math&amp;gt; if and only if &amp;lt;math&amp;gt;a = \mu&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; has finite variance. Show that &amp;lt;math&amp;gt;g(a) = \mathbb{E}((X-a)^2)&amp;lt;/math&amp;gt; achieves the minimum when &amp;lt;math&amp;gt;a = \mu&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Expectation of random variables (II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X, Y&amp;lt;/math&amp;gt; be two independent and identically distributed continuous random variables with cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;. Furthermore, &amp;lt;math&amp;gt;X,Y \ge 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbb{E}[|X-Y|] = 2 \left(\mathbb{E}[X] - \int_{0}^{\infty} (1-F(x))^2 dx\right)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Conditional distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be two random variables. The joint density of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; is given by &amp;lt;math&amp;gt;f(x,y) = c(x^2 - y^2)e^{-x}&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;0\leq x &amp;lt;\infty&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-x\leq y \leq x&amp;lt;/math&amp;gt;. Here, &amp;lt;math&amp;gt;c\in \mathbb{R}_+&amp;lt;/math&amp;gt; is a constant. Find out the conditional distribution of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;, given &amp;lt;math&amp;gt;X = x&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Uniform Distribution (I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;P_i = (X_i,Y_i), 1\leq i\leq n&amp;lt;/math&amp;gt;, be independent, uniformly distributed points in the unit square &amp;lt;math&amp;gt;[0,1]^2&amp;lt;/math&amp;gt;. A point &amp;lt;math&amp;gt;P_i&amp;lt;/math&amp;gt; is called &amp;quot;peripheral&amp;quot; if, for all &amp;lt;math&amp;gt;r = 1,2,\cdots,n&amp;lt;/math&amp;gt;, either &amp;lt;math&amp;gt;X_r \leq X_i&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;Y_r \leq Y_i&amp;lt;/math&amp;gt;, or both. Find out the expected number of peripheral points.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Uniform Distribution (II)&amp;lt;/strong&amp;gt;] Derive the moment generating function of the standard uniform distribution, i.e., uniform distribution on &amp;lt;math&amp;gt;(0,1)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Exponential distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; have an exponential distribution. Show that &amp;lt;math&amp;gt;\textbf{Pr}[X&amp;gt;s+x|X&amp;gt;s] = \textbf{Pr}[X&amp;gt;x]&amp;lt;/math&amp;gt;, for &amp;lt;math&amp;gt;x,s\geq 0&amp;lt;/math&amp;gt;. This is the memoryless property. Show that the exponential distribution is the only continuous distribution with this property.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Normal distribution(I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X,Y\sim N(0,1)&amp;lt;/math&amp;gt; be two independent and identically distributed normal random variables. Let &amp;lt;math&amp;gt;Z = X-Y&amp;lt;/math&amp;gt;. Find the density function of &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;|Z|&amp;lt;/math&amp;gt; respectively.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Normal distribution(II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; have the &amp;lt;math&amp;gt;N(0,1)&amp;lt;/math&amp;gt; distribution and let &amp;lt;math&amp;gt;a&amp;gt;0&amp;lt;/math&amp;gt;. Show that the random variable &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; given by&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{equation*}&lt;br /&gt;
Y = \begin{cases}&lt;br /&gt;
X, &amp;amp; |X|&amp;lt; a \\&lt;br /&gt;
-X, &amp;amp; |X|\geq a&lt;br /&gt;
\end{cases}&lt;br /&gt;
\end{equation*}&amp;lt;/math&amp;gt;&lt;br /&gt;
has the &amp;lt;math&amp;gt;N(0,1)&amp;lt;/math&amp;gt; distribution, and find an expression for &amp;lt;math&amp;gt;\rho(a) = \textbf{Cov}(X,Y)&amp;lt;/math&amp;gt; in terms of the density function &amp;lt;math&amp;gt;\phi&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random process (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Given a real number &amp;lt;math&amp;gt;U&amp;lt;1&amp;lt;/math&amp;gt; as input of the following process, find out the expected returning value.&lt;br /&gt;
{{Theorem|&#039;&#039;Process 1&#039;&#039;|&lt;br /&gt;
:&#039;&#039;&#039;Input:&#039;&#039;&#039;  real numbers &amp;lt;math&amp;gt;U &amp;lt; 1&amp;lt;/math&amp;gt;;&lt;br /&gt;
----&lt;br /&gt;
:initialize &amp;lt;math&amp;gt;x = 1&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;count = 0&amp;lt;/math&amp;gt;;&lt;br /&gt;
:while &amp;lt;math&amp;gt; x &amp;gt; U &amp;lt;/math&amp;gt; do&lt;br /&gt;
:* choose &amp;lt;math&amp;gt;y \in (0,1)&amp;lt;/math&amp;gt; uniformly at random;&lt;br /&gt;
:* update &amp;lt;math&amp;gt;x = x * y&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;count = count + 1&amp;lt;/math&amp;gt;;&lt;br /&gt;
:return &amp;lt;math&amp;gt;count&amp;lt;/math&amp;gt;;&lt;br /&gt;
}}&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random process (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Given a real number &amp;lt;math&amp;gt;U&amp;lt;1&amp;lt;/math&amp;gt; as input of the following process, find out the expected returning value.&lt;br /&gt;
{{Theorem|&#039;&#039;Process 2&#039;&#039;|&lt;br /&gt;
:&#039;&#039;&#039;Input:&#039;&#039;&#039;  real numbers &amp;lt;math&amp;gt;U &amp;lt; 1&amp;lt;/math&amp;gt;;&lt;br /&gt;
----&lt;br /&gt;
:initialize &amp;lt;math&amp;gt;x = 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;count = 0&amp;lt;/math&amp;gt;;&lt;br /&gt;
:while &amp;lt;math&amp;gt; x &amp;lt; U &amp;lt;/math&amp;gt; do&lt;br /&gt;
:* choose &amp;lt;math&amp;gt;y \in (0,1)&amp;lt;/math&amp;gt; uniformly at random;&lt;br /&gt;
:* update &amp;lt;math&amp;gt;x = x + y&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;count = count + 1&amp;lt;/math&amp;gt;;&lt;br /&gt;
:return &amp;lt;math&amp;gt;count&amp;lt;/math&amp;gt;;&lt;br /&gt;
}}&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Random semicircle&amp;lt;/strong&amp;gt;] We sample &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; points within a circle &amp;lt;math&amp;gt;C=\{(x,y) \in \mathbb{R}^2 \mid x^2+y^2 \le 1\}&amp;lt;/math&amp;gt; independently and uniformly at random (i.e., the density function &amp;lt;math&amp;gt;f(x,y) \propto 1_{(x,y) \in C}&amp;lt;/math&amp;gt;). Find out the probability that they all lie within some semicircle of the original circle &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;. (Hint: you may apply the technique of change of variables, see [https://en.wikipedia.org/wiki/Random_variable#Functions_of_random_variables function of random variables] or Chapter 4.7 in [GS])&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Stochastic domination&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X, Y&amp;lt;/math&amp;gt; be continuous random variables. Show that &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; dominates &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; stochastically if and only if &amp;lt;math&amp;gt;\mathbb{E}[f(X)]\geq \mathbb{E}[f(Y)]&amp;lt;/math&amp;gt; for any non-decreasing function &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; for which the expectations exist.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Concentration of measure, 10 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Tossing coins&amp;lt;/strong&amp;gt;] We repeatedly toss a fair coin (with an equal probability of heads and tails). Let the random variable &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of throws required to obtain a total of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; heads. Show that &amp;lt;math&amp;gt;\textbf{Pr}[X &amp;gt; 2n + 2\sqrt{n\log n}]\leq O(1/n)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Chernoff vs Chebyshev&amp;lt;/strong&amp;gt;] We have a standard six-sided die. Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of times a 6 occurs in &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; throws off the die. Compare the best upper bounds on &amp;lt;math&amp;gt;\textbf{Pr}[X\geq n/4]&amp;lt;/math&amp;gt; that you can obtain using Chebyshev&#039;s inequality and Chernoff bounds.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;&amp;lt;strong&amp;gt;-th moment bound]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a random variable with expectation &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; such that moment generating function &amp;lt;math&amp;gt;\mathbf{E}[\exp(t|X|)]&amp;lt;/math&amp;gt; is finite for some &amp;lt;math&amp;gt; t &amp;gt; 0 &amp;lt;/math&amp;gt;. We can use the following two kinds of tail inequalities for &amp;lt;math&amp;gt; X &amp;lt;/math&amp;gt;: &lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&amp;lt;nowiki&amp;gt;     &amp;lt;/nowiki&amp;gt;&#039;&#039;&#039;&#039;&#039;Chernoff Bound&#039;&#039;&#039;&#039;&#039;&lt;br /&gt;
:&amp;lt;math&amp;gt;&lt;br /&gt;
\begin{align}&lt;br /&gt;
\mathbf{Pr}[|X| \geq \delta] \leq \min_{t \geq 0} \frac{\mathbf{E}[e^{t|X|}]}{e^{t\delta}}&lt;br /&gt;
\end{align}&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&#039;&#039;&#039;&#039;&#039;&amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;th-Moment Bound&#039;&#039;&#039;&#039;&#039;&lt;br /&gt;
:&amp;lt;math&amp;gt;&lt;br /&gt;
\begin{align}&lt;br /&gt;
\mathbf{Pr}[|X| \geq \delta] \leq \frac{\mathbf{E}[|X|^k]}{\delta^k}&lt;br /&gt;
\end{align}&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
# Show that for each &amp;lt;math&amp;gt;\delta&amp;lt;/math&amp;gt;, there exists a choice of &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; such that the &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;th-moment bound is no weaker than the Chernoff bound. (Hint: Use the probabilistic method.)&lt;br /&gt;
# Why would we still prefer the Chernoff bound to the (seemingly) stronger &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;-th moment bound?&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Chernoff bound meets graph theory&amp;lt;/strong&amp;gt;]&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; Show that with a probability approaching 1 (as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity), the Erdős–Rényi random graph &amp;lt;math&amp;gt;\textbf{G}(n,1/2)&amp;lt;/math&amp;gt; has the property that the maximum degree is &amp;lt;math&amp;gt;(\frac{n}{2} + O(\sqrt{n\log n}))&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; Show that with a probability approaching 1 (as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity), the Erdős–Rényi random graph &amp;lt;math&amp;gt;\textbf{G}(n,1/2)&amp;lt;/math&amp;gt; has the property that the diameter is exactly 2. The diameter of a graph &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; is the maximum distance between any pair of vertices.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Modes of Convergence, 15 points) (&amp;lt;strong&amp;gt;Bonus problem&amp;lt;/strong&amp;gt;)==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Connection of convergence modes (I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;(X_n)_{n \ge 1}, (Y_n)_{n \ge 1}, X, Y&amp;lt;/math&amp;gt; be random variables and &amp;lt;math&amp;gt;c\in\mathbb{R}&amp;lt;/math&amp;gt; be a real number. &lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; Suppose &amp;lt;math&amp;gt;X_n \overset{D}{\to} X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_n \overset{D}{\to} c&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;X_nY_n \overset{D}{\to} cX&amp;lt;/math&amp;gt;. &lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Construct an example such that &amp;lt;math&amp;gt;X_n \overset{D}{\to} X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_n \overset{D}{\to} Y&amp;lt;/math&amp;gt; but &amp;lt;math&amp;gt;X_nY_n&amp;lt;/math&amp;gt; does not converge to &amp;lt;math&amp;gt;XY&amp;lt;/math&amp;gt; in distribution.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Connection of convergence modes (II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;(X_n)_{n \ge 1}, X&amp;lt;/math&amp;gt; be random variables. Prove that &amp;lt;math&amp;gt;X_n \overset{P}{\to} X&amp;lt;/math&amp;gt; if and only if for every subsequence &amp;lt;math&amp;gt;X_{n(m)}&amp;lt;/math&amp;gt;, there exists a further subsequence &amp;lt;math&amp;gt;Y_k = X_{n(m_k)}&amp;lt;/math&amp;gt; that converges almost surely to &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;. (Hint: you may use the first Borel-Cantelli lemma.)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Extension of Borel-Cantelli Lemma&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;(A_n)_{n \ge 1}&amp;lt;/math&amp;gt; be events. Suppose &amp;lt;math&amp;gt;\sum_{n \ge 1} \mathbf{Pr}(A_n)=+\infty&amp;lt;/math&amp;gt;. Show that &lt;br /&gt;
&amp;lt;math&amp;gt;\mathbf{Pr}(A_n \text{ i.o.}) \ge \limsup_{n \to \infty} \frac{ \left(\sum_{k=1}^n\mathbf{Pr}(A_k)\right)^2 }{\sum_{1\le j,k \le n} \mathbf{Pr}(A_j \cap A_k)}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 4 (LLN and CLT, 15 points + 5 points) ==&lt;br /&gt;
&amp;lt;strong&amp;gt;In this problem, you may apply the results of Laws of Large Numbers (LLN) and the Central Limit Theorem (CLT) to solve the problems.&amp;lt;/strong&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;St. Petersburg paradox&amp;lt;/strong&amp;gt;] Consider the well-known game involving a fair coin. In this game, if it takes &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; tosses to obtain a head, you will win &amp;lt;math&amp;gt;2^k&amp;lt;/math&amp;gt; dollars as the reward. Despite the game&#039;s expected reward being infinite, people tend to offer relatively modest amounts to participate. The following provides a mathematical explanation for this phenomenon.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
For each &amp;lt;math&amp;gt;n \ge 1&amp;lt;/math&amp;gt;, let &amp;lt;math&amp;gt;X_{n,1}, X_{n,2},\ldots, X_{n,k}&amp;lt;/math&amp;gt; be independent random variables. Furthermore, let &amp;lt;math&amp;gt;b_n &amp;gt; 0&amp;lt;/math&amp;gt; be real numbers with &amp;lt;math&amp;gt;b_n \to \infty&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widetilde{X}_{n,k} = X_{n,k} \mathbf{1}_{|X_{n,k}| \le b_n}&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;1 \le k \le n&amp;lt;/math&amp;gt;. If &amp;lt;math&amp;gt;\sum_{k=1}^n \mathbf{Pr}(|X_{n,k}| &amp;gt; b_n) \to 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;b_n^{-2} \sum_{k=1}^n \mathbf{E}[\widetilde{X}_{n,k}^2] \to 0&amp;lt;/math&amp;gt; when &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, then &amp;lt;math&amp;gt;(S_n-a_n)/b_n \overset{P}{\to} 0 &amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;S_n = \sum_{k=1}^n X_{n,k}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;a_n = \sum_{k=1}^n \mathbf{E}[\widetilde{X}_{n,k}]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;S_n&amp;lt;/math&amp;gt; be the total winnings after playing &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds of the game. Prove that &amp;lt;math&amp;gt;\frac{S_n}{n \log_2 n} \overset{P}{\to} 1&amp;lt;/math&amp;gt;. (Therefore, a fair price to play this game &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; times is roughly &amp;lt;math&amp;gt;n \log_2 n&amp;lt;/math&amp;gt; dollars)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; (&amp;lt;strong&amp;gt;Bonus problem, 5 points&amp;lt;/strong&amp;gt;)&lt;br /&gt;
Let &amp;lt;math&amp;gt;S_n&amp;lt;/math&amp;gt; be the total winnings after playing &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds of the game. Prove that &amp;lt;math&amp;gt; \limsup_{n \to \infty} \frac{S_n}{n \log_2 n} = \infty&amp;lt;/math&amp;gt; almost surely. (Hint: You may use Borel-Cantelli lemmas)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Monte Carlo Integration&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; be a continuous function on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;U_1,U_2,\ldots,U_n \sim U(0,1)&amp;lt;/math&amp;gt; be independent random variables. Show that &amp;lt;math&amp;gt;I = \frac{1}{n}\sum_{i=1}^n f(U_i) \to \int_0^1 f(x) \mathrm{d}x&amp;lt;/math&amp;gt; in probability. (Remark: This holds so long as &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; is a measurable function on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;\int_0^1 |f(x)| \mathrm{d}x&amp;lt;/math&amp;gt;)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Square root&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X_1,X_2,\ldots&amp;lt;/math&amp;gt; be i.i.d. random variables with &amp;lt;math&amp;gt;X_i \ge 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{E}[X_1] = 1&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{Var}[X_1] = \sigma^2 \in (0,+\infty)&amp;lt;/math&amp;gt;. Show &amp;lt;math&amp;gt;\sqrt{S_n} - \sqrt{n} \overset{D}{\to} N(0,1)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;S_n = \sum_{i=1}^n X_i&amp;lt;/math&amp;gt;. (Hint: You may use the conclusion in )&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2024)/Problem_Set_4&amp;diff=12437</id>
		<title>概率论与数理统计 (Spring 2024)/Problem Set 4</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2024)/Problem_Set_4&amp;diff=12437"/>
		<updated>2024-05-24T13:08:23Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: /* Problem 4 (LLN and CLT, 15 points + 5 points) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
*Bonus problem为附加题（选做）。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 4==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Continuous Random Variables, 30 points)==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Density function&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Determine the value of &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;f(x) = C\exp(-x-e^{-x}), x\in \mathbb{R}&amp;lt;/math&amp;gt; is a probability density function (PDF) for a continuous random variable.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Independence&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be independent and identically distributed continuous random variables with cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; and probability density function (PDF) &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt;. Find out the density functions of &amp;lt;math&amp;gt;V = \max\{X,Y\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;U = \min\{X,Y\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Correlation&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be uniformly distributed on &amp;lt;math&amp;gt;(-1,1)&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_k = \cos(k \pi X)&amp;lt;/math&amp;gt; for &amp;lt;math&amp;gt;k=1,2,\ldots,n&amp;lt;/math&amp;gt;. Are the random variables &amp;lt;math&amp;gt;Y_1, Y_2, \ldots, Y_n&amp;lt;/math&amp;gt; correlated? independent? You should prove your claim rigorously.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Expectation of random variables (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a continuous random variable with mean &amp;lt;math&amp;gt;\mu&amp;lt;/math&amp;gt; and cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X \ge 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\int_{0}^a F(x) dx = \int_{a}^{\infty} [1-F(x)] dx&amp;lt;/math&amp;gt; if and only if &amp;lt;math&amp;gt;a = \mu&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; has finite variance. Show that &amp;lt;math&amp;gt;g(a) = \mathbb{E}((X-a)^2)&amp;lt;/math&amp;gt; achieves the minimum when &amp;lt;math&amp;gt;a = \mu&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Expectation of random variables (II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X, Y&amp;lt;/math&amp;gt; be two independent and identically distributed continuous random variables with cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;. Furthermore, &amp;lt;math&amp;gt;X,Y \ge 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbb{E}[|X-Y|] = 2 \left(\mathbb{E}[X] - \int_{0}^{\infty} (1-F(x))^2 dx\right)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Conditional distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be two random variables. The joint density of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; is given by &amp;lt;math&amp;gt;f(x,y) = c(x^2 - y^2)e^{-x}&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;0\leq x &amp;lt;\infty&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-x\leq y \leq x&amp;lt;/math&amp;gt;. Here, &amp;lt;math&amp;gt;c\in \mathbb{R}_+&amp;lt;/math&amp;gt; is a constant. Find out the conditional distribution of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;, given &amp;lt;math&amp;gt;X = x&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Uniform Distribution (I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;P_i = (X_i,Y_i), 1\leq i\leq n&amp;lt;/math&amp;gt;, be independent, uniformly distributed points in the unit square &amp;lt;math&amp;gt;[0,1]^2&amp;lt;/math&amp;gt;. A point &amp;lt;math&amp;gt;P_i&amp;lt;/math&amp;gt; is called &amp;quot;peripheral&amp;quot; if, for all &amp;lt;math&amp;gt;r = 1,2,\cdots,n&amp;lt;/math&amp;gt;, either &amp;lt;math&amp;gt;X_r \leq X_i&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;Y_r \leq Y_i&amp;lt;/math&amp;gt;, or both. Find out the expected number of peripheral points.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Uniform Distribution (II)&amp;lt;/strong&amp;gt;] Derive the moment generating function of the standard uniform distribution, i.e., uniform distribution on &amp;lt;math&amp;gt;(0,1)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Exponential distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; have an exponential distribution. Show that &amp;lt;math&amp;gt;\textbf{Pr}[X&amp;gt;s+x|X&amp;gt;s] = \textbf{Pr}[X&amp;gt;x]&amp;lt;/math&amp;gt;, for &amp;lt;math&amp;gt;x,s\geq 0&amp;lt;/math&amp;gt;. This is the memoryless property. Show that the exponential distribution is the only continuous distribution with this property.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Normal distribution(I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X,Y\sim N(0,1)&amp;lt;/math&amp;gt; be two independent and identically distributed normal random variables. Let &amp;lt;math&amp;gt;Z = X-Y&amp;lt;/math&amp;gt;. Find the density function of &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;|Z|&amp;lt;/math&amp;gt; respectively.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Normal distribution(II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; have the &amp;lt;math&amp;gt;N(0,1)&amp;lt;/math&amp;gt; distribution and let &amp;lt;math&amp;gt;a&amp;gt;0&amp;lt;/math&amp;gt;. Show that the random variable &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; given by&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{equation*}&lt;br /&gt;
Y = \begin{cases}&lt;br /&gt;
X, &amp;amp; |X|&amp;lt; a \\&lt;br /&gt;
-X, &amp;amp; |X|\geq a&lt;br /&gt;
\end{cases}&lt;br /&gt;
\end{equation*}&amp;lt;/math&amp;gt;&lt;br /&gt;
has the &amp;lt;math&amp;gt;N(0,1)&amp;lt;/math&amp;gt; distribution, and find an expression for &amp;lt;math&amp;gt;\rho(a) = \textbf{Cov}(X,Y)&amp;lt;/math&amp;gt; in terms of the density function &amp;lt;math&amp;gt;\phi&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random process (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Given a real number &amp;lt;math&amp;gt;U&amp;lt;1&amp;lt;/math&amp;gt; as input of the following process, find out the expected returning value.&lt;br /&gt;
{{Theorem|&#039;&#039;Process 1&#039;&#039;|&lt;br /&gt;
:&#039;&#039;&#039;Input:&#039;&#039;&#039;  real numbers &amp;lt;math&amp;gt;U &amp;lt; 1&amp;lt;/math&amp;gt;;&lt;br /&gt;
----&lt;br /&gt;
:initialize &amp;lt;math&amp;gt;x = 1&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;count = 0&amp;lt;/math&amp;gt;;&lt;br /&gt;
:while &amp;lt;math&amp;gt; x &amp;gt; U &amp;lt;/math&amp;gt; do&lt;br /&gt;
:* choose &amp;lt;math&amp;gt;y \in (0,1)&amp;lt;/math&amp;gt; uniformly at random;&lt;br /&gt;
:* update &amp;lt;math&amp;gt;x = x * y&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;count = count + 1&amp;lt;/math&amp;gt;;&lt;br /&gt;
:return &amp;lt;math&amp;gt;count&amp;lt;/math&amp;gt;;&lt;br /&gt;
}}&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random process (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Given a real number &amp;lt;math&amp;gt;U&amp;lt;1&amp;lt;/math&amp;gt; as input of the following process, find out the expected returning value.&lt;br /&gt;
{{Theorem|&#039;&#039;Process 2&#039;&#039;|&lt;br /&gt;
:&#039;&#039;&#039;Input:&#039;&#039;&#039;  real numbers &amp;lt;math&amp;gt;U &amp;lt; 1&amp;lt;/math&amp;gt;;&lt;br /&gt;
----&lt;br /&gt;
:initialize &amp;lt;math&amp;gt;x = 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;count = 0&amp;lt;/math&amp;gt;;&lt;br /&gt;
:while &amp;lt;math&amp;gt; x &amp;lt; U &amp;lt;/math&amp;gt; do&lt;br /&gt;
:* choose &amp;lt;math&amp;gt;y \in (0,1)&amp;lt;/math&amp;gt; uniformly at random;&lt;br /&gt;
:* update &amp;lt;math&amp;gt;x = x + y&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;count = count + 1&amp;lt;/math&amp;gt;;&lt;br /&gt;
:return &amp;lt;math&amp;gt;count&amp;lt;/math&amp;gt;;&lt;br /&gt;
}}&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Random semicircle&amp;lt;/strong&amp;gt;] We sample &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; points within a circle &amp;lt;math&amp;gt;C=\{(x,y) \in \mathbb{R}^2 \mid x^2+y^2 \le 1\}&amp;lt;/math&amp;gt; independently and uniformly at random (i.e., the density function &amp;lt;math&amp;gt;f(x,y) \propto 1_{(x,y) \in C}&amp;lt;/math&amp;gt;). Find out the probability that they all lie within some semicircle of the original circle &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;. (Hint: you may apply the technique of change of variables, see [https://en.wikipedia.org/wiki/Random_variable#Functions_of_random_variables function of random variables] or Chapter 4.7 in [GS])&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Stochastic domination&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X, Y&amp;lt;/math&amp;gt; be continuous random variables. Show that &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; dominates &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; stochastically if and only if &amp;lt;math&amp;gt;\mathbb{E}[f(X)]\geq \mathbb{E}[f(Y)]&amp;lt;/math&amp;gt; for any non-decreasing function &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; for which the expectations exist.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Concentration of measure, 10 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Tossing coins&amp;lt;/strong&amp;gt;] We repeatedly toss a fair coin (with an equal probability of heads and tails). Let the random variable &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of throws required to obtain a total of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; heads. Show that &amp;lt;math&amp;gt;\textbf{Pr}[X &amp;gt; 2n + 2\sqrt{n\log n}]\leq O(1/n)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Chernoff vs Chebyshev&amp;lt;/strong&amp;gt;] We have a standard six-sided die. Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of times a 6 occurs in &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; throws off the die. Compare the best upper bounds on &amp;lt;math&amp;gt;\textbf{Pr}[X\geq n/4]&amp;lt;/math&amp;gt; that you can obtain using Chebyshev&#039;s inequality and Chernoff bounds.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;&amp;lt;strong&amp;gt;-th moment bound]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a random variable with expectation &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; such that moment generating function &amp;lt;math&amp;gt;\mathbf{E}[\exp(t|X|)]&amp;lt;/math&amp;gt; is finite for some &amp;lt;math&amp;gt; t &amp;gt; 0 &amp;lt;/math&amp;gt;. We can use the following two kinds of tail inequalities for &amp;lt;math&amp;gt; X &amp;lt;/math&amp;gt;: &lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&amp;lt;nowiki&amp;gt;     &amp;lt;/nowiki&amp;gt;&#039;&#039;&#039;&#039;&#039;Chernoff Bound&#039;&#039;&#039;&#039;&#039;&lt;br /&gt;
:&amp;lt;math&amp;gt;&lt;br /&gt;
\begin{align}&lt;br /&gt;
\mathbf{Pr}[|X| \geq \delta] \leq \min_{t \geq 0} \frac{\mathbf{E}[e^{t|X|}]}{e^{t\delta}}&lt;br /&gt;
\end{align}&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&#039;&#039;&#039;&#039;&#039;&amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;th-Moment Bound&#039;&#039;&#039;&#039;&#039;&lt;br /&gt;
:&amp;lt;math&amp;gt;&lt;br /&gt;
\begin{align}&lt;br /&gt;
\mathbf{Pr}[|X| \geq \delta] \leq \frac{\mathbf{E}[|X|^k]}{\delta^k}&lt;br /&gt;
\end{align}&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
# Show that for each &amp;lt;math&amp;gt;\delta&amp;lt;/math&amp;gt;, there exists a choice of &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; such that the &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;th-moment bound is no weaker than the Chernoff bound. (Hint: Use the probabilistic method.)&lt;br /&gt;
# Why would we still prefer the Chernoff bound to the (seemingly) stronger &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;-th moment bound?&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Chernoff bound meets graph theory&amp;lt;/strong&amp;gt;]&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; Show that with a probability approaching 1 (as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity), the Erdős–Rényi random graph &amp;lt;math&amp;gt;\textbf{G}(n,1/2)&amp;lt;/math&amp;gt; has the property that the maximum degree is &amp;lt;math&amp;gt;(\frac{n}{2} + O(\sqrt{n\log n}))&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; Show that with a probability approaching 1 (as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity), the Erdős–Rényi random graph &amp;lt;math&amp;gt;\textbf{G}(n,1/2)&amp;lt;/math&amp;gt; has the property that the diameter is exactly 2. The diameter of a graph &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; is the maximum distance between any pair of vertices.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Modes of Convergence, 15 points) (&amp;lt;strong&amp;gt;Bonus problem&amp;lt;/strong&amp;gt;)==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Connection of convergence modes (I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;(X_n)_{n \ge 1}, (Y_n)_{n \ge 1}, X, Y&amp;lt;/math&amp;gt; be random variables and &amp;lt;math&amp;gt;c\in\mathbb{R}&amp;lt;/math&amp;gt; be a real number. &lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; Suppose &amp;lt;math&amp;gt;X_n \overset{D}{\to} X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_n \overset{D}{\to} c&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;X_nY_n \overset{D}{\to} cX&amp;lt;/math&amp;gt;. &lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Construct an example such that &amp;lt;math&amp;gt;X_n \overset{D}{\to} X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_n \overset{D}{\to} Y&amp;lt;/math&amp;gt; but &amp;lt;math&amp;gt;X_nY_n&amp;lt;/math&amp;gt; does not converge to &amp;lt;math&amp;gt;XY&amp;lt;/math&amp;gt; in distribution.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Connection of convergence modes (II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;(X_n)_{n \ge 1}, X&amp;lt;/math&amp;gt; be random variables. Prove that &amp;lt;math&amp;gt;X_n \overset{P}{\to} X&amp;lt;/math&amp;gt; if and only if for every subsequence &amp;lt;math&amp;gt;X_{n(m)}&amp;lt;/math&amp;gt;, there exists a further subsequence &amp;lt;math&amp;gt;Y_k = X_{n(m_k)}&amp;lt;/math&amp;gt; that converges almost surely to &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;. (Hint: you may use the first Borel-Cantelli lemma.)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Extension of Borel-Cantelli Lemma&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;(A_n)_{n \ge 1}&amp;lt;/math&amp;gt; be events. Suppose &amp;lt;math&amp;gt;\sum_{n \ge 1} \mathbf{Pr}(A_n)=+\infty&amp;lt;/math&amp;gt;. Show that &lt;br /&gt;
&amp;lt;math&amp;gt;\mathbf{Pr}(A_n \text{ i.o.}) \ge \limsup_{n \to \infty} \frac{ \left(\sum_{k=1}^n\mathbf{Pr}(A_k)\right)^2 }{\sum_{1\le j,k \le n} \mathbf{Pr}(A_j \cap A_k)}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 4 (LLN and CLT, 15 points + 5 points) ==&lt;br /&gt;
&amp;lt;strong&amp;gt;In this problem, you may apply the results of Laws of Large Numbers (LLN) and the Central Limit Theorem (CLT) to solve the problems.&amp;lt;/strong&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;St. Petersburg paradox&amp;lt;/strong&amp;gt;] Consider the well-known game involving a fair coin. In this game, if it takes &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; tosses to obtain a head, you will win &amp;lt;math&amp;gt;2^k&amp;lt;/math&amp;gt; dollars as the reward. Despite the game&#039;s expected reward being infinite, people tend to offer relatively modest amounts to participate. The following provides a mathematical explanation for this phenomenon.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
For each &amp;lt;math&amp;gt;n \ge 1&amp;lt;/math&amp;gt;, let &amp;lt;math&amp;gt;X_{n,1}, X_{n,2},\ldots, X_{n,k}&amp;lt;/math&amp;gt; be independent random variables. Furthermore, let &amp;lt;math&amp;gt;b_n &amp;gt; 0&amp;lt;/math&amp;gt; be real numbers with &amp;lt;math&amp;gt;b_n \to \infty&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widetilde{X}_{n,k} = X_{n,k} \mathbf{1}_{|X_{n,k}| \le b_n}&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;1 \le k \le n&amp;lt;/math&amp;gt;. If &amp;lt;math&amp;gt;\sum_{k=1}^n \mathbf{Pr}(|X_{n,k}| &amp;gt; b_n) \to 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;b_n^{-2} \sum_{k=1}^n \mathbf{E}[\widetilde{X}_{n,k}^2] \to 0&amp;lt;/math&amp;gt; when &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, then &amp;lt;math&amp;gt;(S_n-a_n)/b_n \overset{P}{\to} 0 &amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;S_n = \sum_{k=1}^n X_{n,k}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;a_n = \sum_{k=1}^n \mathbf{E}[\widetilde{X}_{n,k}]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;S_n&amp;lt;/math&amp;gt; be the total winnings after playing &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds of the game. Prove that &amp;lt;math&amp;gt;\frac{S_n}{n \log_2 n} \overset{P}{\to} 1&amp;lt;/math&amp;gt;. (Therefore, a fair price to play this game &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; times is roughly &amp;lt;math&amp;gt;n \log_2 n&amp;lt;/math&amp;gt; dollars)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; (&amp;lt;strong&amp;gt;Bonus problem, 5 points&amp;lt;/strong&amp;gt;)&lt;br /&gt;
Let &amp;lt;math&amp;gt;S_n&amp;lt;/math&amp;gt; be the total winnings after playing &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds of the game. Prove that &amp;lt;math&amp;gt; \limsup_{n \to \infty} \frac{S_n}{n \log_2 n} = \infty&amp;lt;/math&amp;gt; almost surely. (Hint: You may use Borel-Cantelli lemmas)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Monte Carlo Integration&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; be a continuous function on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;U_1,U_2,\ldots,U_n \sim U(0,1)&amp;lt;/math&amp;gt; be independent random variables. Show that &amp;lt;math&amp;gt;I = \frac{1}{n}\sum_{i=1}^n f(U_i) \to \int_0^1 f(x) \mathrm{d}x&amp;lt;/math&amp;gt; in probability. (Remark: This holds so long as &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; is a measurable function on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;\int_0^1 |f(x)| \mathrm{d}x&amp;lt;/math&amp;gt;)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Square root&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X_1,X_2,\ldots&amp;lt;/math&amp;gt; be i.i.d. random variables with &amp;lt;math&amp;gt;X_i \ge 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{E}[X_1] = 1&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{Var}[X_1] = \sigma^2 \in (0,+\infty)&amp;lt;/math&amp;gt;. Show &amp;lt;math&amp;gt;\sqrt{S_n} - \sqrt{n} \overset{D}{\to} N(0,1)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, where $S_n = \sum_{i=1}^n X_i$.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2024)/Problem_Set_4&amp;diff=12435</id>
		<title>概率论与数理统计 (Spring 2024)/Problem Set 4</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2024)/Problem_Set_4&amp;diff=12435"/>
		<updated>2024-05-24T13:01:28Z</updated>

		<summary type="html">&lt;p&gt;Zhangxy: /* Problem 3 (LLN and CLT, 15 points + 5 points) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
*Bonus problem为附加题（选做）。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 4==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Continuous Random Variables, 30 points)==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Density function&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Determine the value of &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;f(x) = C\exp(-x-e^{-x}), x\in \mathbb{R}&amp;lt;/math&amp;gt; is a probability density function (PDF) for a continuous random variable.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Independence&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be independent and identically distributed continuous random variables with cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; and probability density function (PDF) &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt;. Find out the density functions of &amp;lt;math&amp;gt;V = \max\{X,Y\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;U = \min\{X,Y\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Correlation&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be uniformly distributed on &amp;lt;math&amp;gt;(-1,1)&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_k = \cos(k \pi X)&amp;lt;/math&amp;gt; for &amp;lt;math&amp;gt;k=1,2,\ldots,n&amp;lt;/math&amp;gt;. Are the random variables &amp;lt;math&amp;gt;Y_1, Y_2, \ldots, Y_n&amp;lt;/math&amp;gt; correlated? independent? You should prove your claim rigorously.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Expectation of random variables (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a continuous random variable with mean &amp;lt;math&amp;gt;\mu&amp;lt;/math&amp;gt; and cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X \ge 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\int_{0}^a F(x) dx = \int_{a}^{\infty} [1-F(x)] dx&amp;lt;/math&amp;gt; if and only if &amp;lt;math&amp;gt;a = \mu&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; has finite variance. Show that &amp;lt;math&amp;gt;g(a) = \mathbb{E}((X-a)^2)&amp;lt;/math&amp;gt; achieves the minimum when &amp;lt;math&amp;gt;a = \mu&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Expectation of random variables (II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X, Y&amp;lt;/math&amp;gt; be two independent and identically distributed continuous random variables with cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;. Furthermore, &amp;lt;math&amp;gt;X,Y \ge 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbb{E}[|X-Y|] = 2 \left(\mathbb{E}[X] - \int_{0}^{\infty} (1-F(x))^2 dx\right)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Conditional distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be two random variables. The joint density of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; is given by &amp;lt;math&amp;gt;f(x,y) = c(x^2 - y^2)e^{-x}&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;0\leq x &amp;lt;\infty&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-x\leq y \leq x&amp;lt;/math&amp;gt;. Here, &amp;lt;math&amp;gt;c\in \mathbb{R}_+&amp;lt;/math&amp;gt; is a constant. Find out the conditional distribution of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;, given &amp;lt;math&amp;gt;X = x&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Uniform Distribution (I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;P_i = (X_i,Y_i), 1\leq i\leq n&amp;lt;/math&amp;gt;, be independent, uniformly distributed points in the unit square &amp;lt;math&amp;gt;[0,1]^2&amp;lt;/math&amp;gt;. A point &amp;lt;math&amp;gt;P_i&amp;lt;/math&amp;gt; is called &amp;quot;peripheral&amp;quot; if, for all &amp;lt;math&amp;gt;r = 1,2,\cdots,n&amp;lt;/math&amp;gt;, either &amp;lt;math&amp;gt;X_r \leq X_i&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;Y_r \leq Y_i&amp;lt;/math&amp;gt;, or both. Find out the expected number of peripheral points.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Uniform Distribution (II)&amp;lt;/strong&amp;gt;] Derive the moment generating function of the standard uniform distribution, i.e., uniform distribution on &amp;lt;math&amp;gt;(0,1)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Exponential distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; have an exponential distribution. Show that &amp;lt;math&amp;gt;\textbf{Pr}[X&amp;gt;s+x|X&amp;gt;s] = \textbf{Pr}[X&amp;gt;x]&amp;lt;/math&amp;gt;, for &amp;lt;math&amp;gt;x,s\geq 0&amp;lt;/math&amp;gt;. This is the memoryless property. Show that the exponential distribution is the only continuous distribution with this property.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Normal distribution(I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X,Y\sim N(0,1)&amp;lt;/math&amp;gt; be two independent and identically distributed normal random variables. Let &amp;lt;math&amp;gt;Z = X-Y&amp;lt;/math&amp;gt;. Find the density function of &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;|Z|&amp;lt;/math&amp;gt; respectively.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Normal distribution(II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; have the &amp;lt;math&amp;gt;N(0,1)&amp;lt;/math&amp;gt; distribution and let &amp;lt;math&amp;gt;a&amp;gt;0&amp;lt;/math&amp;gt;. Show that the random variable &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; given by&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{equation*}&lt;br /&gt;
Y = \begin{cases}&lt;br /&gt;
X, &amp;amp; |X|&amp;lt; a \\&lt;br /&gt;
-X, &amp;amp; |X|\geq a&lt;br /&gt;
\end{cases}&lt;br /&gt;
\end{equation*}&amp;lt;/math&amp;gt;&lt;br /&gt;
has the &amp;lt;math&amp;gt;N(0,1)&amp;lt;/math&amp;gt; distribution, and find an expression for &amp;lt;math&amp;gt;\rho(a) = \textbf{Cov}(X,Y)&amp;lt;/math&amp;gt; in terms of the density function &amp;lt;math&amp;gt;\phi&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random process (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Given a real number &amp;lt;math&amp;gt;U&amp;lt;1&amp;lt;/math&amp;gt; as input of the following process, find out the expected returning value.&lt;br /&gt;
{{Theorem|&#039;&#039;Process 1&#039;&#039;|&lt;br /&gt;
:&#039;&#039;&#039;Input:&#039;&#039;&#039;  real numbers &amp;lt;math&amp;gt;U &amp;lt; 1&amp;lt;/math&amp;gt;;&lt;br /&gt;
----&lt;br /&gt;
:initialize &amp;lt;math&amp;gt;x = 1&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;count = 0&amp;lt;/math&amp;gt;;&lt;br /&gt;
:while &amp;lt;math&amp;gt; x &amp;gt; U &amp;lt;/math&amp;gt; do&lt;br /&gt;
:* choose &amp;lt;math&amp;gt;y \in (0,1)&amp;lt;/math&amp;gt; uniformly at random;&lt;br /&gt;
:* update &amp;lt;math&amp;gt;x = x * y&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;count = count + 1&amp;lt;/math&amp;gt;;&lt;br /&gt;
:return &amp;lt;math&amp;gt;count&amp;lt;/math&amp;gt;;&lt;br /&gt;
}}&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random process (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Given a real number &amp;lt;math&amp;gt;U&amp;lt;1&amp;lt;/math&amp;gt; as input of the following process, find out the expected returning value.&lt;br /&gt;
{{Theorem|&#039;&#039;Process 2&#039;&#039;|&lt;br /&gt;
:&#039;&#039;&#039;Input:&#039;&#039;&#039;  real numbers &amp;lt;math&amp;gt;U &amp;lt; 1&amp;lt;/math&amp;gt;;&lt;br /&gt;
----&lt;br /&gt;
:initialize &amp;lt;math&amp;gt;x = 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;count = 0&amp;lt;/math&amp;gt;;&lt;br /&gt;
:while &amp;lt;math&amp;gt; x &amp;lt; U &amp;lt;/math&amp;gt; do&lt;br /&gt;
:* choose &amp;lt;math&amp;gt;y \in (0,1)&amp;lt;/math&amp;gt; uniformly at random;&lt;br /&gt;
:* update &amp;lt;math&amp;gt;x = x + y&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;count = count + 1&amp;lt;/math&amp;gt;;&lt;br /&gt;
:return &amp;lt;math&amp;gt;count&amp;lt;/math&amp;gt;;&lt;br /&gt;
}}&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Random semicircle&amp;lt;/strong&amp;gt;] We sample &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; points within a circle &amp;lt;math&amp;gt;C=\{(x,y) \in \mathbb{R}^2 \mid x^2+y^2 \le 1\}&amp;lt;/math&amp;gt; independently and uniformly at random (i.e., the density function &amp;lt;math&amp;gt;f(x,y) \propto 1_{(x,y) \in C}&amp;lt;/math&amp;gt;). Find out the probability that they all lie within some semicircle of the original circle &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;. (Hint: you may apply the technique of change of variables, see [https://en.wikipedia.org/wiki/Random_variable#Functions_of_random_variables function of random variables] or Chapter 4.7 in [GS])&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Stochastic domination&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X, Y&amp;lt;/math&amp;gt; be continuous random variables. Show that &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; dominates &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; stochastically if and only if &amp;lt;math&amp;gt;\mathbb{E}[f(X)]\geq \mathbb{E}[f(Y)]&amp;lt;/math&amp;gt; for any non-decreasing function &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; for which the expectations exist.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Modes of Convergence, 15 points) (&amp;lt;strong&amp;gt;Bonus problem&amp;lt;/strong&amp;gt;)==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Connection of convergence modes (I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;(X_n)_{n \ge 1}, (Y_n)_{n \ge 1}, X, Y&amp;lt;/math&amp;gt; be random variables and &amp;lt;math&amp;gt;c\in\mathbb{R}&amp;lt;/math&amp;gt; be a real number. &lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; Suppose &amp;lt;math&amp;gt;X_n \overset{D}{\to} X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_n \overset{D}{\to} c&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;X_nY_n \overset{D}{\to} cX&amp;lt;/math&amp;gt;. &lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Construct an example such that &amp;lt;math&amp;gt;X_n \overset{D}{\to} X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_n \overset{D}{\to} Y&amp;lt;/math&amp;gt; but &amp;lt;math&amp;gt;X_nY_n&amp;lt;/math&amp;gt; does not converge to &amp;lt;math&amp;gt;XY&amp;lt;/math&amp;gt; in distribution.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Connection of convergence modes (II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;(X_n)_{n \ge 1}, X&amp;lt;/math&amp;gt; be random variables. Prove that &amp;lt;math&amp;gt;X_n \overset{P}{\to} X&amp;lt;/math&amp;gt; if and only if for every subsequence &amp;lt;math&amp;gt;X_{n(m)}&amp;lt;/math&amp;gt;, there exists a further subsequence &amp;lt;math&amp;gt;Y_k = X_{n(m_k)}&amp;lt;/math&amp;gt; that converges almost surely to &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;. (Hint: you may use the first Borel-Cantelli lemma.)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Extension of Borel-Cantelli Lemma&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;(A_n)_{n \ge 1}&amp;lt;/math&amp;gt; be events. Suppose &amp;lt;math&amp;gt;\sum_{n \ge 1} \mathbf{Pr}(A_n)=+\infty&amp;lt;/math&amp;gt;. Show that &lt;br /&gt;
&amp;lt;math&amp;gt;\mathbf{Pr}(A_n \text{ i.o.}) \ge \limsup_{n \to \infty} \frac{ \left(\sum_{k=1}^n\mathbf{Pr}(A_k)\right)^2 }{\sum_{1\le j,k \le n} \mathbf{Pr}(A_j \cap A_k)}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (LLN and CLT, 15 points + 5 points) ==&lt;br /&gt;
&amp;lt;strong&amp;gt;In this problem, you may apply the results of Laws of Large Numbers (LLN) and the Central Limit Theorem (CLT) to solve the problems.&amp;lt;/strong&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;St. Petersburg paradox&amp;lt;/strong&amp;gt;] Consider the well-known game involving a fair coin. In this game, if it takes &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; tosses to obtain a head, you will win &amp;lt;math&amp;gt;2^k&amp;lt;/math&amp;gt; dollars as the reward. Despite the game&#039;s expected reward being infinite, people tend to offer relatively modest amounts to participate. The following provides a mathematical explanation for this phenomenon.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
For each &amp;lt;math&amp;gt;n \ge 1&amp;lt;/math&amp;gt;, let &amp;lt;math&amp;gt;X_{n,1}, X_{n,2},\ldots, X_{n,k}&amp;lt;/math&amp;gt; be independent random variables. Furthermore, let &amp;lt;math&amp;gt;b_n &amp;gt; 0&amp;lt;/math&amp;gt; be real numbers with &amp;lt;math&amp;gt;b_n \to \infty&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widetilde{X}_{n,k} = X_{n,k} \mathbf{1}_{|X_{n,k}| \le b_n}&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;1 \le k \le n&amp;lt;/math&amp;gt;. If &amp;lt;math&amp;gt;\sum_{k=1}^n \mathbf{Pr}(|X_{n,k}| &amp;gt; b_n) \to 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;b_n^{-2} \sum_{k=1}^n \mathbf{E}[\widetilde{X}_{n,k}^2] \to 0&amp;lt;/math&amp;gt; when &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, then &amp;lt;math&amp;gt;(S_n-a_n)/b_n \overset{P}{\to} 0 &amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;S_n = \sum_{k=1}^n X_{n,k}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;a_n = \sum_{k=1}^n \mathbf{E}[\widetilde{X}_{n,k}]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;S_n&amp;lt;/math&amp;gt; be the total winnings after playing &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds of the game. Prove that &amp;lt;math&amp;gt;\frac{S_n}{n \log_2 n} \overset{P}{\to} 1&amp;lt;/math&amp;gt;. (Therefore, a fair price to play this game &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; times is roughly &amp;lt;math&amp;gt;n \log_2 n&amp;lt;/math&amp;gt; dollars)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; (&amp;lt;strong&amp;gt;Bonus problem, 5 points&amp;lt;/strong&amp;gt;)&lt;br /&gt;
Let &amp;lt;math&amp;gt;S_n&amp;lt;/math&amp;gt; be the total winnings after playing &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds of the game. Prove that &amp;lt;math&amp;gt; \limsup_{n \to \infty} \frac{S_n}{n \log_2 n} = \infty&amp;lt;/math&amp;gt; almost surely. (Hint: You may use Borel-Cantelli lemmas)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Coupon collector&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; be a continuous function on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;U_1,U_2,\ldots,U_n \sim U(0,1)&amp;lt;/math&amp;gt; be independent random variables. Show that &amp;lt;math&amp;gt;I = \frac{1}{n}\sum_{i=1}^n f(U_i) \to \int_0^1 f(x) \mathrm{d}x&amp;lt;/math&amp;gt; in probability. (Remark: This holds so long as &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; is a measurable function on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;\int_0^1 |f(x)| \mathrm{d}x&amp;lt;/math&amp;gt;)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Normalized sum&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X_1,X_2,\ldots&amp;lt;/math&amp;gt; be i.i.d. random variables with &amp;lt;math&amp;gt;\mathbf{E}[X_1] = 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{Var}[X_1] = \sigma^2 \in (0,+\infty)&amp;lt;/math&amp;gt;. Show &amp;lt;math&amp;gt;\frac{\sum_{k=1}^n X_k}{\left(\sum_{k=1}^n X_k^2\right)^{1/2}} \overset{D}{\to} N(0,1)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 4 (Concentration of measure) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Tossing coins&amp;lt;/strong&amp;gt;] We repeatedly toss a fair coin (with an equal probability of heads and tails). Let the random variable &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of throws required to obtain a total of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; heads. Show that &amp;lt;math&amp;gt;\textbf{Pr}[X &amp;gt; 2n + 2\sqrt{n\log n}]\leq O(1/n)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Chernoff vs Chebyshev&amp;lt;/strong&amp;gt;] We have a standard six-sided die. Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of times a 6 occurs in &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; throws off the die. Compare the best upper bounds on &amp;lt;math&amp;gt;\textbf{Pr}[X\geq n/4]&amp;lt;/math&amp;gt; that you can obtain using Chebyshev&#039;s inequality and Chernoff bounds.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;&amp;lt;strong&amp;gt;-th moment bound]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a random variable with expectation &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; such that moment generating function &amp;lt;math&amp;gt;\mathbf{E}[\exp(t|X|)]&amp;lt;/math&amp;gt; is finite for some &amp;lt;math&amp;gt; t &amp;gt; 0 &amp;lt;/math&amp;gt;. We can use the following two kinds of tail inequalities for &amp;lt;math&amp;gt; X &amp;lt;/math&amp;gt;: &lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&amp;lt;nowiki&amp;gt;     &amp;lt;/nowiki&amp;gt;&#039;&#039;&#039;&#039;&#039;Chernoff Bound&#039;&#039;&#039;&#039;&#039;&lt;br /&gt;
:&amp;lt;math&amp;gt;&lt;br /&gt;
\begin{align}&lt;br /&gt;
\mathbf{Pr}[|X| \geq \delta] \leq \min_{t \geq 0} \frac{\mathbf{E}[e^{t|X|}]}{e^{t\delta}}&lt;br /&gt;
\end{align}&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&#039;&#039;&#039;&#039;&#039;&amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;th-Moment Bound&#039;&#039;&#039;&#039;&#039;&lt;br /&gt;
:&amp;lt;math&amp;gt;&lt;br /&gt;
\begin{align}&lt;br /&gt;
\mathbf{Pr}[|X| \geq \delta] \leq \frac{\mathbf{E}[|X|^k]}{\delta^k}&lt;br /&gt;
\end{align}&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
# Show that for each &amp;lt;math&amp;gt;\delta&amp;lt;/math&amp;gt;, there exists a choice of &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; such that the &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;th-moment bound is no weaker than the Chernoff bound. (Hint: Use the probabilistic method.)&lt;br /&gt;
# Why would we still prefer the Chernoff bound to the (seemingly) stronger &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;-th moment bound?&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Chernoff bound meets graph theory&amp;lt;/strong&amp;gt;]&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; Show that with a probability approaching 1 (as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity), the Erdős–Rényi random graph &amp;lt;math&amp;gt;\textbf{G}(n,1/2)&amp;lt;/math&amp;gt; has the property that the maximum degree is &amp;lt;math&amp;gt;(\frac{n}{2} + O(\sqrt{n\log n}))&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; Show that with a probability approaching 1 (as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity), the Erdős–Rényi random graph &amp;lt;math&amp;gt;\textbf{G}(n,1/2)&amp;lt;/math&amp;gt; has the property that the diameter is exactly 2. The diameter of a graph &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; is the maximum distance between any pair of vertices.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zhangxy</name></author>
	</entry>
</feed>