<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://tcs.nju.edu.cn/wiki/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Zouzongrui</id>
	<title>TCS Wiki - User contributions [en]</title>
	<link rel="self" type="application/atom+xml" href="https://tcs.nju.edu.cn/wiki/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Zouzongrui"/>
	<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=Special:Contributions/Zouzongrui"/>
	<updated>2026-04-30T18:31:42Z</updated>
	<subtitle>User contributions</subtitle>
	<generator>MediaWiki 1.45.3</generator>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/%E7%AC%AC%E5%9B%9B%E6%AC%A1%E4%BD%9C%E4%B8%9A%E6%8F%90%E4%BA%A4%E5%90%8D%E5%8D%95&amp;diff=13155</id>
		<title>概率论与数理统计 (Spring 2025)/第四次作业提交名单</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/%E7%AC%AC%E5%9B%9B%E6%AC%A1%E4%BD%9C%E4%B8%9A%E6%8F%90%E4%BA%A4%E5%90%8D%E5%8D%95&amp;diff=13155"/>
		<updated>2025-05-18T08:41:59Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: Created page with &amp;quot; 如有错漏请邮件联系助教. &amp;lt;center&amp;gt; {| class=&amp;quot;wikitable&amp;quot; |- ! 学号 !! 姓名 |- | 231240009 || 陈心怡  |- | 231240055 || 魏全烨  |- | 231240005 || 王甚易  |- | 231840058 || 王润希  |- | 231200035 || 葛翰飞  |- | 231840166 || 赵少凡  |- | 231240056 || 靳濡搏  |- | 231880394 || 翟笑晨  |- | 231098046 || 李方泽  |- | 231840242 || 李昊凌  |- | 231240018 || 渠翔凯  |- | 231240015 || 邱文宣  |- | 231220001 || 陈梓睿  |- | 23123...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt; 如有错漏请邮件联系助教.&lt;br /&gt;
&amp;lt;center&amp;gt;&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! 学号 !! 姓名&lt;br /&gt;
|-&lt;br /&gt;
| 231240009 || 陈心怡 &lt;br /&gt;
|-&lt;br /&gt;
| 231240055 || 魏全烨 &lt;br /&gt;
|-&lt;br /&gt;
| 231240005 || 王甚易 &lt;br /&gt;
|-&lt;br /&gt;
| 231840058 || 王润希 &lt;br /&gt;
|-&lt;br /&gt;
| 231200035 || 葛翰飞 &lt;br /&gt;
|-&lt;br /&gt;
| 231840166 || 赵少凡 &lt;br /&gt;
|-&lt;br /&gt;
| 231240056 || 靳濡搏 &lt;br /&gt;
|-&lt;br /&gt;
| 231880394 || 翟笑晨 &lt;br /&gt;
|-&lt;br /&gt;
| 231098046 || 李方泽 &lt;br /&gt;
|-&lt;br /&gt;
| 231840242 || 李昊凌 &lt;br /&gt;
|-&lt;br /&gt;
| 231240018 || 渠翔凯 &lt;br /&gt;
|-&lt;br /&gt;
| 231240015 || 邱文宣 &lt;br /&gt;
|-&lt;br /&gt;
| 231220001 || 陈梓睿 &lt;br /&gt;
|-&lt;br /&gt;
| 231230102 || 庄铸锴 &lt;br /&gt;
|-&lt;br /&gt;
| 231880152 || 李佳骏 &lt;br /&gt;
|-&lt;br /&gt;
| 231880123 || 陆博文 &lt;br /&gt;
|-&lt;br /&gt;
| 231240047 || 姜淮仁 &lt;br /&gt;
|-&lt;br /&gt;
| 231240029 || 朱非凡 &lt;br /&gt;
|-&lt;br /&gt;
| 231240060 || 侯佳松 &lt;br /&gt;
|-&lt;br /&gt;
| 231240002 || 余孟凡 &lt;br /&gt;
|-&lt;br /&gt;
| 231240024 || 陈越 &lt;br /&gt;
|-&lt;br /&gt;
| 231240025 || 李熠辰 &lt;br /&gt;
|-&lt;br /&gt;
| 231240051 || 王昱霏 &lt;br /&gt;
|-&lt;br /&gt;
| 231240011 || 陈力琰 &lt;br /&gt;
|-&lt;br /&gt;
| 231880140 || 桂天麟 &lt;br /&gt;
|-&lt;br /&gt;
| 231240038 || 肖金宇 &lt;br /&gt;
|-&lt;br /&gt;
| 231880320 || 尹铁越 &lt;br /&gt;
|-&lt;br /&gt;
| 231240013 || 南晨曦 &lt;br /&gt;
|-&lt;br /&gt;
| 231240058 || 陈小川 &lt;br /&gt;
|-&lt;br /&gt;
| 231240004 || 何梓杨 &lt;br /&gt;
|-&lt;br /&gt;
| 231240007 || 高亮节 &lt;br /&gt;
|-&lt;br /&gt;
| 231220090 || 邓淇 &lt;br /&gt;
|-&lt;br /&gt;
| 231240043 || 朱宇曦 &lt;br /&gt;
|-&lt;br /&gt;
| 231240053 || 王艺文 &lt;br /&gt;
|-&lt;br /&gt;
| 211250032 || 吕卉森 &lt;br /&gt;
|-&lt;br /&gt;
| 231220002 || 潘谟天 &lt;br /&gt;
|-&lt;br /&gt;
| 231240027 || 彭浩楠 &lt;br /&gt;
|-&lt;br /&gt;
| 231240016 || 徐冰冰 &lt;br /&gt;
|-&lt;br /&gt;
| 231240019 || 邵与乔 &lt;br /&gt;
|-&lt;br /&gt;
| 231240026 || 蒙航 &lt;br /&gt;
|-&lt;br /&gt;
| 231220005 || 樊书岩 &lt;br /&gt;
|-&lt;br /&gt;
| 231098142 || 陈逸帆 &lt;br /&gt;
|-&lt;br /&gt;
| 231240023 || 叶宸见 &lt;br /&gt;
|-&lt;br /&gt;
| 231240021 || 陈奕帆 &lt;br /&gt;
|}&lt;br /&gt;
&amp;lt;/center&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)&amp;diff=13154</id>
		<title>概率论与数理统计 (Spring 2025)</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)&amp;diff=13154"/>
		<updated>2025-05-18T08:41:31Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: /* Assignments */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Infobox&lt;br /&gt;
|name         = Infobox&lt;br /&gt;
|bodystyle    = &lt;br /&gt;
|title        = &amp;lt;font size=3&amp;gt;&#039;&#039;&#039;概率论与数理统计&#039;&#039;&#039;&amp;lt;br&amp;gt;&lt;br /&gt;
&#039;&#039;&#039;Probability Theory&#039;&#039;&#039; &amp;lt;br&amp;gt; &amp;amp; &#039;&#039;&#039;Mathematical Statistics&#039;&#039;&#039;&amp;lt;/font&amp;gt;&lt;br /&gt;
|titlestyle   = &lt;br /&gt;
&lt;br /&gt;
|image        = &lt;br /&gt;
|imagestyle   = &lt;br /&gt;
|caption      = &lt;br /&gt;
|captionstyle = &lt;br /&gt;
|headerstyle  = background:#ccf;&lt;br /&gt;
|labelstyle   = background:#ddf;&lt;br /&gt;
|datastyle    = &lt;br /&gt;
&lt;br /&gt;
|header1 =Instructor&lt;br /&gt;
|label1  = &lt;br /&gt;
|data1   = &lt;br /&gt;
|header2 = &lt;br /&gt;
|label2  = &lt;br /&gt;
|data2   = &#039;&#039;&#039;尹一通&#039;&#039;&#039;&lt;br /&gt;
|header3 = &lt;br /&gt;
|label3  = Email&lt;br /&gt;
|data3   = yinyt@nju.edu.cn  &lt;br /&gt;
|header4 =&lt;br /&gt;
|label4  = office&lt;br /&gt;
|data4   = 计算机系 804&lt;br /&gt;
|header5 = &lt;br /&gt;
|label5  = &lt;br /&gt;
|data5   = &#039;&#039;&#039;刘景铖&#039;&#039;&#039;&lt;br /&gt;
|header6 = &lt;br /&gt;
|label6  = Email&lt;br /&gt;
|data6   = liu@nju.edu.cn  &lt;br /&gt;
|header7 =&lt;br /&gt;
|label7  = office&lt;br /&gt;
|data7   = 计算机系 516&lt;br /&gt;
|header8 = Class&lt;br /&gt;
|label8  = &lt;br /&gt;
|data8   = &lt;br /&gt;
|header9 =&lt;br /&gt;
|label9  = Class meeting&lt;br /&gt;
|data9   = Monday, 10am-12pm&amp;lt;br&amp;gt;&lt;br /&gt;
Wednesday (单), 10am-12pm&amp;lt;br&amp;gt;&lt;br /&gt;
仙Ⅰ-204&lt;br /&gt;
|header10=&lt;br /&gt;
|label10 = Office hour&lt;br /&gt;
|data10  = Friday, 2pm-3pm &amp;lt;br&amp;gt;计算机系 804（尹一通）&amp;lt;br&amp;gt;计算机系 516（刘景铖）&lt;br /&gt;
|header11= Textbook&lt;br /&gt;
|label11 = &lt;br /&gt;
|data11  = &lt;br /&gt;
|header12=&lt;br /&gt;
|label12 = &lt;br /&gt;
|data12  = [[File:概率导论.jpeg|border|100px]]&lt;br /&gt;
|header13=&lt;br /&gt;
|label13 = &lt;br /&gt;
|data13  = &#039;&#039;&#039;概率导论&#039;&#039;&#039;（第2版·修订版）&amp;lt;br&amp;gt; Dimitri P. Bertsekas and John N. Tsitsiklis&amp;lt;br&amp;gt; 郑忠国 童行伟 译；人民邮电出版社 (2022)&lt;br /&gt;
|header14=&lt;br /&gt;
|label14 = &lt;br /&gt;
|data14  = [[File:Grimmett_probability.jpg|border|100px]]&lt;br /&gt;
|header15=&lt;br /&gt;
|label15 = &lt;br /&gt;
|data15  = &#039;&#039;&#039;Probability and Random Processes&#039;&#039;&#039; (4E) &amp;lt;br&amp;gt; Geoffrey Grimmett and David Stirzaker &amp;lt;br&amp;gt;  Oxford University Press (2020)&lt;br /&gt;
|header16=&lt;br /&gt;
|label16 = &lt;br /&gt;
|data16  = [[File:Probability_and_Computing_2ed.jpg|border|100px]]&lt;br /&gt;
|header17=&lt;br /&gt;
|label17 = &lt;br /&gt;
|data17  = &#039;&#039;&#039;Probability and Computing&#039;&#039;&#039; (2E) &amp;lt;br&amp;gt; Michael Mitzenmacher and Eli Upfal &amp;lt;br&amp;gt;   Cambridge University Press (2017)&lt;br /&gt;
|belowstyle = background:#ddf;&lt;br /&gt;
|below = &lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
This is the webpage for the &#039;&#039;Probability Theory and Mathematical Statistics&#039;&#039; (概率论与数理统计) class of Spring 2025. Students who take this class should check this page periodically for content updates and new announcements. &lt;br /&gt;
&lt;br /&gt;
= Announcement =&lt;br /&gt;
* TBA&lt;br /&gt;
&lt;br /&gt;
= Course info =&lt;br /&gt;
* &#039;&#039;&#039;Instructor &#039;&#039;&#039;: &lt;br /&gt;
:* [http://tcs.nju.edu.cn/yinyt/ 尹一通]：[mailto:yinyt@nju.edu.cn &amp;lt;yinyt@nju.edu.cn&amp;gt;]，计算机系 804 &lt;br /&gt;
:* [https://liuexp.github.io 刘景铖]：[mailto:liu@nju.edu.cn &amp;lt;liu@nju.edu.cn&amp;gt;]，计算机系 516 &lt;br /&gt;
* &#039;&#039;&#039;Teaching assistant&#039;&#039;&#039;:&lt;br /&gt;
** [https://sites.google.com/view/xinyuanzhang 张昕渊]：[mailto:zhangxy@smail.nju.edu.cn &amp;lt;zhangxy@smail.nju.edu.cn&amp;gt;]，计算机系 410&lt;br /&gt;
** 邹宗瑞：[mailto:zou.zongrui@smail.nju.edu.cn &amp;lt;zou.zongrui@smail.nju.edu.cn&amp;gt;]，计算机系 410&lt;br /&gt;
* &#039;&#039;&#039;Class meeting&#039;&#039;&#039;:&lt;br /&gt;
** 周一：10am-12pm，仙Ⅰ-204&lt;br /&gt;
** 周三（单）：10am-12pm，仙Ⅰ-204&lt;br /&gt;
* &#039;&#039;&#039;Office hour&#039;&#039;&#039;: &lt;br /&gt;
:* 周五 2pm-3pm, 计算机系 804（尹一通）&lt;br /&gt;
:* 周五 2pm-3pm, 计算机系 516（刘景铖）&lt;br /&gt;
:* &#039;&#039;&#039;QQ群&#039;&#039;&#039;: 974742320（申请加入需提供姓名、院系、学号）&lt;br /&gt;
&lt;br /&gt;
= Syllabus =&lt;br /&gt;
课程内容分为三大部分：&lt;br /&gt;
* &#039;&#039;&#039;经典概率论&#039;&#039;&#039;：概率空间、随机变量及其数字特征、多维与连续随机变量、极限定理等内容&lt;br /&gt;
* &#039;&#039;&#039;概率与计算&#039;&#039;&#039;：测度集中现象 (concentration of measure)、概率法 (the probabilistic method)、离散随机过程的相关专题&lt;br /&gt;
* &#039;&#039;&#039;数理统计&#039;&#039;&#039;：参数估计、假设检验、贝叶斯估计、线性回归等统计推断等概念&lt;br /&gt;
&lt;br /&gt;
对于第一和第二部分，要求清楚掌握基本概念，深刻理解关键的现象与规律以及背后的原理，并可以灵活运用所学方法求解相关问题。对于第三部分，要求熟悉数理统计的若干基本概念，以及典型的统计模型和统计推断问题。&lt;br /&gt;
&lt;br /&gt;
经过本课程的训练，力求使学生能够熟悉掌握概率的语言，并会利用概率思维来理解客观世界并对其建模，以及驾驭概率的数学工具来分析和求解专业问题。&lt;br /&gt;
&lt;br /&gt;
=== 教材与参考书 Course Materials ===&lt;br /&gt;
* &#039;&#039;&#039;[BT]&#039;&#039;&#039; 概率导论（第2版·修订版），[美]伯特瑟卡斯（Dimitri P.Bertsekas）[美]齐齐克利斯（John N.Tsitsiklis）著，郑忠国 童行伟 译，人民邮电出版社（2022）。&lt;br /&gt;
* &#039;&#039;&#039;[GS]&#039;&#039;&#039; &#039;&#039;Probability and Random Processes&#039;&#039;, by Geoffrey Grimmett and David Stirzaker; Oxford University Press; 4th edition (2020).&lt;br /&gt;
* &#039;&#039;&#039;[MU]&#039;&#039;&#039; &#039;&#039;Probability and Computing: Randomization and Probabilistic Techniques in Algorithms and Data Analysis&#039;&#039;, by Michael Mitzenmacher, Eli Upfal; Cambridge University Press; 2nd edition (2017).&lt;br /&gt;
&lt;br /&gt;
=== 成绩 Grading Policy ===&lt;br /&gt;
* 课程成绩：本课程将会有若干次作业和一次期末考试。最终成绩将由平时作业成绩和期末考试成绩综合得出。&lt;br /&gt;
* 迟交：如果有特殊的理由，无法按时完成作业，请提前联系授课老师，给出正当理由。否则迟交的作业将不被接受。&lt;br /&gt;
&lt;br /&gt;
=== &amp;lt;font color=red&amp;gt; 学术诚信 Academic Integrity &amp;lt;/font&amp;gt;===&lt;br /&gt;
学术诚信是所有从事学术活动的学生和学者最基本的职业道德底线，本课程将不遗余力的维护学术诚信规范，违反这一底线的行为将不会被容忍。&lt;br /&gt;
&lt;br /&gt;
作业完成的原则：署你名字的工作必须是你个人的贡献。在完成作业的过程中，允许讨论，前提是讨论的所有参与者均处于同等完成度。但关键想法的执行、以及作业文本的写作必须独立完成，并在作业中致谢（acknowledge）所有参与讨论的人。符合规则的讨论与致谢将不会影响得分。不允许其他任何形式的合作——尤其是与已经完成作业的同学“讨论”。&lt;br /&gt;
&lt;br /&gt;
本课程将对剽窃行为采取零容忍的态度。在完成作业过程中，对他人工作（出版物、互联网资料、其他人的作业等）直接的文本抄袭和对关键思想、关键元素的抄袭，按照 [http://www.acm.org/publications/policies/plagiarism_policy ACM Policy on Plagiarism]的解释，都将视为剽窃。剽窃者成绩将被取消。如果发现互相抄袭行为，&amp;lt;font color=red&amp;gt; 抄袭和被抄袭双方的成绩都将被取消&amp;lt;/font&amp;gt;。因此请主动防止自己的作业被他人抄袭。&lt;br /&gt;
&lt;br /&gt;
学术诚信影响学生个人的品行，也关乎整个教育系统的正常运转。为了一点分数而做出学术不端的行为，不仅使自己沦为一个欺骗者，也使他人的诚实努力失去意义。让我们一起努力维护一个诚信的环境。&lt;br /&gt;
&lt;br /&gt;
= Assignments =&lt;br /&gt;
*[[概率论与数理统计 (Spring 2025)/Problem Set 1|Problem Set 1]]  请在 2025/3/10 上课之前(10am UTC+8)提交到 [mailto:pr_nju@163.com pr_nju@163.com] (文件名为&#039;&amp;lt;font color=red &amp;gt;学号_姓名_A1.pdf&amp;lt;/font&amp;gt;&#039;).&lt;br /&gt;
** [[概率论与数理统计 (Spring 2025)/第一次作业提交名单|第一次作业提交名单]]&lt;br /&gt;
&lt;br /&gt;
*[[概率论与数理统计 (Spring 2025)/Problem Set 2|Problem Set 2]]  请在 2025/3/31 上课之前(10am UTC+8)提交到 [mailto:pr_nju@163.com pr_nju@163.com] (文件名为&#039;&amp;lt;font color=red &amp;gt;学号_姓名_A2.pdf&amp;lt;/font&amp;gt;&#039;).&lt;br /&gt;
** [[概率论与数理统计 (Spring 2025)/第二次作业提交名单|第二次作业提交名单]]&lt;br /&gt;
&lt;br /&gt;
*[[概率论与数理统计 (Spring 2025)/Problem Set 3|Problem Set 3]]  请在 2025/4/21 上课之前(10am UTC+8)提交到 [mailto:pr_nju@163.com pr_nju@163.com] (文件名为&#039;&amp;lt;font color=red &amp;gt;学号_姓名_A3.pdf&amp;lt;/font&amp;gt;&#039;).&lt;br /&gt;
** [[概率论与数理统计 (Spring 2025)/第三次作业提交名单|第三次作业提交名单]]&lt;br /&gt;
&lt;br /&gt;
*[[概率论与数理统计 (Spring 2025)/Problem Set 4|Problem Set 4]]  请在 2025/5/14 上课之前(10am UTC+8)提交到 [mailto:pr_nju@163.com pr_nju@163.com] (文件名为&#039;&amp;lt;font color=red &amp;gt;学号_姓名_A4.pdf&amp;lt;/font&amp;gt;&#039;).&lt;br /&gt;
** [[概率论与数理统计 (Spring 2025)/第四次作业提交名单|第四次作业提交名单]]&lt;br /&gt;
&lt;br /&gt;
= Lectures =&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2025/Intro.pdf 课程简介]&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2025/ProbSpace.pdf 概率空间]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 第1章&#039;&#039;&#039; 或 &#039;&#039;&#039;[GS] Chapter 1&#039;&#039;&#039;&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2025)/Entropy and volume of Hamming balls|Entropy and volume of Hamming balls]]&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2025/RandVar.pdf 随机变量]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 第2章&#039;&#039;&#039; 或 &#039;&#039;&#039;[GS] Chapter 2, Sections 3.1~3.5, 3.7&#039;&#039;&#039;&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[MU] Chapter 2&#039;&#039;&#039;&lt;br /&gt;
#* the [http://tcs.nju.edu.cn/slides/prob2024/discrete-pmf.nb &#039;&#039;&#039;Mathematica notebook file&#039;&#039;&#039;] for the PMFs of basic discrete distributions&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2025)/Average-case analysis of QuickSort|Average-case analysis of &#039;&#039;&#039;&#039;&#039;QuickSort&#039;&#039;&#039;&#039;&#039;]]&lt;br /&gt;
#* [https://www.bilibili.com/video/BV1ta411A7fp/ 高尔顿板（Galton board）视频] 和 [https://en.wikipedia.org/wiki/Galton_board 维基百科页面]&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2025/Deviation.pdf 矩与偏差]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[MU] Chapter 3&#039;&#039;&#039;&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 章节 2.4, 4.2, 4.3, 5.1&#039;&#039;&#039; 或 &#039;&#039;&#039;[GS] Sections 3.3, 3.6, 7.3&#039;&#039;&#039;&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2025)/Two-point sampling|Two-point sampling]]&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2025)/Threshold of k-clique in random graph|Threshold of &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;-clique in random graph]]&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2025)/Weierstrass Approximation Theorem|Weierstrass approximation]]&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2025/Continuous.pdf 连续分布]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 第3章, 和4.1节&#039;&#039;&#039; 或 &#039;&#039;&#039;[GS] Chapter 4&#039;&#039;&#039;&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[MU] Chapters 8, 9&#039;&#039;&#039;&lt;br /&gt;
#* [https://measure.axler.net/MIRA.pdf Measure, Integration &amp;amp; Real Analysis] by Sheldon Axler&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2025)/An exercise on induced distribution|An exercise on induced distribution]]&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2025/Convergence.pdf 极限定理]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 第5章&#039;&#039;&#039; &lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[GS] Sections 5.7~5.10, 7.1~7.5&#039;&#039;&#039;&lt;br /&gt;
# [[Media:Concentration-Prob2025.pdf|测度集中]]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[MU] Chapters 4&#039;&#039;&#039; and &#039;&#039;&#039;Sections 13.1, 13.4~13.5&#039;&#039;&#039;&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[GS] Sections 5.11, 12.1~12.3, 7.8~7.9&#039;&#039;&#039;&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2024)/Hoeffding&#039;s lemma|Hoeffding&#039;s lemma]]&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2025)/Bounded Difference Method|The bounded difference method]]&lt;br /&gt;
# [[Media:Process-Prob2025.pdf|随机过程]]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 第6章, 第7章&#039;&#039;&#039;&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[MU]  Chapters 7, Sections 13.1~13.3&#039;&#039;&#039; or &#039;&#039;&#039;[GS] Chapters 6, Sections 12.4~12.5&#039;&#039;&#039;&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2024)/OST and applications|OST and applications]]&lt;br /&gt;
&lt;br /&gt;
= Concepts =&lt;br /&gt;
* [https://plato.stanford.edu/entries/probability-interpret/ Interpretations of probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/History_of_probability History of probability]&lt;br /&gt;
* Example problems:&lt;br /&gt;
** [https://dornsifecms.usc.edu/assets/sites/520/docs/VonNeumann-ams12p36-38.pdf von Neumann&#039;s Bernoulli factory] and other [https://peteroupc.github.io/bernoulli.html Bernoulli factory algorithms]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Boy_or_Girl_paradox Boy or Girl paradox]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Monty_Hall_problem Monty Hall problem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bertrand_paradox_(probability) Bertrand paradox]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hard_spheres Hard spheres model] and [https://en.wikipedia.org/wiki/Ising_model Ising model]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/PageRank &#039;&#039;PageRank&#039;&#039;] and stationary [https://en.wikipedia.org/wiki/Random_walk random walk]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Diffusion_process Diffusion process] and [https://en.wikipedia.org/wiki/Diffusion_model diffusion model]&lt;br /&gt;
*[https://en.wikipedia.org/wiki/Probability_space Probability space]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Sample_space Sample space]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Event_(probability_theory) Event] and [https://en.wikipedia.org/wiki/Σ-algebra &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-algebra]&lt;br /&gt;
** Kolmogorov&#039;s [https://en.wikipedia.org/wiki/Probability_axioms axioms of probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Discrete_uniform_distribution Classical] and [https://en.wikipedia.org/wiki/Geometric_probability goemetric probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Boole%27s_inequality Union bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Inclusion%E2%80%93exclusion_principle Inclusion-Exclusion principle]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Boole%27s_inequality#Bonferroni_inequalities Bonferroni inequalities]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Conditional_probability Conditional probability]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chain_rule_(probability) Chain rule]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_total_probability Law of total probability]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bayes%27_theorem Bayes&#039; law]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Independence_(probability_theory) Independence] &lt;br /&gt;
** [https://en.wikipedia.org/wiki/Pairwise_independence Pairwise independence]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Random_variable Random variable]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cumulative_distribution_function Cumulative distribution function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_mass_function Probability mass function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_density_function Probability density function]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Multivariate_random_variable Random vector]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Joint_probability_distribution Joint probability distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Conditional_probability_distribution Conditional probability distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Marginal_distribution Marginal distribution]&lt;br /&gt;
* Some &#039;&#039;&#039;discrete&#039;&#039;&#039; probability distributions&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bernoulli_trial Bernoulli trial] and [https://en.wikipedia.org/wiki/Bernoulli_distribution Bernoulli distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Discrete_uniform_distribution Discrete uniform distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Binomial_distribution Binomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Geometric_distribution Geometric distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Negative_binomial_distribution Negative binomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hypergeometric_distribution Hypergeometric distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Poisson_distribution Poisson distribution]&lt;br /&gt;
** and [https://en.wikipedia.org/wiki/List_of_probability_distributions#Discrete_distributions others]&lt;br /&gt;
* Balls into bins model&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Multinomial_distribution Multinomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Birthday_problem Birthday problem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Coupon_collector%27s_problem Coupon collector]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Balls_into_bins_problem Occupancy problem]&lt;br /&gt;
* Random graphs&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Erd%C5%91s%E2%80%93R%C3%A9nyi_model Erdős–Rényi random graph model]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Galton%E2%80%93Watson_process Galton–Watson branching process]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Expected_value Expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_the_unconscious_statistician Law of the unconscious statistician, &#039;&#039;LOTUS&#039;&#039;]&lt;br /&gt;
** [https://dlsun.github.io/probability/linearity.html Linearity of expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Conditional_expectation Conditional expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_total_expectation Law of total expectation]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Markov%27s_inequality Markov&#039;s inequality]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Chebyshev%27s_inequality Chebyshev&#039;s inequality]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Moment_(mathematics) Moment] related&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Central_moment Central moment]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Variance Variance]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Standard_deviation Standard deviation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Covariance Covariance]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Correlation Correlation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Correlation_does_not_imply_causation Correlation does not imply causation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Skewness Skewness]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Kurtosis Kurtosis]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Moment_problem Moment problem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Measure_(mathematics) Measure]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Borel_set Borel set]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Lebesgue_integration Lebesgue integration]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Measurable_function Measurable function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cantor_set Cantor set]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Non-measurable_set Non-measurable set]&lt;br /&gt;
*[https://en.wikipedia.org/wiki/Probability_density_function Probability density function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_distribution#Absolutely_continuous_probability_distribution Continuous probability distribution]&lt;br /&gt;
**[https://en.wikipedia.org/wiki/Conditional_probability_distribution#Conditional_continuous_distributions Conditional continuous distributions]&lt;br /&gt;
**[https://en.wikipedia.org/wiki/Convolution_of_probability_distributions Convolution of probability distributions] and [https://en.wikipedia.org/wiki/List_of_convolutions_of_probability_distributions List of convolutions of probability distributions]&lt;br /&gt;
* Some &#039;&#039;&#039;continuous&#039;&#039;&#039; probability distributions&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Continuous_uniform_distribution Continuous uniform distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Exponential_distribution Exponential distribution] and [https://en.wikipedia.org/wiki/Poisson_point_process Poisson point process]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Normal_distribution Normal (Gaussion) distribution]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Gaussian_function Gaussian function] and [https://en.wikipedia.org/wiki/Gaussian_integral Gaussian integral]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Standard_normal_table Standard normal table]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/68%E2%80%9395%E2%80%9399.7_rule 68–95–99.7 rule]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Sum_of_normally_distributed_random_variables Sum of normally distributed random variables]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Multivariate_normal_distribution Multivariate normal distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chi-squared_distribution Chi-squared distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Gamma_distribution Gamma distribution] and [https://en.wikipedia.org/wiki/Gamma_function Gamma function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Beta_distribution Beta distribution] and [https://en.wikipedia.org/wiki/Beta_function Beta function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cauchy_distribution Cauchy distribution]&lt;br /&gt;
** and [https://en.wikipedia.org/wiki/List_of_probability_distributions#Absolutely_continuous_distributions others]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Inverse_transform_sampling Inverse transform sampling]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stochastic_dominance Stochastic dominance] and [https://en.wikipedia.org/wiki/Coupling_(probability) Coupling] &lt;br /&gt;
* [https://en.wikipedia.org/wiki/Moment-generating_function Moment-generating function]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Convergence_of_random_variables Convergence of random variables]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Pointwise_convergence Pointwise convergence]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Convergence_of_measures#Weak_convergence_of_measures Weak convergence of measures]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Convergence_in_measure Convergence in measure]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Skorokhod%27s_representation_theorem Skorokhod&#039;s representation theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Continuous_mapping_theorem Continuous mapping theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Borel%E2%80%93Cantelli_lemma Borel–Cantelli lemma] and [https://en.wikipedia.org/wiki/Zero%E2%80%93one_law zero-one laws]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Law_of_large_numbers Law of large numbers]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Central_limit_theorem Central limit theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/De_Moivre%E2%80%93Laplace_theorem De Moivre–Laplace theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Berry%E2%80%93Esseen_theorem Berry–Esseen theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Characteristic_function_(probability_theory) Characteristic function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Lévy%27s_continuity_theorem Lévy&#039;s continuity theorem]&lt;br /&gt;
* Concentration of measures:&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Moment-generating_function Moment-generating function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chernoff_bound Chernoff bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hoeffding%27s_inequality Hoeffding&#039;s bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Doob_martingale#McDiarmid&#039;s_inequality McDiarmid&#039;s inequality]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Doob_martingale Doob martingale] and [https://en.wikipedia.org/wiki/Azuma%27s_inequality Azuma&#039;s inequality]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Sub-Gaussian_distribution Sub-Gaussian distribution]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stochastic_process Stochastic process]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stopping_time Stopping time]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Martingale_(probability_theory) Martingale]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Optional_stopping_theorem Optional stopping theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Wald%27s_equation Wald&#039;s equation]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Markov_chain Markov chain]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Markov_property Markov property] and [https://en.wikipedia.org/wiki/Memorylessness memorylessness]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Stochastic_matrix Stochastic matrix]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Stationary_distribution Stationary distribution]&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_4&amp;diff=13105</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 4</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_4&amp;diff=13105"/>
		<updated>2025-04-28T13:25:42Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: /* Problem 1 (Continuous Random Variables, 30 points) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
*Bonus problem为附加题（选做）。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 4==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Continuous Random Variables, 30 points)==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Density function&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Determine the value of &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;f(x) = C\exp(-x-e^{-x}), x\in \mathbb{R}&amp;lt;/math&amp;gt; is a probability density function (PDF) for a continuous random variable.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Independence&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be independent and identically distributed continuous random variables with cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; and probability density function (PDF) &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt;. Find out the density functions of &amp;lt;math&amp;gt;V = \max\{X,Y\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;U = \min\{X,Y\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Correlation&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be uniformly distributed on &amp;lt;math&amp;gt;(-1,1)&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_k = \cos(k \pi X)&amp;lt;/math&amp;gt; for &amp;lt;math&amp;gt;k=1,2,\ldots,n&amp;lt;/math&amp;gt;. Are the random variables &amp;lt;math&amp;gt;Y_1, Y_2, \ldots, Y_n&amp;lt;/math&amp;gt; correlated? independent? You should prove your claim rigorously.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Expectation of random variables (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a continuous random variable with mean &amp;lt;math&amp;gt;\mu&amp;lt;/math&amp;gt; and cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X \ge 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\int_{0}^a F(x) dx = \int_{a}^{\infty} [1-F(x)] dx&amp;lt;/math&amp;gt; if and only if &amp;lt;math&amp;gt;a = \mu&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; has finite variance. Show that &amp;lt;math&amp;gt;g(a) = \mathbb{E}((X-a)^2)&amp;lt;/math&amp;gt; achieves the minimum when &amp;lt;math&amp;gt;a = \mu&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Expectation of random variables (II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X, Y&amp;lt;/math&amp;gt; be two independent and identically distributed continuous random variables with cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;. Furthermore, &amp;lt;math&amp;gt;X,Y \ge 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbb{E}[|X-Y|] = 2 \left(\mathbb{E}[X] - \int_{0}^{\infty} (1-F(x))^2 dx\right)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Conditional distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be two random variables. The joint density of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; is given by &amp;lt;math&amp;gt;f(x,y) = c(x^2 - y^2)e^{-x}&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;0\leq x &amp;lt;\infty&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-x\leq y \leq x&amp;lt;/math&amp;gt;. Here, &amp;lt;math&amp;gt;c\in \mathbb{R}_+&amp;lt;/math&amp;gt; is a constant. Find out the conditional distribution of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;, given &amp;lt;math&amp;gt;X = x&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Uniform Distribution (I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;P_i = (X_i,Y_i), 1\leq i\leq n&amp;lt;/math&amp;gt;, be independent, uniformly distributed points in the unit square &amp;lt;math&amp;gt;[0,1]^2&amp;lt;/math&amp;gt;. A point &amp;lt;math&amp;gt;P_i&amp;lt;/math&amp;gt; is called &amp;quot;peripheral&amp;quot; if, for all &amp;lt;math&amp;gt;r = 1,2,\cdots,n&amp;lt;/math&amp;gt;, either &amp;lt;math&amp;gt;X_r \leq X_i&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;Y_r \leq Y_i&amp;lt;/math&amp;gt;, or both. Find out the expected number of peripheral points.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Uniform Distribution (II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X,Y&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; be independent and uniformly distributed on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt;. Find the joint density function of &amp;lt;math&amp;gt;XY&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Z^2&amp;lt;/math&amp;gt;, and show that &amp;lt;math&amp;gt;\textbf{Pr}[XY&amp;lt;Z^2] = 5/9&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Exponential distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; have an exponential distribution. Show that &amp;lt;math&amp;gt;\textbf{Pr}[X&amp;gt;s+x|X&amp;gt;s] = \textbf{Pr}[X&amp;gt;x]&amp;lt;/math&amp;gt;, for &amp;lt;math&amp;gt;x,s\geq 0&amp;lt;/math&amp;gt;. This is the memoryless property. Show that the exponential distribution is the only continuous distribution with this property.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Normal distribution(I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X,Y\sim N(0,1)&amp;lt;/math&amp;gt; be two independent and identically distributed normal random variables. Let &amp;lt;math&amp;gt;Z = X-Y&amp;lt;/math&amp;gt;. Find the density function of &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;|Z|&amp;lt;/math&amp;gt; respectively.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Normal distribution(II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; have the &amp;lt;math&amp;gt;N(0,1)&amp;lt;/math&amp;gt; distribution and let &amp;lt;math&amp;gt;a&amp;gt;0&amp;lt;/math&amp;gt;. Show that the random variable &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; given by&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{equation*}&lt;br /&gt;
Y = \begin{cases}&lt;br /&gt;
X, &amp;amp; |X|&amp;lt; a \\&lt;br /&gt;
-X, &amp;amp; |X|\geq a&lt;br /&gt;
\end{cases}&lt;br /&gt;
\end{equation*}&amp;lt;/math&amp;gt;&lt;br /&gt;
has the &amp;lt;math&amp;gt;N(0,1)&amp;lt;/math&amp;gt; distribution, and find an expression for &amp;lt;math&amp;gt;\rho(a) = \textbf{Cov}(X,Y)&amp;lt;/math&amp;gt; in terms of the density function &amp;lt;math&amp;gt;\phi&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random process&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Given a real number &amp;lt;math&amp;gt;U&amp;lt;1&amp;lt;/math&amp;gt; as input of the following process, find out the expected returning value.&lt;br /&gt;
{{Theorem|&#039;&#039;Process&#039;&#039;|&lt;br /&gt;
:&#039;&#039;&#039;Input:&#039;&#039;&#039;  real numbers &amp;lt;math&amp;gt;U &amp;lt; 1&amp;lt;/math&amp;gt;;&lt;br /&gt;
----&lt;br /&gt;
:initialize &amp;lt;math&amp;gt;x = 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;count = 0&amp;lt;/math&amp;gt;;&lt;br /&gt;
:while &amp;lt;math&amp;gt; x &amp;lt; U &amp;lt;/math&amp;gt; do&lt;br /&gt;
:* choose &amp;lt;math&amp;gt;y \in (0,1)&amp;lt;/math&amp;gt; uniformly at random;&lt;br /&gt;
:* update &amp;lt;math&amp;gt;x = x + y&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;count = count + 1&amp;lt;/math&amp;gt;;&lt;br /&gt;
:return &amp;lt;math&amp;gt;count&amp;lt;/math&amp;gt;;&lt;br /&gt;
}}&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Random semicircle&amp;lt;/strong&amp;gt;] We sample &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; points within a circle &amp;lt;math&amp;gt;C=\{(x,y) \in \mathbb{R}^2 \mid x^2+y^2 \le 1\}&amp;lt;/math&amp;gt; independently and uniformly at random (i.e., the density function &amp;lt;math&amp;gt;f(x,y) \propto 1_{(x,y) \in C}&amp;lt;/math&amp;gt;). Find out the probability that they all lie within some semicircle of the original circle &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;. (Hint: you may apply the technique of change of variables, see [https://en.wikipedia.org/wiki/Random_variable#Functions_of_random_variables function of random variables] or Chapter 4.7 in [GS])&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Stochastic domination&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X, Y&amp;lt;/math&amp;gt; be continuous random variables. Show that &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; dominates &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; stochastically if and only if &amp;lt;math&amp;gt;\mathbb{E}[f(X)]\geq \mathbb{E}[f(Y)]&amp;lt;/math&amp;gt; for any non-decreasing function &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; for which the expectations exist.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Concentration of measure, 10 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Tossing coins&amp;lt;/strong&amp;gt;] We repeatedly toss a fair coin (with an equal probability of heads and tails). Let the random variable &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of throws required to obtain a total of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; heads. Show that &amp;lt;math&amp;gt;\textbf{Pr}[X &amp;gt; 2n + 2\sqrt{n\log n}]\leq O(1/n)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Chernoff vs Chebyshev&amp;lt;/strong&amp;gt;] We have a standard six-sided die. Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of times a 6 occurs in &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; throws off the die. Compare the best upper bounds on &amp;lt;math&amp;gt;\textbf{Pr}[X\geq n/4]&amp;lt;/math&amp;gt; that you can obtain using Chebyshev&#039;s inequality and Chernoff bounds.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;&amp;lt;strong&amp;gt;-th moment bound]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a random variable with expectation &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; such that moment generating function &amp;lt;math&amp;gt;\mathbf{E}[\exp(t|X|)]&amp;lt;/math&amp;gt; is finite for some &amp;lt;math&amp;gt; t &amp;gt; 0 &amp;lt;/math&amp;gt;. We can use the following two kinds of tail inequalities for &amp;lt;math&amp;gt; X &amp;lt;/math&amp;gt;: &lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&amp;lt;nowiki&amp;gt;     &amp;lt;/nowiki&amp;gt;&#039;&#039;&#039;&#039;&#039;Chernoff Bound&#039;&#039;&#039;&#039;&#039;&lt;br /&gt;
:&amp;lt;math&amp;gt;&lt;br /&gt;
\begin{align}&lt;br /&gt;
\mathbf{Pr}[|X| \geq \delta] \leq \min_{t \geq 0} \frac{\mathbf{E}[e^{t|X|}]}{e^{t\delta}}&lt;br /&gt;
\end{align}&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&#039;&#039;&#039;&#039;&#039;&amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;th-Moment Bound&#039;&#039;&#039;&#039;&#039;&lt;br /&gt;
:&amp;lt;math&amp;gt;&lt;br /&gt;
\begin{align}&lt;br /&gt;
\mathbf{Pr}[|X| \geq \delta] \leq \frac{\mathbf{E}[|X|^k]}{\delta^k}&lt;br /&gt;
\end{align}&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
# Show that for each &amp;lt;math&amp;gt;\delta&amp;lt;/math&amp;gt;, there exists a choice of &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; such that the &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;th-moment bound is no weaker than the Chernoff bound. (Hint: Use the probabilistic method.)&lt;br /&gt;
# Why would we still prefer the Chernoff bound to the (seemingly) stronger &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;-th moment bound?&lt;br /&gt;
&lt;br /&gt;
* [&#039;&#039;&#039;Cut size in random graph&#039;&#039;&#039;] Show that with probability at least &amp;lt;math&amp;gt;2/3&amp;lt;/math&amp;gt;, the size of the max-cut in [[wikipedia:Erdős–Rényi_model#Definition|Erdős–Rényi random graph]] &amp;lt;math&amp;gt;G(n,1/2)&amp;lt;/math&amp;gt; is at most &amp;lt;math&amp;gt;n^2/8 + O(n^{1.5})&amp;lt;/math&amp;gt;. In the &amp;lt;math&amp;gt;G(n,1/2)&amp;lt;/math&amp;gt; model, each edge is included in the graph with probability &amp;lt;math&amp;gt;1/2&amp;lt;/math&amp;gt;, independently of every other edge.&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Strong Law of Large Numbers, 15 points, &#039;&#039;&#039;Bonus Problem&#039;&#039;&#039;)==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
Throughout this problem, we assume &amp;lt;math&amp;gt;X_1,X_2,\ldots,&amp;lt;/math&amp;gt; be jointly independent square-integrable real random variables of mean zero. We will prove the strong law of large numbers by Kolmogorov maximal inequality.&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Kolmogorov maximal inequality&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;S_n = \sum_{i=1}^n X_i&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{Pr}\left(\max_{1 \le i \le n} |S_i| \ge t \right) \le \frac{n \mathbf{Var}(X_1)}{t^2}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Convergence of random series&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Var}(X_i) &amp;lt; \infty&amp;lt;/math&amp;gt;. Prove that the series &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} X_i&amp;lt;/math&amp;gt; is almost surely convergent.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Strong law of large numbers&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Prove the strong law of large numbers using previous propositions.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 4 (LLN and CLT, 15 points + 10 points) ==&lt;br /&gt;
&amp;lt;strong&amp;gt;In this problem, you may apply the results of Laws of Large Numbers (LLN) and the Central Limit Theorem (CLT) to solve the problems.&amp;lt;/strong&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;St. Petersburg paradox&amp;lt;/strong&amp;gt;] Consider the well-known game involving a fair coin. In this game, if it takes &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; tosses to obtain a head, you will win &amp;lt;math&amp;gt;2^k&amp;lt;/math&amp;gt; dollars as the reward. Despite the game&#039;s expected reward being infinite, people tend to offer relatively modest amounts to participate. The following provides a mathematical explanation for this phenomenon.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
For each &amp;lt;math&amp;gt;n \ge 1&amp;lt;/math&amp;gt;, let &amp;lt;math&amp;gt;X_{n,1}, X_{n,2},\ldots, X_{n,k}&amp;lt;/math&amp;gt; be independent random variables. Furthermore, let &amp;lt;math&amp;gt;b_n &amp;gt; 0&amp;lt;/math&amp;gt; be real numbers with &amp;lt;math&amp;gt;b_n \to \infty&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widetilde{X}_{n,k} = X_{n,k} \mathbf{1}_{|X_{n,k}| \le b_n}&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;1 \le k \le n&amp;lt;/math&amp;gt;. If &amp;lt;math&amp;gt;\sum_{k=1}^n \mathbf{Pr}(|X_{n,k}| &amp;gt; b_n) \to 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;b_n^{-2} \sum_{k=1}^n \mathbf{E}[\widetilde{X}_{n,k}^2] \to 0&amp;lt;/math&amp;gt; when &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, then &amp;lt;math&amp;gt;(S_n-a_n)/b_n \overset{P}{\to} 0 &amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;S_n = \sum_{k=1}^n X_{n,k}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;a_n = \sum_{k=1}^n \mathbf{E}[\widetilde{X}_{n,k}]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;S_n&amp;lt;/math&amp;gt; be the total winnings after playing &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds of the game. Prove that &amp;lt;math&amp;gt;\frac{S_n}{n \log_2 n} \overset{P}{\to} 1&amp;lt;/math&amp;gt;. (Therefore, a fair price to play this game &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; times is roughly &amp;lt;math&amp;gt;n \log_2 n&amp;lt;/math&amp;gt; dollars)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; (&amp;lt;strong&amp;gt;Bonus problem, 5 points&amp;lt;/strong&amp;gt;)&lt;br /&gt;
Let &amp;lt;math&amp;gt;S_n&amp;lt;/math&amp;gt; be the total winnings after playing &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds of the game. Prove that &amp;lt;math&amp;gt; \limsup_{n \to \infty} \frac{S_n}{n \log_2 n} = \infty&amp;lt;/math&amp;gt; almost surely. (Hint: You may use Borel-Cantelli lemmas)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; (&amp;lt;strong&amp;gt;Bonus problem, 5 points&amp;lt;/strong&amp;gt;)&lt;br /&gt;
Let &amp;lt;math&amp;gt;X_1,X_2,\ldots,X_n&amp;lt;/math&amp;gt; are i.i.d. copies of a random variable &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; with the standard Cauchy distribution (i.e., the probability density function of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; is &amp;lt;math&amp;gt;\frac{1}{\pi} \frac{1}{1+x^2}&amp;lt;/math&amp;gt;), show that &amp;lt;math&amp;gt;\frac{S_n}{n \log n}&amp;lt;/math&amp;gt; converges in probability to a constant &amp;lt;math&amp;gt;\frac{2}{\pi}&amp;lt;/math&amp;gt; but is almost surely unbounded, where &amp;lt;math&amp;gt;S_n = \sum_{i=1}^n |X_i|&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Monte Carlo integration&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; be a continuous function on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;U_1,U_2,\ldots,U_n \sim U(0,1)&amp;lt;/math&amp;gt; be independent random variables. Show that &amp;lt;math&amp;gt;I = \frac{1}{n}\sum_{i=1}^n f(U_i) \to \int_0^1 f(x) \mathrm{d}x&amp;lt;/math&amp;gt; in probability. (Remark: This holds so long as &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; is a measurable function on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;\int_0^1 |f(x)| \mathrm{d}x &amp;lt; \infty&amp;lt;/math&amp;gt;)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Stirling&#039;s formula&amp;lt;/strong&amp;gt;]&lt;br /&gt;
By considering the central limit theorem for the sum of independent Poisson-distributed random variables, show that&lt;br /&gt;
&amp;lt;math&amp;gt;n! \sim \sqrt{2\pi n} \cdot \left(\frac{n}{\mathrm{e}}\right)^n&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_4&amp;diff=13098</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 4</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_4&amp;diff=13098"/>
		<updated>2025-04-27T12:53:39Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: Created page with &amp;quot;*每道题目的解答都要有完整的解题过程，中英文不限。  *我们推荐大家使用LaTeX, markdown等对作业进行排版。  *Bonus problem为附加题（选做）。  == Assumption throughout Problem Set 4== &amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;  &amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;  &amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; u...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
*Bonus problem为附加题（选做）。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 4==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Continuous Random Variables, 30 points)==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Density function&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Determine the value of &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;f(x) = C\exp(-x-e^{-x}), x\in \mathbb{R}&amp;lt;/math&amp;gt; is a probability density function (PDF) for a continuous random variable.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Independence&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be independent and identically distributed continuous random variables with cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; and probability density function (PDF) &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt;. Find out the density functions of &amp;lt;math&amp;gt;V = \max\{X,Y\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;U = \min\{X,Y\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Correlation&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be uniformly distributed on &amp;lt;math&amp;gt;(-1,1)&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_k = \cos(k \pi X)&amp;lt;/math&amp;gt; for &amp;lt;math&amp;gt;k=1,2,\ldots,n&amp;lt;/math&amp;gt;. Are the random variables &amp;lt;math&amp;gt;Y_1, Y_2, \ldots, Y_n&amp;lt;/math&amp;gt; correlated? independent? You should prove your claim rigorously.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Expectation of random variables (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a continuous random variable with mean &amp;lt;math&amp;gt;\mu&amp;lt;/math&amp;gt; and cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X \ge 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\int_{0}^a F(x) dx = \int_{a}^{\infty} [1-F(x)] dx&amp;lt;/math&amp;gt; if and only if &amp;lt;math&amp;gt;a = \mu&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; has finite variance. Show that &amp;lt;math&amp;gt;g(a) = \mathbb{E}((X-a)^2)&amp;lt;/math&amp;gt; achieves the minimum when &amp;lt;math&amp;gt;a = \mu&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Expectation of random variables (II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X, Y&amp;lt;/math&amp;gt; be two independent and identically distributed continuous random variables with cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;. Furthermore, &amp;lt;math&amp;gt;X,Y \ge 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbb{E}[|X-Y|] = 2 \left(\mathbb{E}[X] - \int_{0}^{\infty} (1-F(x))^2 dx\right)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Conditional distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be two random variables. The joint density of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; is given by &amp;lt;math&amp;gt;f(x,y) = c(x^2 - y^2)e^{-x}&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;0\leq x &amp;lt;\infty&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-x\leq y \leq x&amp;lt;/math&amp;gt;. Here, &amp;lt;math&amp;gt;c\in \mathbb{R}_+&amp;lt;/math&amp;gt; is a constant. Find out the conditional distribution of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;, given &amp;lt;math&amp;gt;X = x&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Uniform Distribution (I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;P_i = (X_i,Y_i), 1\leq i\leq n&amp;lt;/math&amp;gt;, be independent, uniformly distributed points in the unit square &amp;lt;math&amp;gt;[0,1]^2&amp;lt;/math&amp;gt;. A point &amp;lt;math&amp;gt;P_i&amp;lt;/math&amp;gt; is called &amp;quot;peripheral&amp;quot; if, for all &amp;lt;math&amp;gt;r = 1,2,\cdots,n&amp;lt;/math&amp;gt;, either &amp;lt;math&amp;gt;X_r \leq X_i&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;Y_r \leq Y_i&amp;lt;/math&amp;gt;, or both. Find out the expected number of peripheral points.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Uniform Distribution (II)&amp;lt;/strong&amp;gt;] Derive the moment generating function of the standard uniform distribution, i.e., uniform distribution on &amp;lt;math&amp;gt;(0,1)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Exponential distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; have an exponential distribution. Show that &amp;lt;math&amp;gt;\textbf{Pr}[X&amp;gt;s+x|X&amp;gt;s] = \textbf{Pr}[X&amp;gt;x]&amp;lt;/math&amp;gt;, for &amp;lt;math&amp;gt;x,s\geq 0&amp;lt;/math&amp;gt;. This is the memoryless property. Show that the exponential distribution is the only continuous distribution with this property.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Normal distribution(I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X,Y\sim N(0,1)&amp;lt;/math&amp;gt; be two independent and identically distributed normal random variables. Let &amp;lt;math&amp;gt;Z = X-Y&amp;lt;/math&amp;gt;. Find the density function of &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;|Z|&amp;lt;/math&amp;gt; respectively.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Normal distribution(II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; have the &amp;lt;math&amp;gt;N(0,1)&amp;lt;/math&amp;gt; distribution and let &amp;lt;math&amp;gt;a&amp;gt;0&amp;lt;/math&amp;gt;. Show that the random variable &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; given by&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{equation*}&lt;br /&gt;
Y = \begin{cases}&lt;br /&gt;
X, &amp;amp; |X|&amp;lt; a \\&lt;br /&gt;
-X, &amp;amp; |X|\geq a&lt;br /&gt;
\end{cases}&lt;br /&gt;
\end{equation*}&amp;lt;/math&amp;gt;&lt;br /&gt;
has the &amp;lt;math&amp;gt;N(0,1)&amp;lt;/math&amp;gt; distribution, and find an expression for &amp;lt;math&amp;gt;\rho(a) = \textbf{Cov}(X,Y)&amp;lt;/math&amp;gt; in terms of the density function &amp;lt;math&amp;gt;\phi&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random process&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Given a real number &amp;lt;math&amp;gt;U&amp;lt;1&amp;lt;/math&amp;gt; as input of the following process, find out the expected returning value.&lt;br /&gt;
{{Theorem|&#039;&#039;Process&#039;&#039;|&lt;br /&gt;
:&#039;&#039;&#039;Input:&#039;&#039;&#039;  real numbers &amp;lt;math&amp;gt;U &amp;lt; 1&amp;lt;/math&amp;gt;;&lt;br /&gt;
----&lt;br /&gt;
:initialize &amp;lt;math&amp;gt;x = 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;count = 0&amp;lt;/math&amp;gt;;&lt;br /&gt;
:while &amp;lt;math&amp;gt; x &amp;lt; U &amp;lt;/math&amp;gt; do&lt;br /&gt;
:* choose &amp;lt;math&amp;gt;y \in (0,1)&amp;lt;/math&amp;gt; uniformly at random;&lt;br /&gt;
:* update &amp;lt;math&amp;gt;x = x + y&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;count = count + 1&amp;lt;/math&amp;gt;;&lt;br /&gt;
:return &amp;lt;math&amp;gt;count&amp;lt;/math&amp;gt;;&lt;br /&gt;
}}&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Random semicircle&amp;lt;/strong&amp;gt;] We sample &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; points within a circle &amp;lt;math&amp;gt;C=\{(x,y) \in \mathbb{R}^2 \mid x^2+y^2 \le 1\}&amp;lt;/math&amp;gt; independently and uniformly at random (i.e., the density function &amp;lt;math&amp;gt;f(x,y) \propto 1_{(x,y) \in C}&amp;lt;/math&amp;gt;). Find out the probability that they all lie within some semicircle of the original circle &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;. (Hint: you may apply the technique of change of variables, see [https://en.wikipedia.org/wiki/Random_variable#Functions_of_random_variables function of random variables] or Chapter 4.7 in [GS])&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Stochastic domination&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X, Y&amp;lt;/math&amp;gt; be continuous random variables. Show that &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; dominates &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; stochastically if and only if &amp;lt;math&amp;gt;\mathbb{E}[f(X)]\geq \mathbb{E}[f(Y)]&amp;lt;/math&amp;gt; for any non-decreasing function &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; for which the expectations exist.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Concentration of measure, 10 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Tossing coins&amp;lt;/strong&amp;gt;] We repeatedly toss a fair coin (with an equal probability of heads and tails). Let the random variable &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of throws required to obtain a total of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; heads. Show that &amp;lt;math&amp;gt;\textbf{Pr}[X &amp;gt; 2n + 2\sqrt{n\log n}]\leq O(1/n)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Chernoff vs Chebyshev&amp;lt;/strong&amp;gt;] We have a standard six-sided die. Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of times a 6 occurs in &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; throws off the die. Compare the best upper bounds on &amp;lt;math&amp;gt;\textbf{Pr}[X\geq n/4]&amp;lt;/math&amp;gt; that you can obtain using Chebyshev&#039;s inequality and Chernoff bounds.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;&amp;lt;strong&amp;gt;-th moment bound]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a random variable with expectation &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; such that moment generating function &amp;lt;math&amp;gt;\mathbf{E}[\exp(t|X|)]&amp;lt;/math&amp;gt; is finite for some &amp;lt;math&amp;gt; t &amp;gt; 0 &amp;lt;/math&amp;gt;. We can use the following two kinds of tail inequalities for &amp;lt;math&amp;gt; X &amp;lt;/math&amp;gt;: &lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&amp;lt;nowiki&amp;gt;     &amp;lt;/nowiki&amp;gt;&#039;&#039;&#039;&#039;&#039;Chernoff Bound&#039;&#039;&#039;&#039;&#039;&lt;br /&gt;
:&amp;lt;math&amp;gt;&lt;br /&gt;
\begin{align}&lt;br /&gt;
\mathbf{Pr}[|X| \geq \delta] \leq \min_{t \geq 0} \frac{\mathbf{E}[e^{t|X|}]}{e^{t\delta}}&lt;br /&gt;
\end{align}&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&#039;&#039;&#039;&#039;&#039;&amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;th-Moment Bound&#039;&#039;&#039;&#039;&#039;&lt;br /&gt;
:&amp;lt;math&amp;gt;&lt;br /&gt;
\begin{align}&lt;br /&gt;
\mathbf{Pr}[|X| \geq \delta] \leq \frac{\mathbf{E}[|X|^k]}{\delta^k}&lt;br /&gt;
\end{align}&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
# Show that for each &amp;lt;math&amp;gt;\delta&amp;lt;/math&amp;gt;, there exists a choice of &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; such that the &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;th-moment bound is no weaker than the Chernoff bound. (Hint: Use the probabilistic method.)&lt;br /&gt;
# Why would we still prefer the Chernoff bound to the (seemingly) stronger &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;-th moment bound?&lt;br /&gt;
&lt;br /&gt;
* [&#039;&#039;&#039;Cut size in random graph&#039;&#039;&#039;] Show that with probability at least &amp;lt;math&amp;gt;2/3&amp;lt;/math&amp;gt;, the size of the max-cut in [[wikipedia:Erdős–Rényi_model#Definition|Erdős–Rényi random graph]] &amp;lt;math&amp;gt;G(n,1/2)&amp;lt;/math&amp;gt; is at most &amp;lt;math&amp;gt;n^2/8 + O(n^{1.5})&amp;lt;/math&amp;gt;. In the &amp;lt;math&amp;gt;G(n,1/2)&amp;lt;/math&amp;gt; model, each edge is included in the graph with probability &amp;lt;math&amp;gt;1/2&amp;lt;/math&amp;gt;, independently of every other edge.&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Modes of Convergence and Characteristic Function, 15 points) (&amp;lt;strong&amp;gt;Bonus problem&amp;lt;/strong&amp;gt;)==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Connection of convergence modes (I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;(X_n)_{n \ge 1}, (Y_n)_{n \ge 1}, X, Y&amp;lt;/math&amp;gt; be random variables and &amp;lt;math&amp;gt;c\in\mathbb{R}&amp;lt;/math&amp;gt; be a real number. &lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; Suppose &amp;lt;math&amp;gt;X_n \overset{D}{\to} X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_n \overset{D}{\to} c&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;X_nY_n \overset{D}{\to} cX&amp;lt;/math&amp;gt;. &lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Construct an example such that &amp;lt;math&amp;gt;X_n \overset{D}{\to} X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_n \overset{D}{\to} Y&amp;lt;/math&amp;gt; but &amp;lt;math&amp;gt;X_nY_n&amp;lt;/math&amp;gt; does not converge to &amp;lt;math&amp;gt;XY&amp;lt;/math&amp;gt; in distribution.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Connection of convergence modes (II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;(X_n)_{n \ge 1}, X&amp;lt;/math&amp;gt; be random variables. Prove that &amp;lt;math&amp;gt;X_n \overset{P}{\to} X&amp;lt;/math&amp;gt; if and only if for every subsequence &amp;lt;math&amp;gt;X_{n(m)}&amp;lt;/math&amp;gt;, there exists a further subsequence &amp;lt;math&amp;gt;Y_k = X_{n(m_k)}&amp;lt;/math&amp;gt; that converges almost surely to &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;. (Hint: you may use the first Borel-Cantelli lemma.)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Characteristic Function&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; is a discrete random variable only taking integer values and &amp;lt;math&amp;gt;\mathbf{E}[X]=0&amp;lt;/math&amp;gt;. Suppose furthermore that there is no infinite subprogression &amp;lt;math&amp;gt;a+q\mathbb{Z}&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;\mathbb{Z}&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;q&amp;gt;1&amp;lt;/math&amp;gt; for which &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; takes values almost surely in &amp;lt;math&amp;gt;a+q \mathbb{Z}&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;|\phi_X(t)|&amp;lt;1&amp;lt;/math&amp;gt; for &amp;lt;math&amp;gt;0 &amp;lt; t \le \pi&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\phi_X(t)=\mathbf{E}[\mathrm{e}^{itX}]&amp;lt;/math&amp;gt; is the characteristic function of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 4 (LLN and CLT, 15 points + 10 points) ==&lt;br /&gt;
&amp;lt;strong&amp;gt;In this problem, you may apply the results of Laws of Large Numbers (LLN) and the Central Limit Theorem (CLT) to solve the problems.&amp;lt;/strong&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;St. Petersburg paradox&amp;lt;/strong&amp;gt;] Consider the well-known game involving a fair coin. In this game, if it takes &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; tosses to obtain a head, you will win &amp;lt;math&amp;gt;2^k&amp;lt;/math&amp;gt; dollars as the reward. Despite the game&#039;s expected reward being infinite, people tend to offer relatively modest amounts to participate. The following provides a mathematical explanation for this phenomenon.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
For each &amp;lt;math&amp;gt;n \ge 1&amp;lt;/math&amp;gt;, let &amp;lt;math&amp;gt;X_{n,1}, X_{n,2},\ldots, X_{n,k}&amp;lt;/math&amp;gt; be independent random variables. Furthermore, let &amp;lt;math&amp;gt;b_n &amp;gt; 0&amp;lt;/math&amp;gt; be real numbers with &amp;lt;math&amp;gt;b_n \to \infty&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widetilde{X}_{n,k} = X_{n,k} \mathbf{1}_{|X_{n,k}| \le b_n}&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;1 \le k \le n&amp;lt;/math&amp;gt;. If &amp;lt;math&amp;gt;\sum_{k=1}^n \mathbf{Pr}(|X_{n,k}| &amp;gt; b_n) \to 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;b_n^{-2} \sum_{k=1}^n \mathbf{E}[\widetilde{X}_{n,k}^2] \to 0&amp;lt;/math&amp;gt; when &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, then &amp;lt;math&amp;gt;(S_n-a_n)/b_n \overset{P}{\to} 0 &amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;S_n = \sum_{k=1}^n X_{n,k}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;a_n = \sum_{k=1}^n \mathbf{E}[\widetilde{X}_{n,k}]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;S_n&amp;lt;/math&amp;gt; be the total winnings after playing &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds of the game. Prove that &amp;lt;math&amp;gt;\frac{S_n}{n \log_2 n} \overset{P}{\to} 1&amp;lt;/math&amp;gt;. (Therefore, a fair price to play this game &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; times is roughly &amp;lt;math&amp;gt;n \log_2 n&amp;lt;/math&amp;gt; dollars)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; (&amp;lt;strong&amp;gt;Bonus problem, 5 points&amp;lt;/strong&amp;gt;)&lt;br /&gt;
Let &amp;lt;math&amp;gt;S_n&amp;lt;/math&amp;gt; be the total winnings after playing &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds of the game. Prove that &amp;lt;math&amp;gt; \limsup_{n \to \infty} \frac{S_n}{n \log_2 n} = \infty&amp;lt;/math&amp;gt; almost surely. (Hint: You may use Borel-Cantelli lemmas)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; (&amp;lt;strong&amp;gt;Bonus problem, 5 points&amp;lt;/strong&amp;gt;)&lt;br /&gt;
Let &amp;lt;math&amp;gt;X_1,X_2,\ldots,X_n&amp;lt;/math&amp;gt; are i.i.d. copies of a random variable &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; with the standard Cauchy distribution (i.e., the probability density function of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; is &amp;lt;math&amp;gt;\frac{1}{\pi} \frac{1}{1+x^2}&amp;lt;/math&amp;gt;), show that &amp;lt;math&amp;gt;\frac{S_n}{n \log n}&amp;lt;/math&amp;gt; converges in probability to a constant &amp;lt;math&amp;gt;\frac{2}{\pi}&amp;lt;/math&amp;gt; but is almost surely unbounded, where &amp;lt;math&amp;gt;S_n = \sum_{i=1}^n |X_i|&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Monte Carlo Integration&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; be a continuous function on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;U_1,U_2,\ldots,U_n \sim U(0,1)&amp;lt;/math&amp;gt; be independent random variables. Show that &amp;lt;math&amp;gt;I = \frac{1}{n}\sum_{i=1}^n f(U_i) \to \int_0^1 f(x) \mathrm{d}x&amp;lt;/math&amp;gt; in probability. (Remark: This holds so long as &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; is a measurable function on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;\int_0^1 |f(x)| \mathrm{d}x &amp;lt; \infty&amp;lt;/math&amp;gt;)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Square root&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X_1,X_2,\ldots&amp;lt;/math&amp;gt; be i.i.d. random variables with &amp;lt;math&amp;gt;X_i \ge 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{E}[X_1] = 1&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{Var}[X_1] = \sigma^2 \in (0,+\infty)&amp;lt;/math&amp;gt;. Show &amp;lt;math&amp;gt;2(\sqrt{S_n} - \sqrt{n}) \overset{D}{\to} \sigma N(0,1)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;S_n = \sum_{i=1}^n X_i&amp;lt;/math&amp;gt;. (Hint: You may use the statement in Problem 3.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)&amp;diff=13097</id>
		<title>概率论与数理统计 (Spring 2025)</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)&amp;diff=13097"/>
		<updated>2025-04-27T12:52:41Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: /* Assignments */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Infobox&lt;br /&gt;
|name         = Infobox&lt;br /&gt;
|bodystyle    = &lt;br /&gt;
|title        = &amp;lt;font size=3&amp;gt;&#039;&#039;&#039;概率论与数理统计&#039;&#039;&#039;&amp;lt;br&amp;gt;&lt;br /&gt;
&#039;&#039;&#039;Probability Theory&#039;&#039;&#039; &amp;lt;br&amp;gt; &amp;amp; &#039;&#039;&#039;Mathematical Statistics&#039;&#039;&#039;&amp;lt;/font&amp;gt;&lt;br /&gt;
|titlestyle   = &lt;br /&gt;
&lt;br /&gt;
|image        = &lt;br /&gt;
|imagestyle   = &lt;br /&gt;
|caption      = &lt;br /&gt;
|captionstyle = &lt;br /&gt;
|headerstyle  = background:#ccf;&lt;br /&gt;
|labelstyle   = background:#ddf;&lt;br /&gt;
|datastyle    = &lt;br /&gt;
&lt;br /&gt;
|header1 =Instructor&lt;br /&gt;
|label1  = &lt;br /&gt;
|data1   = &lt;br /&gt;
|header2 = &lt;br /&gt;
|label2  = &lt;br /&gt;
|data2   = &#039;&#039;&#039;尹一通&#039;&#039;&#039;&lt;br /&gt;
|header3 = &lt;br /&gt;
|label3  = Email&lt;br /&gt;
|data3   = yinyt@nju.edu.cn  &lt;br /&gt;
|header4 =&lt;br /&gt;
|label4  = office&lt;br /&gt;
|data4   = 计算机系 804&lt;br /&gt;
|header5 = &lt;br /&gt;
|label5  = &lt;br /&gt;
|data5   = &#039;&#039;&#039;刘景铖&#039;&#039;&#039;&lt;br /&gt;
|header6 = &lt;br /&gt;
|label6  = Email&lt;br /&gt;
|data6   = liu@nju.edu.cn  &lt;br /&gt;
|header7 =&lt;br /&gt;
|label7  = office&lt;br /&gt;
|data7   = 计算机系 516&lt;br /&gt;
|header8 = Class&lt;br /&gt;
|label8  = &lt;br /&gt;
|data8   = &lt;br /&gt;
|header9 =&lt;br /&gt;
|label9  = Class meeting&lt;br /&gt;
|data9   = Monday, 10am-12pm&amp;lt;br&amp;gt;&lt;br /&gt;
Wednesday (单), 10am-12pm&amp;lt;br&amp;gt;&lt;br /&gt;
仙Ⅰ-204&lt;br /&gt;
|header10=&lt;br /&gt;
|label10 = Office hour&lt;br /&gt;
|data10  = Friday, 2pm-3pm &amp;lt;br&amp;gt;计算机系 804（尹一通）&amp;lt;br&amp;gt;计算机系 516（刘景铖）&lt;br /&gt;
|header11= Textbook&lt;br /&gt;
|label11 = &lt;br /&gt;
|data11  = &lt;br /&gt;
|header12=&lt;br /&gt;
|label12 = &lt;br /&gt;
|data12  = [[File:概率导论.jpeg|border|100px]]&lt;br /&gt;
|header13=&lt;br /&gt;
|label13 = &lt;br /&gt;
|data13  = &#039;&#039;&#039;概率导论&#039;&#039;&#039;（第2版·修订版）&amp;lt;br&amp;gt; Dimitri P. Bertsekas and John N. Tsitsiklis&amp;lt;br&amp;gt; 郑忠国 童行伟 译；人民邮电出版社 (2022)&lt;br /&gt;
|header14=&lt;br /&gt;
|label14 = &lt;br /&gt;
|data14  = [[File:Grimmett_probability.jpg|border|100px]]&lt;br /&gt;
|header15=&lt;br /&gt;
|label15 = &lt;br /&gt;
|data15  = &#039;&#039;&#039;Probability and Random Processes&#039;&#039;&#039; (4E) &amp;lt;br&amp;gt; Geoffrey Grimmett and David Stirzaker &amp;lt;br&amp;gt;  Oxford University Press (2020)&lt;br /&gt;
|header16=&lt;br /&gt;
|label16 = &lt;br /&gt;
|data16  = [[File:Probability_and_Computing_2ed.jpg|border|100px]]&lt;br /&gt;
|header17=&lt;br /&gt;
|label17 = &lt;br /&gt;
|data17  = &#039;&#039;&#039;Probability and Computing&#039;&#039;&#039; (2E) &amp;lt;br&amp;gt; Michael Mitzenmacher and Eli Upfal &amp;lt;br&amp;gt;   Cambridge University Press (2017)&lt;br /&gt;
|belowstyle = background:#ddf;&lt;br /&gt;
|below = &lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
This is the webpage for the &#039;&#039;Probability Theory and Mathematical Statistics&#039;&#039; (概率论与数理统计) class of Spring 2025. Students who take this class should check this page periodically for content updates and new announcements. &lt;br /&gt;
&lt;br /&gt;
= Announcement =&lt;br /&gt;
* TBA&lt;br /&gt;
&lt;br /&gt;
= Course info =&lt;br /&gt;
* &#039;&#039;&#039;Instructor &#039;&#039;&#039;: &lt;br /&gt;
:* [http://tcs.nju.edu.cn/yinyt/ 尹一通]：[mailto:yinyt@nju.edu.cn &amp;lt;yinyt@nju.edu.cn&amp;gt;]，计算机系 804 &lt;br /&gt;
:* [https://liuexp.github.io 刘景铖]：[mailto:liu@nju.edu.cn &amp;lt;liu@nju.edu.cn&amp;gt;]，计算机系 516 &lt;br /&gt;
* &#039;&#039;&#039;Teaching assistant&#039;&#039;&#039;:&lt;br /&gt;
** [https://sites.google.com/view/xinyuanzhang 张昕渊]：[mailto:zhangxy@smail.nju.edu.cn &amp;lt;zhangxy@smail.nju.edu.cn&amp;gt;]，计算机系 410&lt;br /&gt;
** 邹宗瑞：[mailto:zou.zongrui@smail.nju.edu.cn &amp;lt;zou.zongrui@smail.nju.edu.cn&amp;gt;]，计算机系 410&lt;br /&gt;
* &#039;&#039;&#039;Class meeting&#039;&#039;&#039;:&lt;br /&gt;
** 周一：10am-12pm，仙Ⅰ-204&lt;br /&gt;
** 周三（单）：10am-12pm，仙Ⅰ-204&lt;br /&gt;
* &#039;&#039;&#039;Office hour&#039;&#039;&#039;: &lt;br /&gt;
:* 周五 2pm-3pm, 计算机系 804（尹一通）&lt;br /&gt;
:* 周五 2pm-3pm, 计算机系 516（刘景铖）&lt;br /&gt;
:* &#039;&#039;&#039;QQ群&#039;&#039;&#039;: 974742320（申请加入需提供姓名、院系、学号）&lt;br /&gt;
&lt;br /&gt;
= Syllabus =&lt;br /&gt;
课程内容分为三大部分：&lt;br /&gt;
* &#039;&#039;&#039;经典概率论&#039;&#039;&#039;：概率空间、随机变量及其数字特征、多维与连续随机变量、极限定理等内容&lt;br /&gt;
* &#039;&#039;&#039;概率与计算&#039;&#039;&#039;：测度集中现象 (concentration of measure)、概率法 (the probabilistic method)、离散随机过程的相关专题&lt;br /&gt;
* &#039;&#039;&#039;数理统计&#039;&#039;&#039;：参数估计、假设检验、贝叶斯估计、线性回归等统计推断等概念&lt;br /&gt;
&lt;br /&gt;
对于第一和第二部分，要求清楚掌握基本概念，深刻理解关键的现象与规律以及背后的原理，并可以灵活运用所学方法求解相关问题。对于第三部分，要求熟悉数理统计的若干基本概念，以及典型的统计模型和统计推断问题。&lt;br /&gt;
&lt;br /&gt;
经过本课程的训练，力求使学生能够熟悉掌握概率的语言，并会利用概率思维来理解客观世界并对其建模，以及驾驭概率的数学工具来分析和求解专业问题。&lt;br /&gt;
&lt;br /&gt;
=== 教材与参考书 Course Materials ===&lt;br /&gt;
* &#039;&#039;&#039;[BT]&#039;&#039;&#039; 概率导论（第2版·修订版），[美]伯特瑟卡斯（Dimitri P.Bertsekas）[美]齐齐克利斯（John N.Tsitsiklis）著，郑忠国 童行伟 译，人民邮电出版社（2022）。&lt;br /&gt;
* &#039;&#039;&#039;[GS]&#039;&#039;&#039; &#039;&#039;Probability and Random Processes&#039;&#039;, by Geoffrey Grimmett and David Stirzaker; Oxford University Press; 4th edition (2020).&lt;br /&gt;
* &#039;&#039;&#039;[MU]&#039;&#039;&#039; &#039;&#039;Probability and Computing: Randomization and Probabilistic Techniques in Algorithms and Data Analysis&#039;&#039;, by Michael Mitzenmacher, Eli Upfal; Cambridge University Press; 2nd edition (2017).&lt;br /&gt;
&lt;br /&gt;
=== 成绩 Grading Policy ===&lt;br /&gt;
* 课程成绩：本课程将会有若干次作业和一次期末考试。最终成绩将由平时作业成绩和期末考试成绩综合得出。&lt;br /&gt;
* 迟交：如果有特殊的理由，无法按时完成作业，请提前联系授课老师，给出正当理由。否则迟交的作业将不被接受。&lt;br /&gt;
&lt;br /&gt;
=== &amp;lt;font color=red&amp;gt; 学术诚信 Academic Integrity &amp;lt;/font&amp;gt;===&lt;br /&gt;
学术诚信是所有从事学术活动的学生和学者最基本的职业道德底线，本课程将不遗余力的维护学术诚信规范，违反这一底线的行为将不会被容忍。&lt;br /&gt;
&lt;br /&gt;
作业完成的原则：署你名字的工作必须是你个人的贡献。在完成作业的过程中，允许讨论，前提是讨论的所有参与者均处于同等完成度。但关键想法的执行、以及作业文本的写作必须独立完成，并在作业中致谢（acknowledge）所有参与讨论的人。符合规则的讨论与致谢将不会影响得分。不允许其他任何形式的合作——尤其是与已经完成作业的同学“讨论”。&lt;br /&gt;
&lt;br /&gt;
本课程将对剽窃行为采取零容忍的态度。在完成作业过程中，对他人工作（出版物、互联网资料、其他人的作业等）直接的文本抄袭和对关键思想、关键元素的抄袭，按照 [http://www.acm.org/publications/policies/plagiarism_policy ACM Policy on Plagiarism]的解释，都将视为剽窃。剽窃者成绩将被取消。如果发现互相抄袭行为，&amp;lt;font color=red&amp;gt; 抄袭和被抄袭双方的成绩都将被取消&amp;lt;/font&amp;gt;。因此请主动防止自己的作业被他人抄袭。&lt;br /&gt;
&lt;br /&gt;
学术诚信影响学生个人的品行，也关乎整个教育系统的正常运转。为了一点分数而做出学术不端的行为，不仅使自己沦为一个欺骗者，也使他人的诚实努力失去意义。让我们一起努力维护一个诚信的环境。&lt;br /&gt;
&lt;br /&gt;
= Assignments =&lt;br /&gt;
*[[概率论与数理统计 (Spring 2025)/Problem Set 1|Problem Set 1]]  请在 2025/3/10 上课之前(10am UTC+8)提交到 [mailto:pr_nju@163.com pr_nju@163.com] (文件名为&#039;&amp;lt;font color=red &amp;gt;学号_姓名_A1.pdf&amp;lt;/font&amp;gt;&#039;).&lt;br /&gt;
** [[概率论与数理统计 (Spring 2025)/第一次作业提交名单|第一次作业提交名单]]&lt;br /&gt;
&lt;br /&gt;
*[[概率论与数理统计 (Spring 2025)/Problem Set 2|Problem Set 2]]  请在 2025/3/31 上课之前(10am UTC+8)提交到 [mailto:pr_nju@163.com pr_nju@163.com] (文件名为&#039;&amp;lt;font color=red &amp;gt;学号_姓名_A2.pdf&amp;lt;/font&amp;gt;&#039;).&lt;br /&gt;
** [[概率论与数理统计 (Spring 2025)/第二次作业提交名单|第二次作业提交名单]]&lt;br /&gt;
&lt;br /&gt;
*[[概率论与数理统计 (Spring 2025)/Problem Set 3|Problem Set 3]]  请在 2025/4/21 上课之前(10am UTC+8)提交到 [mailto:pr_nju@163.com pr_nju@163.com] (文件名为&#039;&amp;lt;font color=red &amp;gt;学号_姓名_A3.pdf&amp;lt;/font&amp;gt;&#039;).&lt;br /&gt;
** [[概率论与数理统计 (Spring 2025)/第三次作业提交名单|第三次作业提交名单]]&lt;br /&gt;
&lt;br /&gt;
*[[概率论与数理统计 (Spring 2025)/Problem Set 4|Problem Set 4]]  请在 2025/5/14 上课之前(10am UTC+8)提交到 [mailto:pr_nju@163.com pr_nju@163.com] (文件名为&#039;&amp;lt;font color=red &amp;gt;学号_姓名_A4.pdf&amp;lt;/font&amp;gt;&#039;).&lt;br /&gt;
&lt;br /&gt;
= Lectures =&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2025/Intro.pdf 课程简介]&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2025/ProbSpace.pdf 概率空间]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 第1章&#039;&#039;&#039; 或 &#039;&#039;&#039;[GS] Chapter 1&#039;&#039;&#039;&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2025)/Entropy and volume of Hamming balls|Entropy and volume of Hamming balls]]&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2025/RandVar.pdf 随机变量]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 第2章&#039;&#039;&#039; 或 &#039;&#039;&#039;[GS] Chapter 2, Sections 3.1~3.5, 3.7&#039;&#039;&#039;&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[MU] Chapter 2&#039;&#039;&#039;&lt;br /&gt;
#* the [http://tcs.nju.edu.cn/slides/prob2024/discrete-pmf.nb &#039;&#039;&#039;Mathematica notebook file&#039;&#039;&#039;] for the PMFs of basic discrete distributions&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2025)/Average-case analysis of QuickSort|Average-case analysis of &#039;&#039;&#039;&#039;&#039;QuickSort&#039;&#039;&#039;&#039;&#039;]]&lt;br /&gt;
#* [https://www.bilibili.com/video/BV1ta411A7fp/ 高尔顿板（Galton board）视频] 和 [https://en.wikipedia.org/wiki/Galton_board 维基百科页面]&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2025/Deviation.pdf 矩与偏差]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[MU] Chapter 3&#039;&#039;&#039;&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 章节 2.4, 4.2, 4.3, 5.1&#039;&#039;&#039; 或 &#039;&#039;&#039;[GS] Sections 3.3, 3.6, 7.3&#039;&#039;&#039;&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2025)/Two-point sampling|Two-point sampling]]&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2025)/Threshold of k-clique in random graph|Threshold of &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;-clique in random graph]]&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2025)/Weierstrass Approximation Theorem|Weierstrass approximation]]&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2025/Continuous.pdf 连续分布]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 第3章, 和4.1节&#039;&#039;&#039; 或 &#039;&#039;&#039;[GS] Chapter 4&#039;&#039;&#039;&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[MU] Chapters 8, 9&#039;&#039;&#039;&lt;br /&gt;
#* [https://measure.axler.net/MIRA.pdf Measure, Integration &amp;amp; Real Analysis] by Sheldon Axler&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2025)/An exercise on induced distribution|An exercise on induced distribution]]&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2025/Convergence.pdf 极限定理]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 第5章&#039;&#039;&#039; &lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[GS] Sections 5.7~5.10, 7.1~7.5&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= Concepts =&lt;br /&gt;
* [https://plato.stanford.edu/entries/probability-interpret/ Interpretations of probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/History_of_probability History of probability]&lt;br /&gt;
* Example problems:&lt;br /&gt;
** [https://dornsifecms.usc.edu/assets/sites/520/docs/VonNeumann-ams12p36-38.pdf von Neumann&#039;s Bernoulli factory] and other [https://peteroupc.github.io/bernoulli.html Bernoulli factory algorithms]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Boy_or_Girl_paradox Boy or Girl paradox]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Monty_Hall_problem Monty Hall problem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bertrand_paradox_(probability) Bertrand paradox]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hard_spheres Hard spheres model] and [https://en.wikipedia.org/wiki/Ising_model Ising model]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/PageRank &#039;&#039;PageRank&#039;&#039;] and stationary [https://en.wikipedia.org/wiki/Random_walk random walk]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Diffusion_process Diffusion process] and [https://en.wikipedia.org/wiki/Diffusion_model diffusion model]&lt;br /&gt;
*[https://en.wikipedia.org/wiki/Probability_space Probability space]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Sample_space Sample space]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Event_(probability_theory) Event] and [https://en.wikipedia.org/wiki/Σ-algebra &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-algebra]&lt;br /&gt;
** Kolmogorov&#039;s [https://en.wikipedia.org/wiki/Probability_axioms axioms of probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Discrete_uniform_distribution Classical] and [https://en.wikipedia.org/wiki/Geometric_probability goemetric probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Boole%27s_inequality Union bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Inclusion%E2%80%93exclusion_principle Inclusion-Exclusion principle]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Boole%27s_inequality#Bonferroni_inequalities Bonferroni inequalities]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Conditional_probability Conditional probability]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chain_rule_(probability) Chain rule]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_total_probability Law of total probability]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bayes%27_theorem Bayes&#039; law]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Independence_(probability_theory) Independence] &lt;br /&gt;
** [https://en.wikipedia.org/wiki/Pairwise_independence Pairwise independence]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Random_variable Random variable]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cumulative_distribution_function Cumulative distribution function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_mass_function Probability mass function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_density_function Probability density function]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Multivariate_random_variable Random vector]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Joint_probability_distribution Joint probability distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Conditional_probability_distribution Conditional probability distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Marginal_distribution Marginal distribution]&lt;br /&gt;
* Some &#039;&#039;&#039;discrete&#039;&#039;&#039; probability distributions&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bernoulli_trial Bernoulli trial] and [https://en.wikipedia.org/wiki/Bernoulli_distribution Bernoulli distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Discrete_uniform_distribution Discrete uniform distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Binomial_distribution Binomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Geometric_distribution Geometric distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Negative_binomial_distribution Negative binomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hypergeometric_distribution Hypergeometric distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Poisson_distribution Poisson distribution]&lt;br /&gt;
** and [https://en.wikipedia.org/wiki/List_of_probability_distributions#Discrete_distributions others]&lt;br /&gt;
* Balls into bins model&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Multinomial_distribution Multinomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Birthday_problem Birthday problem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Coupon_collector%27s_problem Coupon collector]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Balls_into_bins_problem Occupancy problem]&lt;br /&gt;
* Random graphs&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Erd%C5%91s%E2%80%93R%C3%A9nyi_model Erdős–Rényi random graph model]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Galton%E2%80%93Watson_process Galton–Watson branching process]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Expected_value Expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_the_unconscious_statistician Law of the unconscious statistician, &#039;&#039;LOTUS&#039;&#039;]&lt;br /&gt;
** [https://dlsun.github.io/probability/linearity.html Linearity of expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Conditional_expectation Conditional expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_total_expectation Law of total expectation]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Markov%27s_inequality Markov&#039;s inequality]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Chebyshev%27s_inequality Chebyshev&#039;s inequality]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Moment_(mathematics) Moment] related&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Central_moment Central moment]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Variance Variance]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Standard_deviation Standard deviation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Covariance Covariance]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Correlation Correlation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Correlation_does_not_imply_causation Correlation does not imply causation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Skewness Skewness]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Kurtosis Kurtosis]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Moment_problem Moment problem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Measure_(mathematics) Measure]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Borel_set Borel set]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Lebesgue_integration Lebesgue integration]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Measurable_function Measurable function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cantor_set Cantor set]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Non-measurable_set Non-measurable set]&lt;br /&gt;
*[https://en.wikipedia.org/wiki/Probability_density_function Probability density function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_distribution#Absolutely_continuous_probability_distribution Continuous probability distribution]&lt;br /&gt;
**[https://en.wikipedia.org/wiki/Conditional_probability_distribution#Conditional_continuous_distributions Conditional continuous distributions]&lt;br /&gt;
**[https://en.wikipedia.org/wiki/Convolution_of_probability_distributions Convolution of probability distributions] and [https://en.wikipedia.org/wiki/List_of_convolutions_of_probability_distributions List of convolutions of probability distributions]&lt;br /&gt;
* Some &#039;&#039;&#039;continuous&#039;&#039;&#039; probability distributions&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Continuous_uniform_distribution Continuous uniform distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Exponential_distribution Exponential distribution] and [https://en.wikipedia.org/wiki/Poisson_point_process Poisson point process]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Normal_distribution Normal (Gaussion) distribution]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Gaussian_function Gaussian function] and [https://en.wikipedia.org/wiki/Gaussian_integral Gaussian integral]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Standard_normal_table Standard normal table]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/68%E2%80%9395%E2%80%9399.7_rule 68–95–99.7 rule]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Sum_of_normally_distributed_random_variables Sum of normally distributed random variables]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Multivariate_normal_distribution Multivariate normal distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chi-squared_distribution Chi-squared distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Gamma_distribution Gamma distribution] and [https://en.wikipedia.org/wiki/Gamma_function Gamma function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Beta_distribution Beta distribution] and [https://en.wikipedia.org/wiki/Beta_function Beta function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cauchy_distribution Cauchy distribution]&lt;br /&gt;
** and [https://en.wikipedia.org/wiki/List_of_probability_distributions#Absolutely_continuous_distributions others]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Inverse_transform_sampling Inverse transform sampling]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stochastic_dominance Stochastic dominance] and [https://en.wikipedia.org/wiki/Coupling_(probability) Coupling] &lt;br /&gt;
* [https://en.wikipedia.org/wiki/Moment-generating_function Moment-generating function]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Convergence_of_random_variables Convergence of random variables]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Pointwise_convergence Pointwise convergence]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Convergence_of_measures#Weak_convergence_of_measures Weak convergence of measures]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Convergence_in_measure Convergence in measure]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Skorokhod%27s_representation_theorem Skorokhod&#039;s representation theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Continuous_mapping_theorem Continuous mapping theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Borel%E2%80%93Cantelli_lemma Borel–Cantelli lemma] and [https://en.wikipedia.org/wiki/Zero%E2%80%93one_law zero-one laws]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Law_of_large_numbers Law of large numbers]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Central_limit_theorem Central limit theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/De_Moivre%E2%80%93Laplace_theorem De Moivre–Laplace theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Berry%E2%80%93Esseen_theorem Berry–Esseen theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Characteristic_function_(probability_theory) Characteristic function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Lévy%27s_continuity_theorem Lévy&#039;s continuity theorem]&lt;br /&gt;
* Concentration of measures:&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Moment-generating_function Moment-generating function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chernoff_bound Chernoff bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hoeffding%27s_inequality Hoeffding&#039;s bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Doob_martingale#McDiarmid&#039;s_inequality McDiarmid&#039;s inequality]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Doob_martingale Doob martingale] and [https://en.wikipedia.org/wiki/Azuma%27s_inequality Azuma&#039;s inequality]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Sub-Gaussian_distribution Sub-Gaussian distribution]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stochastic_process Stochastic process]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stopping_time Stopping time]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Martingale_(probability_theory) Martingale]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Optional_stopping_theorem Optional stopping theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Wald%27s_equation Wald&#039;s equation]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Markov_chain Markov chain]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Markov_property Markov property] and [https://en.wikipedia.org/wiki/Memorylessness memorylessness]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Stochastic_matrix Stochastic matrix]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Stationary_distribution Stationary distribution]&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/%E7%AC%AC%E4%B8%89%E6%AC%A1%E4%BD%9C%E4%B8%9A%E6%8F%90%E4%BA%A4%E5%90%8D%E5%8D%95&amp;diff=13078</id>
		<title>概率论与数理统计 (Spring 2025)/第三次作业提交名单</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/%E7%AC%AC%E4%B8%89%E6%AC%A1%E4%BD%9C%E4%B8%9A%E6%8F%90%E4%BA%A4%E5%90%8D%E5%8D%95&amp;diff=13078"/>
		<updated>2025-04-21T06:13:58Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: Created page with &amp;quot; 如有错漏请邮件联系助教. &amp;lt;center&amp;gt; {| class=&amp;quot;wikitable&amp;quot; |- ! 学号 !! 姓名 |- | 231240038 || 肖金宇  |- | 231240002 || 余孟凡  |- | 231200035 || 葛翰飞  |- | 231880394 || 翟笑晨  |- | 231240055 || 魏全烨  |- | 231880320 || 尹铁越  |- | 231880123 || 陆博文  |- | 231840166 || 赵少凡  |- | 231840242 || 李昊凌  |- | 231098046 || 李方泽  |- | 231840058 || 王润希  |- | 231240021 || 陈奕帆  |- | 231240018 || 渠翔凯  |- | 23124...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt; 如有错漏请邮件联系助教.&lt;br /&gt;
&amp;lt;center&amp;gt;&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! 学号 !! 姓名&lt;br /&gt;
|-&lt;br /&gt;
| 231240038 || 肖金宇 &lt;br /&gt;
|-&lt;br /&gt;
| 231240002 || 余孟凡 &lt;br /&gt;
|-&lt;br /&gt;
| 231200035 || 葛翰飞 &lt;br /&gt;
|-&lt;br /&gt;
| 231880394 || 翟笑晨 &lt;br /&gt;
|-&lt;br /&gt;
| 231240055 || 魏全烨 &lt;br /&gt;
|-&lt;br /&gt;
| 231880320 || 尹铁越 &lt;br /&gt;
|-&lt;br /&gt;
| 231880123 || 陆博文 &lt;br /&gt;
|-&lt;br /&gt;
| 231840166 || 赵少凡 &lt;br /&gt;
|-&lt;br /&gt;
| 231840242 || 李昊凌 &lt;br /&gt;
|-&lt;br /&gt;
| 231098046 || 李方泽 &lt;br /&gt;
|-&lt;br /&gt;
| 231840058 || 王润希 &lt;br /&gt;
|-&lt;br /&gt;
| 231240021 || 陈奕帆 &lt;br /&gt;
|-&lt;br /&gt;
| 231240018 || 渠翔凯 &lt;br /&gt;
|-&lt;br /&gt;
| 231240005 || 王甚易 &lt;br /&gt;
|-&lt;br /&gt;
| 231880140 || 桂天麟 &lt;br /&gt;
|-&lt;br /&gt;
| 231240011 || 陈力琰 &lt;br /&gt;
|-&lt;br /&gt;
| 231230102 || 庄铸锴 &lt;br /&gt;
|-&lt;br /&gt;
| 231240025 || 李熠辰 &lt;br /&gt;
|-&lt;br /&gt;
| 231220005 || 樊书岩 &lt;br /&gt;
|-&lt;br /&gt;
| 231240056 || 靳濡搏 &lt;br /&gt;
|-&lt;br /&gt;
| 231240015 || 邱文宣 &lt;br /&gt;
|-&lt;br /&gt;
| 231240060 || 侯佳松 &lt;br /&gt;
|-&lt;br /&gt;
| 231240026 || 蒙航 &lt;br /&gt;
|-&lt;br /&gt;
| 231240013 || 南晨曦 &lt;br /&gt;
|-&lt;br /&gt;
| 231240047 || 姜淮仁 &lt;br /&gt;
|-&lt;br /&gt;
| 231880152 || 李佳骏 &lt;br /&gt;
|-&lt;br /&gt;
| 231240051 || 王昱霏 &lt;br /&gt;
|-&lt;br /&gt;
| 231240058 || 陈小川 &lt;br /&gt;
|-&lt;br /&gt;
| 231240004 || 何梓杨 &lt;br /&gt;
|-&lt;br /&gt;
| 231240029 || 朱非凡 &lt;br /&gt;
|-&lt;br /&gt;
| 231240024 || 陈越 &lt;br /&gt;
|-&lt;br /&gt;
| 231220001 || 陈梓睿 &lt;br /&gt;
|-&lt;br /&gt;
| 231240053 || 王艺文 &lt;br /&gt;
|-&lt;br /&gt;
| 231240045 || 杨俊炜 &lt;br /&gt;
|-&lt;br /&gt;
| 231240016 || 徐冰冰 &lt;br /&gt;
|-&lt;br /&gt;
| 231240023 || 叶宸见 &lt;br /&gt;
|-&lt;br /&gt;
| 231240009 || 陈心怡 &lt;br /&gt;
|-&lt;br /&gt;
| 211250032 || 吕卉森 &lt;br /&gt;
|-&lt;br /&gt;
| 231220090 || 邓淇 &lt;br /&gt;
|-&lt;br /&gt;
| 231240027 || 彭浩楠 &lt;br /&gt;
|-&lt;br /&gt;
| 231220002 || 潘谟天 &lt;br /&gt;
|-&lt;br /&gt;
| 231240019 ||  邵与乔 &lt;br /&gt;
|-&lt;br /&gt;
| 231098142 || 陈逸帆 &lt;br /&gt;
|}&lt;br /&gt;
&amp;lt;/center&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)&amp;diff=13077</id>
		<title>概率论与数理统计 (Spring 2025)</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)&amp;diff=13077"/>
		<updated>2025-04-21T06:13:41Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: /* Assignments */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Infobox&lt;br /&gt;
|name         = Infobox&lt;br /&gt;
|bodystyle    = &lt;br /&gt;
|title        = &amp;lt;font size=3&amp;gt;&#039;&#039;&#039;概率论与数理统计&#039;&#039;&#039;&amp;lt;br&amp;gt;&lt;br /&gt;
&#039;&#039;&#039;Probability Theory&#039;&#039;&#039; &amp;lt;br&amp;gt; &amp;amp; &#039;&#039;&#039;Mathematical Statistics&#039;&#039;&#039;&amp;lt;/font&amp;gt;&lt;br /&gt;
|titlestyle   = &lt;br /&gt;
&lt;br /&gt;
|image        = &lt;br /&gt;
|imagestyle   = &lt;br /&gt;
|caption      = &lt;br /&gt;
|captionstyle = &lt;br /&gt;
|headerstyle  = background:#ccf;&lt;br /&gt;
|labelstyle   = background:#ddf;&lt;br /&gt;
|datastyle    = &lt;br /&gt;
&lt;br /&gt;
|header1 =Instructor&lt;br /&gt;
|label1  = &lt;br /&gt;
|data1   = &lt;br /&gt;
|header2 = &lt;br /&gt;
|label2  = &lt;br /&gt;
|data2   = &#039;&#039;&#039;尹一通&#039;&#039;&#039;&lt;br /&gt;
|header3 = &lt;br /&gt;
|label3  = Email&lt;br /&gt;
|data3   = yinyt@nju.edu.cn  &lt;br /&gt;
|header4 =&lt;br /&gt;
|label4  = office&lt;br /&gt;
|data4   = 计算机系 804&lt;br /&gt;
|header5 = &lt;br /&gt;
|label5  = &lt;br /&gt;
|data5   = &#039;&#039;&#039;刘景铖&#039;&#039;&#039;&lt;br /&gt;
|header6 = &lt;br /&gt;
|label6  = Email&lt;br /&gt;
|data6   = liu@nju.edu.cn  &lt;br /&gt;
|header7 =&lt;br /&gt;
|label7  = office&lt;br /&gt;
|data7   = 计算机系 516&lt;br /&gt;
|header8 = Class&lt;br /&gt;
|label8  = &lt;br /&gt;
|data8   = &lt;br /&gt;
|header9 =&lt;br /&gt;
|label9  = Class meeting&lt;br /&gt;
|data9   = Monday, 10am-12pm&amp;lt;br&amp;gt;&lt;br /&gt;
Wednesday (单), 10am-12pm&amp;lt;br&amp;gt;&lt;br /&gt;
仙Ⅰ-204&lt;br /&gt;
|header10=&lt;br /&gt;
|label10 = Office hour&lt;br /&gt;
|data10  = Friday, 2pm-3pm &amp;lt;br&amp;gt;计算机系 804（尹一通）&amp;lt;br&amp;gt;计算机系 516（刘景铖）&lt;br /&gt;
|header11= Textbook&lt;br /&gt;
|label11 = &lt;br /&gt;
|data11  = &lt;br /&gt;
|header12=&lt;br /&gt;
|label12 = &lt;br /&gt;
|data12  = [[File:概率导论.jpeg|border|100px]]&lt;br /&gt;
|header13=&lt;br /&gt;
|label13 = &lt;br /&gt;
|data13  = &#039;&#039;&#039;概率导论&#039;&#039;&#039;（第2版·修订版）&amp;lt;br&amp;gt; Dimitri P. Bertsekas and John N. Tsitsiklis&amp;lt;br&amp;gt; 郑忠国 童行伟 译；人民邮电出版社 (2022)&lt;br /&gt;
|header14=&lt;br /&gt;
|label14 = &lt;br /&gt;
|data14  = [[File:Grimmett_probability.jpg|border|100px]]&lt;br /&gt;
|header15=&lt;br /&gt;
|label15 = &lt;br /&gt;
|data15  = &#039;&#039;&#039;Probability and Random Processes&#039;&#039;&#039; (4E) &amp;lt;br&amp;gt; Geoffrey Grimmett and David Stirzaker &amp;lt;br&amp;gt;  Oxford University Press (2020)&lt;br /&gt;
|header16=&lt;br /&gt;
|label16 = &lt;br /&gt;
|data16  = [[File:Probability_and_Computing_2ed.jpg|border|100px]]&lt;br /&gt;
|header17=&lt;br /&gt;
|label17 = &lt;br /&gt;
|data17  = &#039;&#039;&#039;Probability and Computing&#039;&#039;&#039; (2E) &amp;lt;br&amp;gt; Michael Mitzenmacher and Eli Upfal &amp;lt;br&amp;gt;   Cambridge University Press (2017)&lt;br /&gt;
|belowstyle = background:#ddf;&lt;br /&gt;
|below = &lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
This is the webpage for the &#039;&#039;Probability Theory and Mathematical Statistics&#039;&#039; (概率论与数理统计) class of Spring 2025. Students who take this class should check this page periodically for content updates and new announcements. &lt;br /&gt;
&lt;br /&gt;
= Announcement =&lt;br /&gt;
* TBA&lt;br /&gt;
&lt;br /&gt;
= Course info =&lt;br /&gt;
* &#039;&#039;&#039;Instructor &#039;&#039;&#039;: &lt;br /&gt;
:* [http://tcs.nju.edu.cn/yinyt/ 尹一通]：[mailto:yinyt@nju.edu.cn &amp;lt;yinyt@nju.edu.cn&amp;gt;]，计算机系 804 &lt;br /&gt;
:* [https://liuexp.github.io 刘景铖]：[mailto:liu@nju.edu.cn &amp;lt;liu@nju.edu.cn&amp;gt;]，计算机系 516 &lt;br /&gt;
* &#039;&#039;&#039;Teaching assistant&#039;&#039;&#039;:&lt;br /&gt;
** [https://sites.google.com/view/xinyuanzhang 张昕渊]：[mailto:zhangxy@smail.nju.edu.cn &amp;lt;zhangxy@smail.nju.edu.cn&amp;gt;]，计算机系 410&lt;br /&gt;
** 邹宗瑞：[mailto:zou.zongrui@smail.nju.edu.cn &amp;lt;zou.zongrui@smail.nju.edu.cn&amp;gt;]，计算机系 410&lt;br /&gt;
* &#039;&#039;&#039;Class meeting&#039;&#039;&#039;:&lt;br /&gt;
** 周一：10am-12pm，仙Ⅰ-204&lt;br /&gt;
** 周三（单）：10am-12pm，仙Ⅰ-204&lt;br /&gt;
* &#039;&#039;&#039;Office hour&#039;&#039;&#039;: &lt;br /&gt;
:* 周五 2pm-3pm, 计算机系 804（尹一通）&lt;br /&gt;
:* 周五 2pm-3pm, 计算机系 516（刘景铖）&lt;br /&gt;
:* &#039;&#039;&#039;QQ群&#039;&#039;&#039;: 974742320（申请加入需提供姓名、院系、学号）&lt;br /&gt;
&lt;br /&gt;
= Syllabus =&lt;br /&gt;
课程内容分为三大部分：&lt;br /&gt;
* &#039;&#039;&#039;经典概率论&#039;&#039;&#039;：概率空间、随机变量及其数字特征、多维与连续随机变量、极限定理等内容&lt;br /&gt;
* &#039;&#039;&#039;概率与计算&#039;&#039;&#039;：测度集中现象 (concentration of measure)、概率法 (the probabilistic method)、离散随机过程的相关专题&lt;br /&gt;
* &#039;&#039;&#039;数理统计&#039;&#039;&#039;：参数估计、假设检验、贝叶斯估计、线性回归等统计推断等概念&lt;br /&gt;
&lt;br /&gt;
对于第一和第二部分，要求清楚掌握基本概念，深刻理解关键的现象与规律以及背后的原理，并可以灵活运用所学方法求解相关问题。对于第三部分，要求熟悉数理统计的若干基本概念，以及典型的统计模型和统计推断问题。&lt;br /&gt;
&lt;br /&gt;
经过本课程的训练，力求使学生能够熟悉掌握概率的语言，并会利用概率思维来理解客观世界并对其建模，以及驾驭概率的数学工具来分析和求解专业问题。&lt;br /&gt;
&lt;br /&gt;
=== 教材与参考书 Course Materials ===&lt;br /&gt;
* &#039;&#039;&#039;[BT]&#039;&#039;&#039; 概率导论（第2版·修订版），[美]伯特瑟卡斯（Dimitri P.Bertsekas）[美]齐齐克利斯（John N.Tsitsiklis）著，郑忠国 童行伟 译，人民邮电出版社（2022）。&lt;br /&gt;
* &#039;&#039;&#039;[GS]&#039;&#039;&#039; &#039;&#039;Probability and Random Processes&#039;&#039;, by Geoffrey Grimmett and David Stirzaker; Oxford University Press; 4th edition (2020).&lt;br /&gt;
* &#039;&#039;&#039;[MU]&#039;&#039;&#039; &#039;&#039;Probability and Computing: Randomization and Probabilistic Techniques in Algorithms and Data Analysis&#039;&#039;, by Michael Mitzenmacher, Eli Upfal; Cambridge University Press; 2nd edition (2017).&lt;br /&gt;
&lt;br /&gt;
=== 成绩 Grading Policy ===&lt;br /&gt;
* 课程成绩：本课程将会有若干次作业和一次期末考试。最终成绩将由平时作业成绩和期末考试成绩综合得出。&lt;br /&gt;
* 迟交：如果有特殊的理由，无法按时完成作业，请提前联系授课老师，给出正当理由。否则迟交的作业将不被接受。&lt;br /&gt;
&lt;br /&gt;
=== &amp;lt;font color=red&amp;gt; 学术诚信 Academic Integrity &amp;lt;/font&amp;gt;===&lt;br /&gt;
学术诚信是所有从事学术活动的学生和学者最基本的职业道德底线，本课程将不遗余力的维护学术诚信规范，违反这一底线的行为将不会被容忍。&lt;br /&gt;
&lt;br /&gt;
作业完成的原则：署你名字的工作必须是你个人的贡献。在完成作业的过程中，允许讨论，前提是讨论的所有参与者均处于同等完成度。但关键想法的执行、以及作业文本的写作必须独立完成，并在作业中致谢（acknowledge）所有参与讨论的人。符合规则的讨论与致谢将不会影响得分。不允许其他任何形式的合作——尤其是与已经完成作业的同学“讨论”。&lt;br /&gt;
&lt;br /&gt;
本课程将对剽窃行为采取零容忍的态度。在完成作业过程中，对他人工作（出版物、互联网资料、其他人的作业等）直接的文本抄袭和对关键思想、关键元素的抄袭，按照 [http://www.acm.org/publications/policies/plagiarism_policy ACM Policy on Plagiarism]的解释，都将视为剽窃。剽窃者成绩将被取消。如果发现互相抄袭行为，&amp;lt;font color=red&amp;gt; 抄袭和被抄袭双方的成绩都将被取消&amp;lt;/font&amp;gt;。因此请主动防止自己的作业被他人抄袭。&lt;br /&gt;
&lt;br /&gt;
学术诚信影响学生个人的品行，也关乎整个教育系统的正常运转。为了一点分数而做出学术不端的行为，不仅使自己沦为一个欺骗者，也使他人的诚实努力失去意义。让我们一起努力维护一个诚信的环境。&lt;br /&gt;
&lt;br /&gt;
= Assignments =&lt;br /&gt;
*[[概率论与数理统计 (Spring 2025)/Problem Set 1|Problem Set 1]]  请在 2025/3/10 上课之前(10am UTC+8)提交到 [mailto:pr_nju@163.com pr_nju@163.com] (文件名为&#039;&amp;lt;font color=red &amp;gt;学号_姓名_A1.pdf&amp;lt;/font&amp;gt;&#039;).&lt;br /&gt;
** [[概率论与数理统计 (Spring 2025)/第一次作业提交名单|第一次作业提交名单]]&lt;br /&gt;
&lt;br /&gt;
*[[概率论与数理统计 (Spring 2025)/Problem Set 2|Problem Set 2]]  请在 2025/3/31 上课之前(10am UTC+8)提交到 [mailto:pr_nju@163.com pr_nju@163.com] (文件名为&#039;&amp;lt;font color=red &amp;gt;学号_姓名_A2.pdf&amp;lt;/font&amp;gt;&#039;).&lt;br /&gt;
** [[概率论与数理统计 (Spring 2025)/第二次作业提交名单|第二次作业提交名单]]&lt;br /&gt;
&lt;br /&gt;
*[[概率论与数理统计 (Spring 2025)/Problem Set 3|Problem Set 3]]  请在 2025/4/21 上课之前(10am UTC+8)提交到 [mailto:pr_nju@163.com pr_nju@163.com] (文件名为&#039;&amp;lt;font color=red &amp;gt;学号_姓名_A3.pdf&amp;lt;/font&amp;gt;&#039;).&lt;br /&gt;
** [[概率论与数理统计 (Spring 2025)/第三次作业提交名单|第三次作业提交名单]]&lt;br /&gt;
&lt;br /&gt;
= Lectures =&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2025/Intro.pdf 课程简介]&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2025/ProbSpace.pdf 概率空间]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 第1章&#039;&#039;&#039; 或 &#039;&#039;&#039;[GS] Chapter 1&#039;&#039;&#039;&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2025)/Entropy and volume of Hamming balls|Entropy and volume of Hamming balls]]&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2025/RandVar.pdf 随机变量]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 第2章&#039;&#039;&#039; 或 &#039;&#039;&#039;[GS] Chapter 2, Sections 3.1~3.5, 3.7&#039;&#039;&#039;&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[MU] Chapter 2&#039;&#039;&#039;&lt;br /&gt;
#* the [http://tcs.nju.edu.cn/slides/prob2024/discrete-pmf.nb &#039;&#039;&#039;Mathematica notebook file&#039;&#039;&#039;] for the PMFs of basic discrete distributions&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2025)/Average-case analysis of QuickSort|Average-case analysis of &#039;&#039;&#039;&#039;&#039;QuickSort&#039;&#039;&#039;&#039;&#039;]]&lt;br /&gt;
#* [https://www.bilibili.com/video/BV1ta411A7fp/ 高尔顿板（Galton board）视频] 和 [https://en.wikipedia.org/wiki/Galton_board 维基百科页面]&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2025/Deviation.pdf 矩与偏差]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[MU] Chapter 3&#039;&#039;&#039;&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 章节 2.4, 4.2, 4.3, 5.1&#039;&#039;&#039; 或 &#039;&#039;&#039;[GS] Sections 3.3, 3.6, 7.3&#039;&#039;&#039;&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2025)/Two-point sampling|Two-point sampling]]&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2025)/Threshold of k-clique in random graph|Threshold of &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;-clique in random graph]]&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2025)/Weierstrass Approximation Theorem|Weierstrass approximation]]&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2025/Continuous.pdf 连续分布]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 第3章, 和4.1节&#039;&#039;&#039; 或 &#039;&#039;&#039;[GS] Chapter 4&#039;&#039;&#039;&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[MU] Chapters 8, 9&#039;&#039;&#039;&lt;br /&gt;
#* [https://measure.axler.net/MIRA.pdf Measure, Integration &amp;amp; Real Analysis] by Sheldon Axler&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2025)/An exercise on induced distribution|An exercise on induced distribution]]&lt;br /&gt;
&lt;br /&gt;
= Concepts =&lt;br /&gt;
* [https://plato.stanford.edu/entries/probability-interpret/ Interpretations of probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/History_of_probability History of probability]&lt;br /&gt;
* Example problems:&lt;br /&gt;
** [https://dornsifecms.usc.edu/assets/sites/520/docs/VonNeumann-ams12p36-38.pdf von Neumann&#039;s Bernoulli factory] and other [https://peteroupc.github.io/bernoulli.html Bernoulli factory algorithms]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Boy_or_Girl_paradox Boy or Girl paradox]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Monty_Hall_problem Monty Hall problem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bertrand_paradox_(probability) Bertrand paradox]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hard_spheres Hard spheres model] and [https://en.wikipedia.org/wiki/Ising_model Ising model]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/PageRank &#039;&#039;PageRank&#039;&#039;] and stationary [https://en.wikipedia.org/wiki/Random_walk random walk]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Diffusion_process Diffusion process] and [https://en.wikipedia.org/wiki/Diffusion_model diffusion model]&lt;br /&gt;
*[https://en.wikipedia.org/wiki/Probability_space Probability space]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Sample_space Sample space]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Event_(probability_theory) Event] and [https://en.wikipedia.org/wiki/Σ-algebra &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-algebra]&lt;br /&gt;
** Kolmogorov&#039;s [https://en.wikipedia.org/wiki/Probability_axioms axioms of probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Discrete_uniform_distribution Classical] and [https://en.wikipedia.org/wiki/Geometric_probability goemetric probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Boole%27s_inequality Union bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Inclusion%E2%80%93exclusion_principle Inclusion-Exclusion principle]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Boole%27s_inequality#Bonferroni_inequalities Bonferroni inequalities]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Conditional_probability Conditional probability]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chain_rule_(probability) Chain rule]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_total_probability Law of total probability]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bayes%27_theorem Bayes&#039; law]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Independence_(probability_theory) Independence] &lt;br /&gt;
** [https://en.wikipedia.org/wiki/Pairwise_independence Pairwise independence]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Random_variable Random variable]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cumulative_distribution_function Cumulative distribution function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_mass_function Probability mass function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_density_function Probability density function]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Multivariate_random_variable Random vector]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Joint_probability_distribution Joint probability distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Conditional_probability_distribution Conditional probability distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Marginal_distribution Marginal distribution]&lt;br /&gt;
* Some &#039;&#039;&#039;discrete&#039;&#039;&#039; probability distributions&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bernoulli_trial Bernoulli trial] and [https://en.wikipedia.org/wiki/Bernoulli_distribution Bernoulli distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Discrete_uniform_distribution Discrete uniform distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Binomial_distribution Binomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Geometric_distribution Geometric distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Negative_binomial_distribution Negative binomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hypergeometric_distribution Hypergeometric distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Poisson_distribution Poisson distribution]&lt;br /&gt;
** and [https://en.wikipedia.org/wiki/List_of_probability_distributions#Discrete_distributions others]&lt;br /&gt;
* Balls into bins model&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Multinomial_distribution Multinomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Birthday_problem Birthday problem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Coupon_collector%27s_problem Coupon collector]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Balls_into_bins_problem Occupancy problem]&lt;br /&gt;
* Random graphs&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Erd%C5%91s%E2%80%93R%C3%A9nyi_model Erdős–Rényi random graph model]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Galton%E2%80%93Watson_process Galton–Watson branching process]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Expected_value Expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_the_unconscious_statistician Law of the unconscious statistician, &#039;&#039;LOTUS&#039;&#039;]&lt;br /&gt;
** [https://dlsun.github.io/probability/linearity.html Linearity of expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Conditional_expectation Conditional expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_total_expectation Law of total expectation]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Markov%27s_inequality Markov&#039;s inequality]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Chebyshev%27s_inequality Chebyshev&#039;s inequality]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Moment_(mathematics) Moment] related&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Central_moment Central moment]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Variance Variance]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Standard_deviation Standard deviation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Covariance Covariance]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Correlation Correlation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Correlation_does_not_imply_causation Correlation does not imply causation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Skewness Skewness]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Kurtosis Kurtosis]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Moment_problem Moment problem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Measure_(mathematics) Measure]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Borel_set Borel set]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Lebesgue_integration Lebesgue integration]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Measurable_function Measurable function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cantor_set Cantor set]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Non-measurable_set Non-measurable set]&lt;br /&gt;
*[https://en.wikipedia.org/wiki/Probability_density_function Probability density function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_distribution#Absolutely_continuous_probability_distribution Continuous probability distribution]&lt;br /&gt;
**[https://en.wikipedia.org/wiki/Conditional_probability_distribution#Conditional_continuous_distributions Conditional continuous distributions]&lt;br /&gt;
**[https://en.wikipedia.org/wiki/Convolution_of_probability_distributions Convolution of probability distributions] and [https://en.wikipedia.org/wiki/List_of_convolutions_of_probability_distributions List of convolutions of probability distributions]&lt;br /&gt;
* Some &#039;&#039;&#039;continuous&#039;&#039;&#039; probability distributions&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Continuous_uniform_distribution Continuous uniform distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Exponential_distribution Exponential distribution] and [https://en.wikipedia.org/wiki/Poisson_point_process Poisson point process]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Normal_distribution Normal (Gaussion) distribution]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Gaussian_function Gaussian function] and [https://en.wikipedia.org/wiki/Gaussian_integral Gaussian integral]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Standard_normal_table Standard normal table]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/68%E2%80%9395%E2%80%9399.7_rule 68–95–99.7 rule]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Sum_of_normally_distributed_random_variables Sum of normally distributed random variables]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Multivariate_normal_distribution Multivariate normal distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chi-squared_distribution Chi-squared distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Gamma_distribution Gamma distribution] and [https://en.wikipedia.org/wiki/Gamma_function Gamma function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Beta_distribution Beta distribution] and [https://en.wikipedia.org/wiki/Beta_function Beta function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cauchy_distribution Cauchy distribution]&lt;br /&gt;
** and [https://en.wikipedia.org/wiki/List_of_probability_distributions#Absolutely_continuous_distributions others]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Inverse_transform_sampling Inverse transform sampling]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stochastic_dominance Stochastic dominance] and [https://en.wikipedia.org/wiki/Coupling_(probability) Coupling] &lt;br /&gt;
* [https://en.wikipedia.org/wiki/Moment-generating_function Moment-generating function]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Convergence_of_random_variables Convergence of random variables]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Pointwise_convergence Pointwise convergence]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Convergence_of_measures#Weak_convergence_of_measures Weak convergence of measures]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Convergence_in_measure Convergence in measure]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Skorokhod%27s_representation_theorem Skorokhod&#039;s representation theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Continuous_mapping_theorem Continuous mapping theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Borel%E2%80%93Cantelli_lemma Borel–Cantelli lemma] and [https://en.wikipedia.org/wiki/Zero%E2%80%93one_law zero-one laws]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Law_of_large_numbers Law of large numbers]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Central_limit_theorem Central limit theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/De_Moivre%E2%80%93Laplace_theorem De Moivre–Laplace theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Berry%E2%80%93Esseen_theorem Berry–Esseen theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Characteristic_function_(probability_theory) Characteristic function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Lévy%27s_continuity_theorem Lévy&#039;s continuity theorem]&lt;br /&gt;
* Concentration of measures:&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Moment-generating_function Moment-generating function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chernoff_bound Chernoff bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hoeffding%27s_inequality Hoeffding&#039;s bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Doob_martingale#McDiarmid&#039;s_inequality McDiarmid&#039;s inequality]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Doob_martingale Doob martingale] and [https://en.wikipedia.org/wiki/Azuma%27s_inequality Azuma&#039;s inequality]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Sub-Gaussian_distribution Sub-Gaussian distribution]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stochastic_process Stochastic process]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stopping_time Stopping time]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Martingale_(probability_theory) Martingale]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Optional_stopping_theorem Optional stopping theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Wald%27s_equation Wald&#039;s equation]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Markov_chain Markov chain]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Markov_property Markov property] and [https://en.wikipedia.org/wiki/Memorylessness memorylessness]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Stochastic_matrix Stochastic matrix]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Stationary_distribution Stationary distribution]&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_3&amp;diff=13043</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 3</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_3&amp;diff=13043"/>
		<updated>2025-04-05T04:12:00Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: /* Problem 1 (Warm-up Problems) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 3==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Warm-up Problems) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
      Let &amp;lt;math&amp;gt;X_1,X_2,...,X_n&amp;lt;/math&amp;gt; be independent random variables, and suppose that &amp;lt;math&amp;gt;X_k&amp;lt;/math&amp;gt; is Bernoulli with parameter &amp;lt;math&amp;gt;p_k&amp;lt;/math&amp;gt;. Let &amp;lt;math&amp;gt;Y= X_1 + X_2 + \dots + X_n&amp;lt;/math&amp;gt;. Show that, for &amp;lt;math&amp;gt;\mathbb E[Y]&amp;lt;/math&amp;gt; fixed, &amp;lt;math&amp;gt;\mathrm{Var}(Y )&amp;lt;/math&amp;gt; is a maximized when &amp;lt;math&amp;gt;p_1 = p_2 = \dots = p_n&amp;lt;/math&amp;gt;. That is to say, the variation in the sum is greatest when individuals are most alike.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Each member of a group of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; players rolls a (fair) 6-sided die. For any pair of players who throw the same number, the group scores &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; point. Find the mean and variance of the total score of the group.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        An urn contains &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; balls numbered &amp;lt;math&amp;gt;1, 2, \ldots, n&amp;lt;/math&amp;gt;. We select &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; balls uniformly at random &amp;lt;strong&amp;gt;without replacement&amp;lt;/strong&amp;gt; and add up their numbers. Find the mean and variance of the sum.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (IV)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
      Let &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; be an integer-valued, positive random variable and let &amp;lt;math&amp;gt;\{X_i\}_{i=1}^{\infty}&amp;lt;/math&amp;gt; be indepedently identically distributed random variables that are independent of &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt;, too. &lt;br /&gt;
Precisely, for any finite subset &amp;lt;math&amp;gt;I \subseteq\mathbb{N}_+&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\{X_i\}_{i \in I}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; are mutually independent. Let &amp;lt;math&amp;gt;X = \sum_{i=1}^N X_i&amp;lt;/math&amp;gt;, show that &amp;lt;math&amp;gt;\textbf{Var}[X] = \textbf{Var}[X_1] \mathbb{E}[N] + \mathbb{E}[X_1]^2 \textbf{Var}[N]&amp;lt;/math&amp;gt;.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Moments (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Show that [math]G(t) = \frac{e^t}{4} + \frac{e^{-t}}{2} + \frac{1}{4}[/math] is a moment-generating function of a random variable, and write the probability mass function of this random variable.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Moments (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X\sim \text{Geo}(p)&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt;. Find &amp;lt;math&amp;gt;\mathbb{E}[X^3]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbb{E}[X^4]&amp;lt;/math&amp;gt;. &lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Moments (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X\sim \text{Pois}(\lambda)&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;\lambda &amp;gt;0 &amp;lt;/math&amp;gt;. Find &amp;lt;math&amp;gt;\mathbb{E}[X^3]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbb{E}[X^4]&amp;lt;/math&amp;gt;. &lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be discrete random variables with correlation &amp;lt;math&amp;gt;\rho&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;|\rho|= 1&amp;lt;/math&amp;gt; if and only if &amp;lt;math&amp;gt;X=aY+b&amp;lt;/math&amp;gt; for some real numbers &amp;lt;math&amp;gt;a,b&amp;lt;/math&amp;gt;.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
    Let [math]X[/math] and [math]Y[/math] be discrete random variables with mean &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;, variance &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;, and correlation [math]\rho[/math]. Show that [math]\mathbb{E}(\max\{X^2,Y^2\})\leq 1+\sqrt{1-\rho^2}[/math]. (Hint: use the identity [math]\max\{a,b\} = \frac{1}{2}(a+b+|a-b|)[/math].)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
   Let [math]X[/math] and [math]Y[/math] be independent Bernoulli random variables with parameter [math]1/2[/math]. Show that [math]X+Y[/math] and [math]|X-Y|[/math] are dependent though uncorrelated.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Inequalities) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Reverse Markov&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable with bounded range &amp;lt;math&amp;gt;0 \le X \le U&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;U &amp;gt; 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbf{Pr}(X \le a) \le \frac{U-\mathbf{E}[X]}{U-a}&amp;lt;/math&amp;gt; for any &amp;lt;math&amp;gt;0 &amp;lt; a &amp;lt; U&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Markov&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable. Show that for all &amp;lt;math&amp;gt;\beta \geq 0&amp;lt;/math&amp;gt; and all &amp;lt;math&amp;gt;x &amp;gt; 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}(X\geq x)\leq \mathbb{E}(e^{\beta X})e^{-\beta x}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Cantelli&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable with mean &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; and variance &amp;lt;math&amp;gt;\sigma^2&amp;lt;/math&amp;gt;. Prove that for any &amp;lt;math&amp;gt;\lambda &amp;gt; 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}[X \ge \lambda] \le \frac{\sigma^2}{\lambda^2+\sigma^2}&amp;lt;/math&amp;gt;. (Hint: You may first show that &amp;lt;math&amp;gt;\mathbf{Pr}[X \ge \lambda] \le \frac{\sigma^2 + u^2}{(\lambda + u)^2}&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;u &amp;gt; 0&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
&amp;lt;strong&amp;gt;[Chebyshev&#039;s inequality (I)]&amp;lt;/strong&amp;gt; Fix &amp;lt;math&amp;gt;0 &amp;lt; b \le a&amp;lt;/math&amp;gt;. Construct a random variable &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;\mathbb{E}[X^2] = b^2&amp;lt;/math&amp;gt; for which &amp;lt;math&amp;gt;\mathbf{Pr}(|X| \ge a) = b^2/a^2&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
&amp;lt;strong&amp;gt;[Chebyshev&#039;s inequality (II)]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a random variable with &amp;lt;math&amp;gt;0 &amp;lt; \mathbb{E}[X^2] &amp;lt; \infty&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\lim_{a \to \infty} \frac{a^2 \mathbf{Pr}(|X| \ge a)}{ \mathbb{E}[X^2] } = 0&amp;lt;/math&amp;gt;. (Hint: Use the dominated convergence theorem. For discrete random variables, it can be formulated as follows: Let &amp;lt;math&amp;gt;Z,X, X_1,X_2,\ldots,X_n,\ldots&amp;lt;/math&amp;gt; be discrete random variables with finite second moments. If &amp;lt;math&amp;gt;|X_n| \le Z&amp;lt;/math&amp;gt; and for any &amp;lt;math&amp;gt;a \in \mathbb{R}&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}(X_n = a) \to \mathbf{Pr}(X = a)&amp;lt;/math&amp;gt; when &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity, then &amp;lt;math&amp;gt;\mathbb{E}[X_n^2] \to \mathbb{E}[X^2]&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Probability meets graph theory) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;    &lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;4-clique threshold&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Prove that &amp;lt;math&amp;gt;p = n^{-2/3}&amp;lt;/math&amp;gt; is the threshold probability for the existence of 4-clique. &lt;br /&gt;
        Formally, you are required to show that &lt;br /&gt;
        &amp;lt;ol type=&amp;quot;a&amp;quot;&amp;gt;&lt;br /&gt;
            &amp;lt;li&amp;gt;&lt;br /&gt;
                with a probability approaching to &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; (as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity) the Erdős–Rényi random graph &amp;lt;math&amp;gt;\mathbf{G} = \mathbf{G}(n,p)&amp;lt;/math&amp;gt; contains a 4-clique when &amp;lt;math&amp;gt;p = \omega(n^{-2/3})&amp;lt;/math&amp;gt;; (Hint: use Chebyshev&#039;s inequality.)&lt;br /&gt;
            &amp;lt;/li&amp;gt;&lt;br /&gt;
            &amp;lt;li&amp;gt;&lt;br /&gt;
                with a probability approaching to &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; (as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity) the Erdős–Rényi random graph &amp;lt;math&amp;gt;\mathbf{G} = \mathbf{G}(n,p)&amp;lt;/math&amp;gt; contains a 4-clique when &amp;lt;math&amp;gt;p = o(n^{-2/3})&amp;lt;/math&amp;gt;. (Hint: use Markov&#039;s inequality.)&lt;br /&gt;
            &amp;lt;/li&amp;gt;&lt;br /&gt;
        &amp;lt;/ol&amp;gt;&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_3&amp;diff=13042</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 3</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_3&amp;diff=13042"/>
		<updated>2025-04-05T04:10:10Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: /* Problem 1 (Warm-up Problems) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 3==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Warm-up Problems) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
      Let &amp;lt;math&amp;gt;X_1,X_2,...,X_n&amp;lt;/math&amp;gt; be independent random variables, and suppose that &amp;lt;math&amp;gt;X_k&amp;lt;/math&amp;gt; is Bernoulli with parameter &amp;lt;math&amp;gt;p_k&amp;lt;/math&amp;gt;. Show that, for &amp;lt;math&amp;gt;\mathbb E[Y]&amp;lt;/math&amp;gt; fixed, &amp;lt;math&amp;gt;\mathrm{Var}(Y )&amp;lt;/math&amp;gt; is a maximized when &amp;lt;math&amp;gt;p_1 = p_2 = \dots = p_n&amp;lt;/math&amp;gt;. That is to say, the variation in the sum is greatest when individuals are most alike.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Each member of a group of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; players rolls a (fair) 6-sided die. For any pair of players who throw the same number, the group scores &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; point. Find the mean and variance of the total score of the group.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        An urn contains &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; balls numbered &amp;lt;math&amp;gt;1, 2, \ldots, n&amp;lt;/math&amp;gt;. We select &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; balls uniformly at random &amp;lt;strong&amp;gt;without replacement&amp;lt;/strong&amp;gt; and add up their numbers. Find the mean and variance of the sum.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (IV)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
      Let &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; be an integer-valued, positive random variable and let &amp;lt;math&amp;gt;\{X_i\}_{i=1}^{\infty}&amp;lt;/math&amp;gt; be indepedently identically distributed random variables that are independent of &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt;, too. &lt;br /&gt;
Precisely, for any finite subset &amp;lt;math&amp;gt;I \subseteq\mathbb{N}_+&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\{X_i\}_{i \in I}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; are mutually independent. Let &amp;lt;math&amp;gt;X = \sum_{i=1}^N X_i&amp;lt;/math&amp;gt;, show that &amp;lt;math&amp;gt;\textbf{Var}[X] = \textbf{Var}[X_1] \mathbb{E}[N] + \mathbb{E}[X_1]^2 \textbf{Var}[N]&amp;lt;/math&amp;gt;.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Moments (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Show that [math]G(t) = \frac{e^t}{4} + \frac{e^{-t}}{2} + \frac{1}{4}[/math] is a moment-generating function of a random variable, and write the probability mass function of this random variable.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Moments (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X\sim \text{Geo}(p)&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt;. Find &amp;lt;math&amp;gt;\mathbb{E}[X^3]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbb{E}[X^4]&amp;lt;/math&amp;gt;. &lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Moments (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X\sim \text{Pois}(\lambda)&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;\lambda &amp;gt;0 &amp;lt;/math&amp;gt;. Find &amp;lt;math&amp;gt;\mathbb{E}[X^3]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbb{E}[X^4]&amp;lt;/math&amp;gt;. &lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be discrete random variables with correlation &amp;lt;math&amp;gt;\rho&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;|\rho|= 1&amp;lt;/math&amp;gt; if and only if &amp;lt;math&amp;gt;X=aY+b&amp;lt;/math&amp;gt; for some real numbers &amp;lt;math&amp;gt;a,b&amp;lt;/math&amp;gt;.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
    Let [math]X[/math] and [math]Y[/math] be discrete random variables with mean &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;, variance &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;, and correlation [math]\rho[/math]. Show that [math]\mathbb{E}(\max\{X^2,Y^2\})\leq 1+\sqrt{1-\rho^2}[/math]. (Hint: use the identity [math]\max\{a,b\} = \frac{1}{2}(a+b+|a-b|)[/math].)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
   Let [math]X[/math] and [math]Y[/math] be independent Bernoulli random variables with parameter [math]1/2[/math]. Show that [math]X+Y[/math] and [math]|X-Y|[/math] are dependent though uncorrelated.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Inequalities) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Reverse Markov&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable with bounded range &amp;lt;math&amp;gt;0 \le X \le U&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;U &amp;gt; 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbf{Pr}(X \le a) \le \frac{U-\mathbf{E}[X]}{U-a}&amp;lt;/math&amp;gt; for any &amp;lt;math&amp;gt;0 &amp;lt; a &amp;lt; U&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Markov&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable. Show that for all &amp;lt;math&amp;gt;\beta \geq 0&amp;lt;/math&amp;gt; and all &amp;lt;math&amp;gt;x &amp;gt; 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}(X\geq x)\leq \mathbb{E}(e^{\beta X})e^{-\beta x}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Cantelli&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable with mean &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; and variance &amp;lt;math&amp;gt;\sigma^2&amp;lt;/math&amp;gt;. Prove that for any &amp;lt;math&amp;gt;\lambda &amp;gt; 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}[X \ge \lambda] \le \frac{\sigma^2}{\lambda^2+\sigma^2}&amp;lt;/math&amp;gt;. (Hint: You may first show that &amp;lt;math&amp;gt;\mathbf{Pr}[X \ge \lambda] \le \frac{\sigma^2 + u^2}{(\lambda + u)^2}&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;u &amp;gt; 0&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
&amp;lt;strong&amp;gt;[Chebyshev&#039;s inequality (I)]&amp;lt;/strong&amp;gt; Fix &amp;lt;math&amp;gt;0 &amp;lt; b \le a&amp;lt;/math&amp;gt;. Construct a random variable &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;\mathbb{E}[X^2] = b^2&amp;lt;/math&amp;gt; for which &amp;lt;math&amp;gt;\mathbf{Pr}(|X| \ge a) = b^2/a^2&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
&amp;lt;strong&amp;gt;[Chebyshev&#039;s inequality (II)]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a random variable with &amp;lt;math&amp;gt;0 &amp;lt; \mathbb{E}[X^2] &amp;lt; \infty&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\lim_{a \to \infty} \frac{a^2 \mathbf{Pr}(|X| \ge a)}{ \mathbb{E}[X^2] } = 0&amp;lt;/math&amp;gt;. (Hint: Use the dominated convergence theorem. For discrete random variables, it can be formulated as follows: Let &amp;lt;math&amp;gt;Z,X, X_1,X_2,\ldots,X_n,\ldots&amp;lt;/math&amp;gt; be discrete random variables with finite second moments. If &amp;lt;math&amp;gt;|X_n| \le Z&amp;lt;/math&amp;gt; and for any &amp;lt;math&amp;gt;a \in \mathbb{R}&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}(X_n = a) \to \mathbf{Pr}(X = a)&amp;lt;/math&amp;gt; when &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity, then &amp;lt;math&amp;gt;\mathbb{E}[X_n^2] \to \mathbb{E}[X^2]&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Probability meets graph theory) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;    &lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;4-clique threshold&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Prove that &amp;lt;math&amp;gt;p = n^{-2/3}&amp;lt;/math&amp;gt; is the threshold probability for the existence of 4-clique. &lt;br /&gt;
        Formally, you are required to show that &lt;br /&gt;
        &amp;lt;ol type=&amp;quot;a&amp;quot;&amp;gt;&lt;br /&gt;
            &amp;lt;li&amp;gt;&lt;br /&gt;
                with a probability approaching to &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; (as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity) the Erdős–Rényi random graph &amp;lt;math&amp;gt;\mathbf{G} = \mathbf{G}(n,p)&amp;lt;/math&amp;gt; contains a 4-clique when &amp;lt;math&amp;gt;p = \omega(n^{-2/3})&amp;lt;/math&amp;gt;; (Hint: use Chebyshev&#039;s inequality.)&lt;br /&gt;
            &amp;lt;/li&amp;gt;&lt;br /&gt;
            &amp;lt;li&amp;gt;&lt;br /&gt;
                with a probability approaching to &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; (as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity) the Erdős–Rényi random graph &amp;lt;math&amp;gt;\mathbf{G} = \mathbf{G}(n,p)&amp;lt;/math&amp;gt; contains a 4-clique when &amp;lt;math&amp;gt;p = o(n^{-2/3})&amp;lt;/math&amp;gt;. (Hint: use Markov&#039;s inequality.)&lt;br /&gt;
            &amp;lt;/li&amp;gt;&lt;br /&gt;
        &amp;lt;/ol&amp;gt;&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_3&amp;diff=13041</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 3</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_3&amp;diff=13041"/>
		<updated>2025-04-05T04:08:56Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: /* Problem 1 (Warm-up Problems) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 3==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Warm-up Problems) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
      Let &amp;lt;math&amp;gt;X_1,X_2,...,X_n&amp;lt;/math&amp;gt; be independent random variables, and suppose that &amp;lt;math&amp;gt;X_k&amp;lt;/math&amp;gt; is Bernoulli with parameter &amp;lt;math&amp;gt;p_k&amp;lt;/math&amp;gt;. Show that, for &amp;lt;math&amp;gt;\mathbb E[Y]&amp;lt;/math&amp;gt; fixed, &amp;lt;math&amp;gt;\mathrm{Var}(Y )&amp;lt;/math&amp;gt; is a maximum when &amp;lt;math&amp;gt;p_1 = p_2 = \dots = p_n&amp;lt;/math&amp;gt;. That is to say, the variation in the sum is greatest when individuals are most alike.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Each member of a group of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; players rolls a (fair) 6-sided die. For any pair of players who throw the same number, the group scores &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; point. Find the mean and variance of the total score of the group.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        An urn contains &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; balls numbered &amp;lt;math&amp;gt;1, 2, \ldots, n&amp;lt;/math&amp;gt;. We select &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; balls uniformly at random &amp;lt;strong&amp;gt;without replacement&amp;lt;/strong&amp;gt; and add up their numbers. Find the mean and variance of the sum.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (IV)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
      Let &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; be an integer-valued, positive random variable and let &amp;lt;math&amp;gt;\{X_i\}_{i=1}^{\infty}&amp;lt;/math&amp;gt; be indepedently identically distributed random variables that are independent of &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt;, too. &lt;br /&gt;
Precisely, for any finite subset &amp;lt;math&amp;gt;I \subseteq\mathbb{N}_+&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\{X_i\}_{i \in I}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; are mutually independent. Let &amp;lt;math&amp;gt;X = \sum_{i=1}^N X_i&amp;lt;/math&amp;gt;, show that &amp;lt;math&amp;gt;\textbf{Var}[X] = \textbf{Var}[X_1] \mathbb{E}[N] + \mathbb{E}[X_1]^2 \textbf{Var}[N]&amp;lt;/math&amp;gt;.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Moments (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Show that [math]G(t) = \frac{e^t}{4} + \frac{e^{-t}}{2} + \frac{1}{4}[/math] is a moment-generating function of a random variable, and write the probability mass function of this random variable.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Moments (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X\sim \text{Geo}(p)&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt;. Find &amp;lt;math&amp;gt;\mathbb{E}[X^3]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbb{E}[X^4]&amp;lt;/math&amp;gt;. &lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Moments (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X\sim \text{Pois}(\lambda)&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;\lambda &amp;gt;0 &amp;lt;/math&amp;gt;. Find &amp;lt;math&amp;gt;\mathbb{E}[X^3]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbb{E}[X^4]&amp;lt;/math&amp;gt;. &lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be discrete random variables with correlation &amp;lt;math&amp;gt;\rho&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;|\rho|= 1&amp;lt;/math&amp;gt; if and only if &amp;lt;math&amp;gt;X=aY+b&amp;lt;/math&amp;gt; for some real numbers &amp;lt;math&amp;gt;a,b&amp;lt;/math&amp;gt;.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
    Let [math]X[/math] and [math]Y[/math] be discrete random variables with mean &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;, variance &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;, and correlation [math]\rho[/math]. Show that [math]\mathbb{E}(\max\{X^2,Y^2\})\leq 1+\sqrt{1-\rho^2}[/math]. (Hint: use the identity [math]\max\{a,b\} = \frac{1}{2}(a+b+|a-b|)[/math].)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
   Let [math]X[/math] and [math]Y[/math] be independent Bernoulli random variables with parameter [math]1/2[/math]. Show that [math]X+Y[/math] and [math]|X-Y|[/math] are dependent though uncorrelated.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Inequalities) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Reverse Markov&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable with bounded range &amp;lt;math&amp;gt;0 \le X \le U&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;U &amp;gt; 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbf{Pr}(X \le a) \le \frac{U-\mathbf{E}[X]}{U-a}&amp;lt;/math&amp;gt; for any &amp;lt;math&amp;gt;0 &amp;lt; a &amp;lt; U&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Markov&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable. Show that for all &amp;lt;math&amp;gt;\beta \geq 0&amp;lt;/math&amp;gt; and all &amp;lt;math&amp;gt;x &amp;gt; 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}(X\geq x)\leq \mathbb{E}(e^{\beta X})e^{-\beta x}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Cantelli&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable with mean &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; and variance &amp;lt;math&amp;gt;\sigma^2&amp;lt;/math&amp;gt;. Prove that for any &amp;lt;math&amp;gt;\lambda &amp;gt; 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}[X \ge \lambda] \le \frac{\sigma^2}{\lambda^2+\sigma^2}&amp;lt;/math&amp;gt;. (Hint: You may first show that &amp;lt;math&amp;gt;\mathbf{Pr}[X \ge \lambda] \le \frac{\sigma^2 + u^2}{(\lambda + u)^2}&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;u &amp;gt; 0&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
&amp;lt;strong&amp;gt;[Chebyshev&#039;s inequality (I)]&amp;lt;/strong&amp;gt; Fix &amp;lt;math&amp;gt;0 &amp;lt; b \le a&amp;lt;/math&amp;gt;. Construct a random variable &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;\mathbb{E}[X^2] = b^2&amp;lt;/math&amp;gt; for which &amp;lt;math&amp;gt;\mathbf{Pr}(|X| \ge a) = b^2/a^2&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
&amp;lt;strong&amp;gt;[Chebyshev&#039;s inequality (II)]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a random variable with &amp;lt;math&amp;gt;0 &amp;lt; \mathbb{E}[X^2] &amp;lt; \infty&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\lim_{a \to \infty} \frac{a^2 \mathbf{Pr}(|X| \ge a)}{ \mathbb{E}[X^2] } = 0&amp;lt;/math&amp;gt;. (Hint: Use the dominated convergence theorem. For discrete random variables, it can be formulated as follows: Let &amp;lt;math&amp;gt;Z,X, X_1,X_2,\ldots,X_n,\ldots&amp;lt;/math&amp;gt; be discrete random variables with finite second moments. If &amp;lt;math&amp;gt;|X_n| \le Z&amp;lt;/math&amp;gt; and for any &amp;lt;math&amp;gt;a \in \mathbb{R}&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}(X_n = a) \to \mathbf{Pr}(X = a)&amp;lt;/math&amp;gt; when &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity, then &amp;lt;math&amp;gt;\mathbb{E}[X_n^2] \to \mathbb{E}[X^2]&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Probability meets graph theory) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;    &lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;4-clique threshold&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Prove that &amp;lt;math&amp;gt;p = n^{-2/3}&amp;lt;/math&amp;gt; is the threshold probability for the existence of 4-clique. &lt;br /&gt;
        Formally, you are required to show that &lt;br /&gt;
        &amp;lt;ol type=&amp;quot;a&amp;quot;&amp;gt;&lt;br /&gt;
            &amp;lt;li&amp;gt;&lt;br /&gt;
                with a probability approaching to &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; (as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity) the Erdős–Rényi random graph &amp;lt;math&amp;gt;\mathbf{G} = \mathbf{G}(n,p)&amp;lt;/math&amp;gt; contains a 4-clique when &amp;lt;math&amp;gt;p = \omega(n^{-2/3})&amp;lt;/math&amp;gt;; (Hint: use Chebyshev&#039;s inequality.)&lt;br /&gt;
            &amp;lt;/li&amp;gt;&lt;br /&gt;
            &amp;lt;li&amp;gt;&lt;br /&gt;
                with a probability approaching to &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; (as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity) the Erdős–Rényi random graph &amp;lt;math&amp;gt;\mathbf{G} = \mathbf{G}(n,p)&amp;lt;/math&amp;gt; contains a 4-clique when &amp;lt;math&amp;gt;p = o(n^{-2/3})&amp;lt;/math&amp;gt;. (Hint: use Markov&#039;s inequality.)&lt;br /&gt;
            &amp;lt;/li&amp;gt;&lt;br /&gt;
        &amp;lt;/ol&amp;gt;&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_3&amp;diff=13040</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 3</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_3&amp;diff=13040"/>
		<updated>2025-04-05T04:07:50Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: /* Problem 1 (Warm-up Problems) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 3==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Warm-up Problems) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
      Let &amp;lt;math&amp;gt;X_1,X_2,...,X_n&amp;lt;/math&amp;gt; be independent random variables, and suppose that &amp;lt;math&amp;gt;X_k&amp;lt;/math&amp;gt; is Bernoulli with parameter &amp;lt;math&amp;gt;p_k&amp;lt;/math&amp;gt; . Show that &amp;lt;math&amp;gt;Y= X_1 + X_2 + \dots + X_n&amp;lt;/math&amp;gt; has mean and variance given by &amp;lt;math&amp;gt;\mathbb E[Y ]= \sum^n_1 p_k , \mathrm{Var}(Y )= \sum^n_1 p_k (1− p_k )&amp;lt;/math&amp;gt;. Show that, for &amp;lt;math&amp;gt;\mathbb E[Y]&amp;lt;/math&amp;gt; fixed, &amp;lt;math&amp;gt;\mathrm{Var}(Y )&amp;lt;/math&amp;gt; is a maximum when &amp;lt;math&amp;gt;p_1 = p_2 = \dots = p_n&amp;lt;/math&amp;gt;. That is to say, the variation in the sum is greatest when individuals are most alike.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Each member of a group of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; players rolls a (fair) 6-sided die. For any pair of players who throw the same number, the group scores &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; point. Find the mean and variance of the total score of the group.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        An urn contains &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; balls numbered &amp;lt;math&amp;gt;1, 2, \ldots, n&amp;lt;/math&amp;gt;. We select &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; balls uniformly at random &amp;lt;strong&amp;gt;without replacement&amp;lt;/strong&amp;gt; and add up their numbers. Find the mean and variance of the sum.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (IV)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
      Let &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; be an integer-valued, positive random variable and let &amp;lt;math&amp;gt;\{X_i\}_{i=1}^{\infty}&amp;lt;/math&amp;gt; be indepedently identically distributed random variables that are independent of &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt;, too. &lt;br /&gt;
Precisely, for any finite subset &amp;lt;math&amp;gt;I \subseteq\mathbb{N}_+&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\{X_i\}_{i \in I}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; are mutually independent. Let &amp;lt;math&amp;gt;X = \sum_{i=1}^N X_i&amp;lt;/math&amp;gt;, show that &amp;lt;math&amp;gt;\textbf{Var}[X] = \textbf{Var}[X_1] \mathbb{E}[N] + \mathbb{E}[X_1]^2 \textbf{Var}[N]&amp;lt;/math&amp;gt;.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Moments (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Show that [math]G(t) = \frac{e^t}{4} + \frac{e^{-t}}{2} + \frac{1}{4}[/math] is a moment-generating function of a random variable, and write the probability mass function of this random variable.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Moments (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X\sim \text{Geo}(p)&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt;. Find &amp;lt;math&amp;gt;\mathbb{E}[X^3]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbb{E}[X^4]&amp;lt;/math&amp;gt;. &lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Moments (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X\sim \text{Pois}(\lambda)&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;\lambda &amp;gt;0 &amp;lt;/math&amp;gt;. Find &amp;lt;math&amp;gt;\mathbb{E}[X^3]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbb{E}[X^4]&amp;lt;/math&amp;gt;. &lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be discrete random variables with correlation &amp;lt;math&amp;gt;\rho&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;|\rho|= 1&amp;lt;/math&amp;gt; if and only if &amp;lt;math&amp;gt;X=aY+b&amp;lt;/math&amp;gt; for some real numbers &amp;lt;math&amp;gt;a,b&amp;lt;/math&amp;gt;.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
    Let [math]X[/math] and [math]Y[/math] be discrete random variables with mean &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;, variance &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;, and correlation [math]\rho[/math]. Show that [math]\mathbb{E}(\max\{X^2,Y^2\})\leq 1+\sqrt{1-\rho^2}[/math]. (Hint: use the identity [math]\max\{a,b\} = \frac{1}{2}(a+b+|a-b|)[/math].)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
   Let [math]X[/math] and [math]Y[/math] be independent Bernoulli random variables with parameter [math]1/2[/math]. Show that [math]X+Y[/math] and [math]|X-Y|[/math] are dependent though uncorrelated.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Inequalities) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Reverse Markov&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable with bounded range &amp;lt;math&amp;gt;0 \le X \le U&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;U &amp;gt; 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbf{Pr}(X \le a) \le \frac{U-\mathbf{E}[X]}{U-a}&amp;lt;/math&amp;gt; for any &amp;lt;math&amp;gt;0 &amp;lt; a &amp;lt; U&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Markov&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable. Show that for all &amp;lt;math&amp;gt;\beta \geq 0&amp;lt;/math&amp;gt; and all &amp;lt;math&amp;gt;x &amp;gt; 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}(X\geq x)\leq \mathbb{E}(e^{\beta X})e^{-\beta x}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Cantelli&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable with mean &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; and variance &amp;lt;math&amp;gt;\sigma^2&amp;lt;/math&amp;gt;. Prove that for any &amp;lt;math&amp;gt;\lambda &amp;gt; 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}[X \ge \lambda] \le \frac{\sigma^2}{\lambda^2+\sigma^2}&amp;lt;/math&amp;gt;. (Hint: You may first show that &amp;lt;math&amp;gt;\mathbf{Pr}[X \ge \lambda] \le \frac{\sigma^2 + u^2}{(\lambda + u)^2}&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;u &amp;gt; 0&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
&amp;lt;strong&amp;gt;[Chebyshev&#039;s inequality (I)]&amp;lt;/strong&amp;gt; Fix &amp;lt;math&amp;gt;0 &amp;lt; b \le a&amp;lt;/math&amp;gt;. Construct a random variable &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;\mathbb{E}[X^2] = b^2&amp;lt;/math&amp;gt; for which &amp;lt;math&amp;gt;\mathbf{Pr}(|X| \ge a) = b^2/a^2&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
&amp;lt;strong&amp;gt;[Chebyshev&#039;s inequality (II)]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a random variable with &amp;lt;math&amp;gt;0 &amp;lt; \mathbb{E}[X^2] &amp;lt; \infty&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\lim_{a \to \infty} \frac{a^2 \mathbf{Pr}(|X| \ge a)}{ \mathbb{E}[X^2] } = 0&amp;lt;/math&amp;gt;. (Hint: Use the dominated convergence theorem. For discrete random variables, it can be formulated as follows: Let &amp;lt;math&amp;gt;Z,X, X_1,X_2,\ldots,X_n,\ldots&amp;lt;/math&amp;gt; be discrete random variables with finite second moments. If &amp;lt;math&amp;gt;|X_n| \le Z&amp;lt;/math&amp;gt; and for any &amp;lt;math&amp;gt;a \in \mathbb{R}&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}(X_n = a) \to \mathbf{Pr}(X = a)&amp;lt;/math&amp;gt; when &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity, then &amp;lt;math&amp;gt;\mathbb{E}[X_n^2] \to \mathbb{E}[X^2]&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Probability meets graph theory) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;    &lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;4-clique threshold&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Prove that &amp;lt;math&amp;gt;p = n^{-2/3}&amp;lt;/math&amp;gt; is the threshold probability for the existence of 4-clique. &lt;br /&gt;
        Formally, you are required to show that &lt;br /&gt;
        &amp;lt;ol type=&amp;quot;a&amp;quot;&amp;gt;&lt;br /&gt;
            &amp;lt;li&amp;gt;&lt;br /&gt;
                with a probability approaching to &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; (as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity) the Erdős–Rényi random graph &amp;lt;math&amp;gt;\mathbf{G} = \mathbf{G}(n,p)&amp;lt;/math&amp;gt; contains a 4-clique when &amp;lt;math&amp;gt;p = \omega(n^{-2/3})&amp;lt;/math&amp;gt;; (Hint: use Chebyshev&#039;s inequality.)&lt;br /&gt;
            &amp;lt;/li&amp;gt;&lt;br /&gt;
            &amp;lt;li&amp;gt;&lt;br /&gt;
                with a probability approaching to &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; (as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity) the Erdős–Rényi random graph &amp;lt;math&amp;gt;\mathbf{G} = \mathbf{G}(n,p)&amp;lt;/math&amp;gt; contains a 4-clique when &amp;lt;math&amp;gt;p = o(n^{-2/3})&amp;lt;/math&amp;gt;. (Hint: use Markov&#039;s inequality.)&lt;br /&gt;
            &amp;lt;/li&amp;gt;&lt;br /&gt;
        &amp;lt;/ol&amp;gt;&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_3&amp;diff=13039</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 3</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_3&amp;diff=13039"/>
		<updated>2025-04-05T04:07:31Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: /* Problem 1 (Warm-up Problems) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 3==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Warm-up Problems) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
      Let &amp;lt;math&amp;gt;X_1,X_2,...,X_n&amp;lt;/math&amp;gt; be independent random variables, and suppose that &amp;lt;math&amp;gt;X_k&amp;lt;/math&amp;gt; is Bernoulli with parameter &amp;lt;math&amp;gt;p_k&amp;lt;/math&amp;gt; . Show that &amp;lt;math&amp;gt;Y= X_1 + X_2 + \dots + X_n&amp;lt;/math&amp;gt; has mean and variance given by &amp;lt;math&amp;gt;\mathbb E[Y ]= \sum^n_1 p_k , \mathrm{Var}(Y )= \sum^n_1 p_k (1− p_k )&amp;lt;/math&amp;gt;. Show that, for &amp;lt;math&amp;gt;\mathbb E[Y]&amp;lt;/math&amp;gt; fixed, &amp;lt;math&amp;gt;\mathrm{Var}(Y )&amp;lt;/math&amp;gt; is a maximum when &amp;lt;math&amp;gt;p_1 = p_2 = \dots = p_n&amp;lt;/math&amp;gt;. That is to say, the variation in the sum is greatest when individuals are most alike. Is this contrary to intuition?&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Each member of a group of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; players rolls a (fair) 6-sided die. For any pair of players who throw the same number, the group scores &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; point. Find the mean and variance of the total score of the group.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        An urn contains &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; balls numbered &amp;lt;math&amp;gt;1, 2, \ldots, n&amp;lt;/math&amp;gt;. We select &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; balls uniformly at random &amp;lt;strong&amp;gt;without replacement&amp;lt;/strong&amp;gt; and add up their numbers. Find the mean and variance of the sum.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (IV)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
      Let &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; be an integer-valued, positive random variable and let &amp;lt;math&amp;gt;\{X_i\}_{i=1}^{\infty}&amp;lt;/math&amp;gt; be indepedently identically distributed random variables that are independent of &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt;, too. &lt;br /&gt;
Precisely, for any finite subset &amp;lt;math&amp;gt;I \subseteq\mathbb{N}_+&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\{X_i\}_{i \in I}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; are mutually independent. Let &amp;lt;math&amp;gt;X = \sum_{i=1}^N X_i&amp;lt;/math&amp;gt;, show that &amp;lt;math&amp;gt;\textbf{Var}[X] = \textbf{Var}[X_1] \mathbb{E}[N] + \mathbb{E}[X_1]^2 \textbf{Var}[N]&amp;lt;/math&amp;gt;.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Moments (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Show that [math]G(t) = \frac{e^t}{4} + \frac{e^{-t}}{2} + \frac{1}{4}[/math] is a moment-generating function of a random variable, and write the probability mass function of this random variable.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Moments (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X\sim \text{Geo}(p)&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt;. Find &amp;lt;math&amp;gt;\mathbb{E}[X^3]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbb{E}[X^4]&amp;lt;/math&amp;gt;. &lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Moments (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X\sim \text{Pois}(\lambda)&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;\lambda &amp;gt;0 &amp;lt;/math&amp;gt;. Find &amp;lt;math&amp;gt;\mathbb{E}[X^3]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbb{E}[X^4]&amp;lt;/math&amp;gt;. &lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be discrete random variables with correlation &amp;lt;math&amp;gt;\rho&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;|\rho|= 1&amp;lt;/math&amp;gt; if and only if &amp;lt;math&amp;gt;X=aY+b&amp;lt;/math&amp;gt; for some real numbers &amp;lt;math&amp;gt;a,b&amp;lt;/math&amp;gt;.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
    Let [math]X[/math] and [math]Y[/math] be discrete random variables with mean &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;, variance &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;, and correlation [math]\rho[/math]. Show that [math]\mathbb{E}(\max\{X^2,Y^2\})\leq 1+\sqrt{1-\rho^2}[/math]. (Hint: use the identity [math]\max\{a,b\} = \frac{1}{2}(a+b+|a-b|)[/math].)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
   Let [math]X[/math] and [math]Y[/math] be independent Bernoulli random variables with parameter [math]1/2[/math]. Show that [math]X+Y[/math] and [math]|X-Y|[/math] are dependent though uncorrelated.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Inequalities) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Reverse Markov&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable with bounded range &amp;lt;math&amp;gt;0 \le X \le U&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;U &amp;gt; 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbf{Pr}(X \le a) \le \frac{U-\mathbf{E}[X]}{U-a}&amp;lt;/math&amp;gt; for any &amp;lt;math&amp;gt;0 &amp;lt; a &amp;lt; U&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Markov&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable. Show that for all &amp;lt;math&amp;gt;\beta \geq 0&amp;lt;/math&amp;gt; and all &amp;lt;math&amp;gt;x &amp;gt; 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}(X\geq x)\leq \mathbb{E}(e^{\beta X})e^{-\beta x}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Cantelli&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable with mean &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; and variance &amp;lt;math&amp;gt;\sigma^2&amp;lt;/math&amp;gt;. Prove that for any &amp;lt;math&amp;gt;\lambda &amp;gt; 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}[X \ge \lambda] \le \frac{\sigma^2}{\lambda^2+\sigma^2}&amp;lt;/math&amp;gt;. (Hint: You may first show that &amp;lt;math&amp;gt;\mathbf{Pr}[X \ge \lambda] \le \frac{\sigma^2 + u^2}{(\lambda + u)^2}&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;u &amp;gt; 0&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
&amp;lt;strong&amp;gt;[Chebyshev&#039;s inequality (I)]&amp;lt;/strong&amp;gt; Fix &amp;lt;math&amp;gt;0 &amp;lt; b \le a&amp;lt;/math&amp;gt;. Construct a random variable &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;\mathbb{E}[X^2] = b^2&amp;lt;/math&amp;gt; for which &amp;lt;math&amp;gt;\mathbf{Pr}(|X| \ge a) = b^2/a^2&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
&amp;lt;strong&amp;gt;[Chebyshev&#039;s inequality (II)]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a random variable with &amp;lt;math&amp;gt;0 &amp;lt; \mathbb{E}[X^2] &amp;lt; \infty&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\lim_{a \to \infty} \frac{a^2 \mathbf{Pr}(|X| \ge a)}{ \mathbb{E}[X^2] } = 0&amp;lt;/math&amp;gt;. (Hint: Use the dominated convergence theorem. For discrete random variables, it can be formulated as follows: Let &amp;lt;math&amp;gt;Z,X, X_1,X_2,\ldots,X_n,\ldots&amp;lt;/math&amp;gt; be discrete random variables with finite second moments. If &amp;lt;math&amp;gt;|X_n| \le Z&amp;lt;/math&amp;gt; and for any &amp;lt;math&amp;gt;a \in \mathbb{R}&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}(X_n = a) \to \mathbf{Pr}(X = a)&amp;lt;/math&amp;gt; when &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity, then &amp;lt;math&amp;gt;\mathbb{E}[X_n^2] \to \mathbb{E}[X^2]&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Probability meets graph theory) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;    &lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;4-clique threshold&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Prove that &amp;lt;math&amp;gt;p = n^{-2/3}&amp;lt;/math&amp;gt; is the threshold probability for the existence of 4-clique. &lt;br /&gt;
        Formally, you are required to show that &lt;br /&gt;
        &amp;lt;ol type=&amp;quot;a&amp;quot;&amp;gt;&lt;br /&gt;
            &amp;lt;li&amp;gt;&lt;br /&gt;
                with a probability approaching to &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; (as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity) the Erdős–Rényi random graph &amp;lt;math&amp;gt;\mathbf{G} = \mathbf{G}(n,p)&amp;lt;/math&amp;gt; contains a 4-clique when &amp;lt;math&amp;gt;p = \omega(n^{-2/3})&amp;lt;/math&amp;gt;; (Hint: use Chebyshev&#039;s inequality.)&lt;br /&gt;
            &amp;lt;/li&amp;gt;&lt;br /&gt;
            &amp;lt;li&amp;gt;&lt;br /&gt;
                with a probability approaching to &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; (as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity) the Erdős–Rényi random graph &amp;lt;math&amp;gt;\mathbf{G} = \mathbf{G}(n,p)&amp;lt;/math&amp;gt; contains a 4-clique when &amp;lt;math&amp;gt;p = o(n^{-2/3})&amp;lt;/math&amp;gt;. (Hint: use Markov&#039;s inequality.)&lt;br /&gt;
            &amp;lt;/li&amp;gt;&lt;br /&gt;
        &amp;lt;/ol&amp;gt;&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_3&amp;diff=13038</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 3</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_3&amp;diff=13038"/>
		<updated>2025-04-05T03:58:21Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: /* Problem 1 (Warm-up Problems) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 3==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Warm-up Problems) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X_1,X_2,\cdots, X_n&amp;lt;/math&amp;gt; be pairwise independent random variables. Show that &amp;lt;math&amp;gt;\textbf{Var}\left[\sum_{i=1}^n X_i\right] =\sum_{i=1}^n \textbf{Var} [X_i]&amp;lt;/math&amp;gt;.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Each member of a group of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; players rolls a (fair) 6-sided die. For any pair of players who throw the same number, the group scores &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; point. Find the mean and variance of the total score of the group.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        An urn contains &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; balls numbered &amp;lt;math&amp;gt;1, 2, \ldots, n&amp;lt;/math&amp;gt;. We select &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; balls uniformly at random &amp;lt;strong&amp;gt;without replacement&amp;lt;/strong&amp;gt; and add up their numbers. Find the mean and variance of the sum.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (IV)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
      Let &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; be an integer-valued, positive random variable and let &amp;lt;math&amp;gt;\{X_i\}_{i=1}^{\infty}&amp;lt;/math&amp;gt; be indepedently identically distributed random variables that are independent of &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt;, too. &lt;br /&gt;
Precisely, for any finite subset &amp;lt;math&amp;gt;I \subseteq\mathbb{N}_+&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\{X_i\}_{i \in I}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; are mutually independent. Let &amp;lt;math&amp;gt;X = \sum_{i=1}^N X_i&amp;lt;/math&amp;gt;, show that &amp;lt;math&amp;gt;\textbf{Var}[X] = \textbf{Var}[X_1] \mathbb{E}[N] + \mathbb{E}[X_1]^2 \textbf{Var}[N]&amp;lt;/math&amp;gt;.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Moments (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Show that [math]G(t) = \frac{e^t}{4} + \frac{e^{-t}}{2} + \frac{1}{4}[/math] is a moment-generating function of a random variable, and write the probability mass function of this random variable.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Moments (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X\sim \text{Geo}(p)&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt;. Find &amp;lt;math&amp;gt;\mathbb{E}[X^3]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbb{E}[X^4]&amp;lt;/math&amp;gt;. &lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Moments (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X\sim \text{Pois}(\lambda)&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;\lambda &amp;gt;0 &amp;lt;/math&amp;gt;. Find &amp;lt;math&amp;gt;\mathbb{E}[X^3]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbb{E}[X^4]&amp;lt;/math&amp;gt;. &lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be discrete random variables with correlation &amp;lt;math&amp;gt;\rho&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;|\rho|= 1&amp;lt;/math&amp;gt; if and only if &amp;lt;math&amp;gt;X=aY+b&amp;lt;/math&amp;gt; for some real numbers &amp;lt;math&amp;gt;a,b&amp;lt;/math&amp;gt;.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
    Let [math]X[/math] and [math]Y[/math] be discrete random variables with mean &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;, variance &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;, and correlation [math]\rho[/math]. Show that [math]\mathbb{E}(\max\{X^2,Y^2\})\leq 1+\sqrt{1-\rho^2}[/math]. (Hint: use the identity [math]\max\{a,b\} = \frac{1}{2}(a+b+|a-b|)[/math].)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
   Let [math]X[/math] and [math]Y[/math] be independent Bernoulli random variables with parameter [math]1/2[/math]. Show that [math]X+Y[/math] and [math]|X-Y|[/math] are dependent though uncorrelated.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Inequalities) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Reverse Markov&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable with bounded range &amp;lt;math&amp;gt;0 \le X \le U&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;U &amp;gt; 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbf{Pr}(X \le a) \le \frac{U-\mathbf{E}[X]}{U-a}&amp;lt;/math&amp;gt; for any &amp;lt;math&amp;gt;0 &amp;lt; a &amp;lt; U&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Markov&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable. Show that for all &amp;lt;math&amp;gt;\beta \geq 0&amp;lt;/math&amp;gt; and all &amp;lt;math&amp;gt;x &amp;gt; 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}(X\geq x)\leq \mathbb{E}(e^{\beta X})e^{-\beta x}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Cantelli&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable with mean &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; and variance &amp;lt;math&amp;gt;\sigma^2&amp;lt;/math&amp;gt;. Prove that for any &amp;lt;math&amp;gt;\lambda &amp;gt; 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}[X \ge \lambda] \le \frac{\sigma^2}{\lambda^2+\sigma^2}&amp;lt;/math&amp;gt;. (Hint: You may first show that &amp;lt;math&amp;gt;\mathbf{Pr}[X \ge \lambda] \le \frac{\sigma^2 + u^2}{(\lambda + u)^2}&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;u &amp;gt; 0&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
&amp;lt;strong&amp;gt;[Chebyshev&#039;s inequality (I)]&amp;lt;/strong&amp;gt; Fix &amp;lt;math&amp;gt;0 &amp;lt; b \le a&amp;lt;/math&amp;gt;. Construct a random variable &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;\mathbb{E}[X^2] = b^2&amp;lt;/math&amp;gt; for which &amp;lt;math&amp;gt;\mathbf{Pr}(|X| \ge a) = b^2/a^2&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
&amp;lt;strong&amp;gt;[Chebyshev&#039;s inequality (II)]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a random variable with &amp;lt;math&amp;gt;0 &amp;lt; \mathbb{E}[X^2] &amp;lt; \infty&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\lim_{a \to \infty} \frac{a^2 \mathbf{Pr}(|X| \ge a)}{ \mathbb{E}[X^2] } = 0&amp;lt;/math&amp;gt;. (Hint: Use the dominated convergence theorem. For discrete random variables, it can be formulated as follows: Let &amp;lt;math&amp;gt;Z,X, X_1,X_2,\ldots,X_n,\ldots&amp;lt;/math&amp;gt; be discrete random variables with finite second moments. If &amp;lt;math&amp;gt;|X_n| \le Z&amp;lt;/math&amp;gt; and for any &amp;lt;math&amp;gt;a \in \mathbb{R}&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}(X_n = a) \to \mathbf{Pr}(X = a)&amp;lt;/math&amp;gt; when &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity, then &amp;lt;math&amp;gt;\mathbb{E}[X_n^2] \to \mathbb{E}[X^2]&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Probability meets graph theory) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;    &lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;4-clique threshold&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Prove that &amp;lt;math&amp;gt;p = n^{-2/3}&amp;lt;/math&amp;gt; is the threshold probability for the existence of 4-clique. &lt;br /&gt;
        Formally, you are required to show that &lt;br /&gt;
        &amp;lt;ol type=&amp;quot;a&amp;quot;&amp;gt;&lt;br /&gt;
            &amp;lt;li&amp;gt;&lt;br /&gt;
                with a probability approaching to &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; (as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity) the Erdős–Rényi random graph &amp;lt;math&amp;gt;\mathbf{G} = \mathbf{G}(n,p)&amp;lt;/math&amp;gt; contains a 4-clique when &amp;lt;math&amp;gt;p = \omega(n^{-2/3})&amp;lt;/math&amp;gt;; (Hint: use Chebyshev&#039;s inequality.)&lt;br /&gt;
            &amp;lt;/li&amp;gt;&lt;br /&gt;
            &amp;lt;li&amp;gt;&lt;br /&gt;
                with a probability approaching to &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; (as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity) the Erdős–Rényi random graph &amp;lt;math&amp;gt;\mathbf{G} = \mathbf{G}(n,p)&amp;lt;/math&amp;gt; contains a 4-clique when &amp;lt;math&amp;gt;p = o(n^{-2/3})&amp;lt;/math&amp;gt;. (Hint: use Markov&#039;s inequality.)&lt;br /&gt;
            &amp;lt;/li&amp;gt;&lt;br /&gt;
        &amp;lt;/ol&amp;gt;&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_3&amp;diff=13037</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 3</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_3&amp;diff=13037"/>
		<updated>2025-04-05T03:56:22Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: /* Problem 1 (Warm-up Problems) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 3==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Warm-up Problems) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X_1,X_2,\cdots, X_n&amp;lt;/math&amp;gt; be pairwise independent random variables. Show that &amp;lt;math&amp;gt;\textbf{Var}\left[\sum_{i=1}^n X_i\right] =\sum_{i=1}^n \textbf{Var} [X_i]&amp;lt;/math&amp;gt;.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Each member of a group of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; players rolls a (fair) 6-sided die. For any pair of players who throw the same number, the group scores &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; point. Find the mean and variance of the total score of the group.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        An urn contains &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; balls numbered &amp;lt;math&amp;gt;1, 2, \ldots, n&amp;lt;/math&amp;gt;. We select &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; balls uniformly at random &amp;lt;strong&amp;gt;without replacement&amp;lt;/strong&amp;gt; and add up their numbers. Find the mean and variance of the sum.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (IV)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
      Let &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; be an integer-valued, positive random variable and let &amp;lt;math&amp;gt;\{X_i\}_{i=1}^{\infty}&amp;lt;/math&amp;gt; be indepedently identically distributed random variables that are independent of &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt;, too. &lt;br /&gt;
Precisely, for any finite subset &amp;lt;math&amp;gt;I \subseteq\mathbb{N}_+&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\{X_i\}_{i \in I}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; are mutually independent. Let &amp;lt;math&amp;gt;X = \sum_{i=1}^N X_i&amp;lt;/math&amp;gt;, show that &amp;lt;math&amp;gt;\textbf{Var}[X] = \textbf{Var}[X_1] \mathbb{E}[N] + \mathbb{E}[X_1]^2 \textbf{Var}[N]&amp;lt;/math&amp;gt;.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Moments (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Show that [math]G(t) = \frac{e^t}{4} + \frac{e^{-t}}{2} + \frac{1}{4}[/math] is a moment-generating function of a random variable, and write the probability mass function of this random variable.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Moments (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X\sim \text{Geo}(p)&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt;. Find &amp;lt;math&amp;gt;\mathbb{E}[X^3]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbb{E}[X^4]&amp;lt;/math&amp;gt;. &lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Moments (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X\sim \text{Pois}(\lambda)&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;\lambda &amp;gt;0 &amp;lt;/math&amp;gt;. Find &amp;lt;math&amp;gt;\mathbb{E}[X^3]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbb{E}[X^4]&amp;lt;/math&amp;gt;. &lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be independent Bernoulli random variables with parameter &amp;lt;math&amp;gt;1/2&amp;lt;/math&amp;gt; . Show that &amp;lt;math&amp;gt;X + Y&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;|X− Y |&amp;lt;/math&amp;gt; are dependent though uncorrelated.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
    Let [math]X[/math] and [math]Y[/math] be discrete random variables with mean &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;, variance &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;, and correlation [math]\rho[/math]. Show that [math]\mathbb{E}(\max\{X^2,Y^2\})\leq 1+\sqrt{1-\rho^2}[/math]. (Hint: use the identity [math]\max\{a,b\} = \frac{1}{2}(a+b+|a-b|)[/math].)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
   Let [math]X[/math] and [math]Y[/math] be independent Bernoulli random variables with parameter [math]1/2[/math]. Show that [math]X+Y[/math] and [math]|X-Y|[/math] are dependent though uncorrelated.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Inequalities) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Reverse Markov&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable with bounded range &amp;lt;math&amp;gt;0 \le X \le U&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;U &amp;gt; 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbf{Pr}(X \le a) \le \frac{U-\mathbf{E}[X]}{U-a}&amp;lt;/math&amp;gt; for any &amp;lt;math&amp;gt;0 &amp;lt; a &amp;lt; U&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Markov&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable. Show that for all &amp;lt;math&amp;gt;\beta \geq 0&amp;lt;/math&amp;gt; and all &amp;lt;math&amp;gt;x &amp;gt; 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}(X\geq x)\leq \mathbb{E}(e^{\beta X})e^{-\beta x}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Cantelli&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable with mean &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; and variance &amp;lt;math&amp;gt;\sigma^2&amp;lt;/math&amp;gt;. Prove that for any &amp;lt;math&amp;gt;\lambda &amp;gt; 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}[X \ge \lambda] \le \frac{\sigma^2}{\lambda^2+\sigma^2}&amp;lt;/math&amp;gt;. (Hint: You may first show that &amp;lt;math&amp;gt;\mathbf{Pr}[X \ge \lambda] \le \frac{\sigma^2 + u^2}{(\lambda + u)^2}&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;u &amp;gt; 0&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
&amp;lt;strong&amp;gt;[Chebyshev&#039;s inequality (I)]&amp;lt;/strong&amp;gt; Fix &amp;lt;math&amp;gt;0 &amp;lt; b \le a&amp;lt;/math&amp;gt;. Construct a random variable &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;\mathbb{E}[X^2] = b^2&amp;lt;/math&amp;gt; for which &amp;lt;math&amp;gt;\mathbf{Pr}(|X| \ge a) = b^2/a^2&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
&amp;lt;strong&amp;gt;[Chebyshev&#039;s inequality (II)]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a random variable with &amp;lt;math&amp;gt;0 &amp;lt; \mathbb{E}[X^2] &amp;lt; \infty&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\lim_{a \to \infty} \frac{a^2 \mathbf{Pr}(|X| \ge a)}{ \mathbb{E}[X^2] } = 0&amp;lt;/math&amp;gt;. (Hint: Use the dominated convergence theorem. For discrete random variables, it can be formulated as follows: Let &amp;lt;math&amp;gt;Z,X, X_1,X_2,\ldots,X_n,\ldots&amp;lt;/math&amp;gt; be discrete random variables with finite second moments. If &amp;lt;math&amp;gt;|X_n| \le Z&amp;lt;/math&amp;gt; and for any &amp;lt;math&amp;gt;a \in \mathbb{R}&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}(X_n = a) \to \mathbf{Pr}(X = a)&amp;lt;/math&amp;gt; when &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity, then &amp;lt;math&amp;gt;\mathbb{E}[X_n^2] \to \mathbb{E}[X^2]&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Probability meets graph theory) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;    &lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;4-clique threshold&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Prove that &amp;lt;math&amp;gt;p = n^{-2/3}&amp;lt;/math&amp;gt; is the threshold probability for the existence of 4-clique. &lt;br /&gt;
        Formally, you are required to show that &lt;br /&gt;
        &amp;lt;ol type=&amp;quot;a&amp;quot;&amp;gt;&lt;br /&gt;
            &amp;lt;li&amp;gt;&lt;br /&gt;
                with a probability approaching to &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; (as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity) the Erdős–Rényi random graph &amp;lt;math&amp;gt;\mathbf{G} = \mathbf{G}(n,p)&amp;lt;/math&amp;gt; contains a 4-clique when &amp;lt;math&amp;gt;p = \omega(n^{-2/3})&amp;lt;/math&amp;gt;; (Hint: use Chebyshev&#039;s inequality.)&lt;br /&gt;
            &amp;lt;/li&amp;gt;&lt;br /&gt;
            &amp;lt;li&amp;gt;&lt;br /&gt;
                with a probability approaching to &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; (as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity) the Erdős–Rényi random graph &amp;lt;math&amp;gt;\mathbf{G} = \mathbf{G}(n,p)&amp;lt;/math&amp;gt; contains a 4-clique when &amp;lt;math&amp;gt;p = o(n^{-2/3})&amp;lt;/math&amp;gt;. (Hint: use Markov&#039;s inequality.)&lt;br /&gt;
            &amp;lt;/li&amp;gt;&lt;br /&gt;
        &amp;lt;/ol&amp;gt;&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_3&amp;diff=13036</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 3</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_3&amp;diff=13036"/>
		<updated>2025-04-05T03:56:03Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: Created page with &amp;quot;*每道题目的解答都要有完整的解题过程，中英文不限。  *我们推荐大家使用LaTeX, markdown等对作业进行排版。  == Assumption throughout Problem Set 3== &amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;  &amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;  &amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural l...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 3==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Warm-up Problems) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X_1,X_2,\cdots, X_n&amp;lt;/math&amp;gt; be pairwise independent random variables. Show that &amp;lt;math&amp;gt;\textbf{Var}\left[\sum_{i=1}^n X_i\right] =\sum_{i=1}^n \textbf{Var} [X_i]&amp;lt;/math&amp;gt;.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Each member of a group of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; players rolls a (fair) 6-sided die. For any pair of players who throw the same number, the group scores &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; point. Find the mean and variance of the total score of the group.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        An urn contains &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; balls numbered &amp;lt;math&amp;gt;1, 2, \ldots, n&amp;lt;/math&amp;gt;. We select &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; balls uniformly at random &amp;lt;strong&amp;gt;without replacement&amp;lt;/strong&amp;gt; and add up their numbers. Find the mean and variance of the sum.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Variance (IV)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
      Let &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; be an integer-valued, positive random variable and let &amp;lt;math&amp;gt;\{X_i\}_{i=1}^{\infty}&amp;lt;/math&amp;gt; be indepedently identically distributed random variables that are independent of &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt;, too. &lt;br /&gt;
Precisely, for any finite subset &amp;lt;math&amp;gt;I \subseteq\mathbb{N}_+&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\{X_i\}_{i \in I}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; are mutually independent. Let &amp;lt;math&amp;gt;X = \sum_{i=1}^N X_i&amp;lt;/math&amp;gt;, show that &amp;lt;math&amp;gt;\textbf{Var}[X] = \textbf{Var}[X_1] \mathbb{E}[N] + \mathbb{E}[X_1]^2 \textbf{Var}[N]&amp;lt;/math&amp;gt;.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Moments (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Show that [math]G(t) = \frac{e^t}{4} + \frac{e^{-t}}{2} + \frac{1}{4}[/math] is a moment-generating function of a random variable, and write the probability mass function of this random variable.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Moments (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X\sim \text{Geo}(p)&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt;. Find &amp;lt;math&amp;gt;\mathbb{E}[X^3]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbb{E}[X^4]&amp;lt;/math&amp;gt;. &lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Moments (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X\sim \text{Pois}(\lambda)&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;\lambda &amp;gt;0 &amp;lt;/math&amp;gt;. Find &amp;lt;math&amp;gt;\mathbb{E}[X^3]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbb{E}[X^4]&amp;lt;/math&amp;gt;. &lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be discrete random variables with correlation &amp;lt;math&amp;gt;\rho&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;|\rho|\leq 1&amp;lt;/math&amp;gt;.&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
    Let [math]X[/math] and [math]Y[/math] be discrete random variables with mean &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;, variance &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;, and correlation [math]\rho[/math]. Show that [math]\mathbb{E}(\max\{X^2,Y^2\})\leq 1+\sqrt{1-\rho^2}[/math]. (Hint: use the identity [math]\max\{a,b\} = \frac{1}{2}(a+b+|a-b|)[/math].)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Covariance and correlation (III)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
   Let [math]X[/math] and [math]Y[/math] be independent Bernoulli random variables with parameter [math]1/2[/math]. Show that [math]X+Y[/math] and [math]|X-Y|[/math] are dependent though uncorrelated.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Inequalities) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Reverse Markov&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable with bounded range &amp;lt;math&amp;gt;0 \le X \le U&amp;lt;/math&amp;gt; for some &amp;lt;math&amp;gt;U &amp;gt; 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbf{Pr}(X \le a) \le \frac{U-\mathbf{E}[X]}{U-a}&amp;lt;/math&amp;gt; for any &amp;lt;math&amp;gt;0 &amp;lt; a &amp;lt; U&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Markov&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable. Show that for all &amp;lt;math&amp;gt;\beta \geq 0&amp;lt;/math&amp;gt; and all &amp;lt;math&amp;gt;x &amp;gt; 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}(X\geq x)\leq \mathbb{E}(e^{\beta X})e^{-\beta x}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
    &amp;lt;strong&amp;gt;[Cantelli&#039;s inequality]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a discrete random variable with mean &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; and variance &amp;lt;math&amp;gt;\sigma^2&amp;lt;/math&amp;gt;. Prove that for any &amp;lt;math&amp;gt;\lambda &amp;gt; 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}[X \ge \lambda] \le \frac{\sigma^2}{\lambda^2+\sigma^2}&amp;lt;/math&amp;gt;. (Hint: You may first show that &amp;lt;math&amp;gt;\mathbf{Pr}[X \ge \lambda] \le \frac{\sigma^2 + u^2}{(\lambda + u)^2}&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;u &amp;gt; 0&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
&amp;lt;strong&amp;gt;[Chebyshev&#039;s inequality (I)]&amp;lt;/strong&amp;gt; Fix &amp;lt;math&amp;gt;0 &amp;lt; b \le a&amp;lt;/math&amp;gt;. Construct a random variable &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;\mathbb{E}[X^2] = b^2&amp;lt;/math&amp;gt; for which &amp;lt;math&amp;gt;\mathbf{Pr}(|X| \ge a) = b^2/a^2&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
&amp;lt;strong&amp;gt;[Chebyshev&#039;s inequality (II)]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a random variable with &amp;lt;math&amp;gt;0 &amp;lt; \mathbb{E}[X^2] &amp;lt; \infty&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\lim_{a \to \infty} \frac{a^2 \mathbf{Pr}(|X| \ge a)}{ \mathbb{E}[X^2] } = 0&amp;lt;/math&amp;gt;. (Hint: Use the dominated convergence theorem. For discrete random variables, it can be formulated as follows: Let &amp;lt;math&amp;gt;Z,X, X_1,X_2,\ldots,X_n,\ldots&amp;lt;/math&amp;gt; be discrete random variables with finite second moments. If &amp;lt;math&amp;gt;|X_n| \le Z&amp;lt;/math&amp;gt; and for any &amp;lt;math&amp;gt;a \in \mathbb{R}&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}(X_n = a) \to \mathbf{Pr}(X = a)&amp;lt;/math&amp;gt; when &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity, then &amp;lt;math&amp;gt;\mathbb{E}[X_n^2] \to \mathbb{E}[X^2]&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Probability meets graph theory) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;    &lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;4-clique threshold&amp;lt;/strong&amp;gt;]&lt;br /&gt;
        Prove that &amp;lt;math&amp;gt;p = n^{-2/3}&amp;lt;/math&amp;gt; is the threshold probability for the existence of 4-clique. &lt;br /&gt;
        Formally, you are required to show that &lt;br /&gt;
        &amp;lt;ol type=&amp;quot;a&amp;quot;&amp;gt;&lt;br /&gt;
            &amp;lt;li&amp;gt;&lt;br /&gt;
                with a probability approaching to &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; (as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity) the Erdős–Rényi random graph &amp;lt;math&amp;gt;\mathbf{G} = \mathbf{G}(n,p)&amp;lt;/math&amp;gt; contains a 4-clique when &amp;lt;math&amp;gt;p = \omega(n^{-2/3})&amp;lt;/math&amp;gt;; (Hint: use Chebyshev&#039;s inequality.)&lt;br /&gt;
            &amp;lt;/li&amp;gt;&lt;br /&gt;
            &amp;lt;li&amp;gt;&lt;br /&gt;
                with a probability approaching to &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; (as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity) the Erdős–Rényi random graph &amp;lt;math&amp;gt;\mathbf{G} = \mathbf{G}(n,p)&amp;lt;/math&amp;gt; contains a 4-clique when &amp;lt;math&amp;gt;p = o(n^{-2/3})&amp;lt;/math&amp;gt;. (Hint: use Markov&#039;s inequality.)&lt;br /&gt;
            &amp;lt;/li&amp;gt;&lt;br /&gt;
        &amp;lt;/ol&amp;gt;&lt;br /&gt;
    &amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)&amp;diff=13035</id>
		<title>概率论与数理统计 (Spring 2025)</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)&amp;diff=13035"/>
		<updated>2025-04-05T03:49:54Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: /* Assignments */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Infobox&lt;br /&gt;
|name         = Infobox&lt;br /&gt;
|bodystyle    = &lt;br /&gt;
|title        = &amp;lt;font size=3&amp;gt;&#039;&#039;&#039;概率论与数理统计&#039;&#039;&#039;&amp;lt;br&amp;gt;&lt;br /&gt;
&#039;&#039;&#039;Probability Theory&#039;&#039;&#039; &amp;lt;br&amp;gt; &amp;amp; &#039;&#039;&#039;Mathematical Statistics&#039;&#039;&#039;&amp;lt;/font&amp;gt;&lt;br /&gt;
|titlestyle   = &lt;br /&gt;
&lt;br /&gt;
|image        = &lt;br /&gt;
|imagestyle   = &lt;br /&gt;
|caption      = &lt;br /&gt;
|captionstyle = &lt;br /&gt;
|headerstyle  = background:#ccf;&lt;br /&gt;
|labelstyle   = background:#ddf;&lt;br /&gt;
|datastyle    = &lt;br /&gt;
&lt;br /&gt;
|header1 =Instructor&lt;br /&gt;
|label1  = &lt;br /&gt;
|data1   = &lt;br /&gt;
|header2 = &lt;br /&gt;
|label2  = &lt;br /&gt;
|data2   = &#039;&#039;&#039;尹一通&#039;&#039;&#039;&lt;br /&gt;
|header3 = &lt;br /&gt;
|label3  = Email&lt;br /&gt;
|data3   = yinyt@nju.edu.cn  &lt;br /&gt;
|header4 =&lt;br /&gt;
|label4  = office&lt;br /&gt;
|data4   = 计算机系 804&lt;br /&gt;
|header5 = &lt;br /&gt;
|label5  = &lt;br /&gt;
|data5   = &#039;&#039;&#039;刘景铖&#039;&#039;&#039;&lt;br /&gt;
|header6 = &lt;br /&gt;
|label6  = Email&lt;br /&gt;
|data6   = liu@nju.edu.cn  &lt;br /&gt;
|header7 =&lt;br /&gt;
|label7  = office&lt;br /&gt;
|data7   = 计算机系 516&lt;br /&gt;
|header8 = Class&lt;br /&gt;
|label8  = &lt;br /&gt;
|data8   = &lt;br /&gt;
|header9 =&lt;br /&gt;
|label9  = Class meeting&lt;br /&gt;
|data9   = Monday, 10am-12pm&amp;lt;br&amp;gt;&lt;br /&gt;
Wednesday (单), 10am-12pm&amp;lt;br&amp;gt;&lt;br /&gt;
仙Ⅰ-204&lt;br /&gt;
|header10=&lt;br /&gt;
|label10 = Office hour&lt;br /&gt;
|data10  = Friday, 2pm-3pm &amp;lt;br&amp;gt;计算机系 804（尹一通）&amp;lt;br&amp;gt;计算机系 516（刘景铖）&lt;br /&gt;
|header11= Textbook&lt;br /&gt;
|label11 = &lt;br /&gt;
|data11  = &lt;br /&gt;
|header12=&lt;br /&gt;
|label12 = &lt;br /&gt;
|data12  = [[File:概率导论.jpeg|border|100px]]&lt;br /&gt;
|header13=&lt;br /&gt;
|label13 = &lt;br /&gt;
|data13  = &#039;&#039;&#039;概率导论&#039;&#039;&#039;（第2版·修订版）&amp;lt;br&amp;gt; Dimitri P. Bertsekas and John N. Tsitsiklis&amp;lt;br&amp;gt; 郑忠国 童行伟 译；人民邮电出版社 (2022)&lt;br /&gt;
|header14=&lt;br /&gt;
|label14 = &lt;br /&gt;
|data14  = [[File:Grimmett_probability.jpg|border|100px]]&lt;br /&gt;
|header15=&lt;br /&gt;
|label15 = &lt;br /&gt;
|data15  = &#039;&#039;&#039;Probability and Random Processes&#039;&#039;&#039; (4E) &amp;lt;br&amp;gt; Geoffrey Grimmett and David Stirzaker &amp;lt;br&amp;gt;  Oxford University Press (2020)&lt;br /&gt;
|header16=&lt;br /&gt;
|label16 = &lt;br /&gt;
|data16  = [[File:Probability_and_Computing_2ed.jpg|border|100px]]&lt;br /&gt;
|header17=&lt;br /&gt;
|label17 = &lt;br /&gt;
|data17  = &#039;&#039;&#039;Probability and Computing&#039;&#039;&#039; (2E) &amp;lt;br&amp;gt; Michael Mitzenmacher and Eli Upfal &amp;lt;br&amp;gt;   Cambridge University Press (2017)&lt;br /&gt;
|belowstyle = background:#ddf;&lt;br /&gt;
|below = &lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
This is the webpage for the &#039;&#039;Probability Theory and Mathematical Statistics&#039;&#039; (概率论与数理统计) class of Spring 2025. Students who take this class should check this page periodically for content updates and new announcements. &lt;br /&gt;
&lt;br /&gt;
= Announcement =&lt;br /&gt;
* TBA&lt;br /&gt;
&lt;br /&gt;
= Course info =&lt;br /&gt;
* &#039;&#039;&#039;Instructor &#039;&#039;&#039;: &lt;br /&gt;
:* [http://tcs.nju.edu.cn/yinyt/ 尹一通]：[mailto:yinyt@nju.edu.cn &amp;lt;yinyt@nju.edu.cn&amp;gt;]，计算机系 804 &lt;br /&gt;
:* [https://liuexp.github.io 刘景铖]：[mailto:liu@nju.edu.cn &amp;lt;liu@nju.edu.cn&amp;gt;]，计算机系 516 &lt;br /&gt;
* &#039;&#039;&#039;Teaching assistant&#039;&#039;&#039;:&lt;br /&gt;
** [https://sites.google.com/view/xinyuanzhang 张昕渊]：[mailto:zhangxy@smail.nju.edu.cn &amp;lt;zhangxy@smail.nju.edu.cn&amp;gt;]，计算机系 410&lt;br /&gt;
** 邹宗瑞：[mailto:zou.zongrui@smail.nju.edu.cn &amp;lt;zou.zongrui@smail.nju.edu.cn&amp;gt;]，计算机系 410&lt;br /&gt;
* &#039;&#039;&#039;Class meeting&#039;&#039;&#039;:&lt;br /&gt;
** 周一：10am-12pm，仙Ⅰ-204&lt;br /&gt;
** 周三（单）：10am-12pm，仙Ⅰ-204&lt;br /&gt;
* &#039;&#039;&#039;Office hour&#039;&#039;&#039;: &lt;br /&gt;
:* 周五 2pm-3pm, 计算机系 804（尹一通）&lt;br /&gt;
:* 周五 2pm-3pm, 计算机系 516（刘景铖）&lt;br /&gt;
:* &#039;&#039;&#039;QQ群&#039;&#039;&#039;: 974742320（申请加入需提供姓名、院系、学号）&lt;br /&gt;
&lt;br /&gt;
= Syllabus =&lt;br /&gt;
课程内容分为三大部分：&lt;br /&gt;
* &#039;&#039;&#039;经典概率论&#039;&#039;&#039;：概率空间、随机变量及其数字特征、多维与连续随机变量、极限定理等内容&lt;br /&gt;
* &#039;&#039;&#039;概率与计算&#039;&#039;&#039;：测度集中现象 (concentration of measure)、概率法 (the probabilistic method)、离散随机过程的相关专题&lt;br /&gt;
* &#039;&#039;&#039;数理统计&#039;&#039;&#039;：参数估计、假设检验、贝叶斯估计、线性回归等统计推断等概念&lt;br /&gt;
&lt;br /&gt;
对于第一和第二部分，要求清楚掌握基本概念，深刻理解关键的现象与规律以及背后的原理，并可以灵活运用所学方法求解相关问题。对于第三部分，要求熟悉数理统计的若干基本概念，以及典型的统计模型和统计推断问题。&lt;br /&gt;
&lt;br /&gt;
经过本课程的训练，力求使学生能够熟悉掌握概率的语言，并会利用概率思维来理解客观世界并对其建模，以及驾驭概率的数学工具来分析和求解专业问题。&lt;br /&gt;
&lt;br /&gt;
=== 教材与参考书 Course Materials ===&lt;br /&gt;
* &#039;&#039;&#039;[BT]&#039;&#039;&#039; 概率导论（第2版·修订版），[美]伯特瑟卡斯（Dimitri P.Bertsekas）[美]齐齐克利斯（John N.Tsitsiklis）著，郑忠国 童行伟 译，人民邮电出版社（2022）。&lt;br /&gt;
* &#039;&#039;&#039;[GS]&#039;&#039;&#039; &#039;&#039;Probability and Random Processes&#039;&#039;, by Geoffrey Grimmett and David Stirzaker; Oxford University Press; 4th edition (2020).&lt;br /&gt;
* &#039;&#039;&#039;[MU]&#039;&#039;&#039; &#039;&#039;Probability and Computing: Randomization and Probabilistic Techniques in Algorithms and Data Analysis&#039;&#039;, by Michael Mitzenmacher, Eli Upfal; Cambridge University Press; 2nd edition (2017).&lt;br /&gt;
&lt;br /&gt;
=== 成绩 Grading Policy ===&lt;br /&gt;
* 课程成绩：本课程将会有若干次作业和一次期末考试。最终成绩将由平时作业成绩和期末考试成绩综合得出。&lt;br /&gt;
* 迟交：如果有特殊的理由，无法按时完成作业，请提前联系授课老师，给出正当理由。否则迟交的作业将不被接受。&lt;br /&gt;
&lt;br /&gt;
=== &amp;lt;font color=red&amp;gt; 学术诚信 Academic Integrity &amp;lt;/font&amp;gt;===&lt;br /&gt;
学术诚信是所有从事学术活动的学生和学者最基本的职业道德底线，本课程将不遗余力的维护学术诚信规范，违反这一底线的行为将不会被容忍。&lt;br /&gt;
&lt;br /&gt;
作业完成的原则：署你名字的工作必须是你个人的贡献。在完成作业的过程中，允许讨论，前提是讨论的所有参与者均处于同等完成度。但关键想法的执行、以及作业文本的写作必须独立完成，并在作业中致谢（acknowledge）所有参与讨论的人。符合规则的讨论与致谢将不会影响得分。不允许其他任何形式的合作——尤其是与已经完成作业的同学“讨论”。&lt;br /&gt;
&lt;br /&gt;
本课程将对剽窃行为采取零容忍的态度。在完成作业过程中，对他人工作（出版物、互联网资料、其他人的作业等）直接的文本抄袭和对关键思想、关键元素的抄袭，按照 [http://www.acm.org/publications/policies/plagiarism_policy ACM Policy on Plagiarism]的解释，都将视为剽窃。剽窃者成绩将被取消。如果发现互相抄袭行为，&amp;lt;font color=red&amp;gt; 抄袭和被抄袭双方的成绩都将被取消&amp;lt;/font&amp;gt;。因此请主动防止自己的作业被他人抄袭。&lt;br /&gt;
&lt;br /&gt;
学术诚信影响学生个人的品行，也关乎整个教育系统的正常运转。为了一点分数而做出学术不端的行为，不仅使自己沦为一个欺骗者，也使他人的诚实努力失去意义。让我们一起努力维护一个诚信的环境。&lt;br /&gt;
&lt;br /&gt;
= Assignments =&lt;br /&gt;
*[[概率论与数理统计 (Spring 2025)/Problem Set 1|Problem Set 1]]  请在 2025/3/10 上课之前(10am UTC+8)提交到 [mailto:pr_nju@163.com pr_nju@163.com] (文件名为&#039;&amp;lt;font color=red &amp;gt;学号_姓名_A1.pdf&amp;lt;/font&amp;gt;&#039;).&lt;br /&gt;
** [[概率论与数理统计 (Spring 2025)/第一次作业提交名单|第一次作业提交名单]]&lt;br /&gt;
&lt;br /&gt;
*[[概率论与数理统计 (Spring 2025)/Problem Set 2|Problem Set 2]]  请在 2025/3/31 上课之前(10am UTC+8)提交到 [mailto:pr_nju@163.com pr_nju@163.com] (文件名为&#039;&amp;lt;font color=red &amp;gt;学号_姓名_A2.pdf&amp;lt;/font&amp;gt;&#039;).&lt;br /&gt;
** [[概率论与数理统计 (Spring 2025)/第二次作业提交名单|第二次作业提交名单]]&lt;br /&gt;
&lt;br /&gt;
*[[概率论与数理统计 (Spring 2025)/Problem Set 3|Problem Set 3]]  请在 2025/4/21 上课之前(10am UTC+8)提交到 [mailto:pr_nju@163.com pr_nju@163.com] (文件名为&#039;&amp;lt;font color=red &amp;gt;学号_姓名_A3.pdf&amp;lt;/font&amp;gt;&#039;).&lt;br /&gt;
&lt;br /&gt;
= Lectures =&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2025/Intro.pdf 课程简介]&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2025/ProbSpace.pdf 概率空间]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 第1章&#039;&#039;&#039; 或 &#039;&#039;&#039;[GS] Chapter 1&#039;&#039;&#039;&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2025)/Entropy and volume of Hamming balls|Entropy and volume of Hamming balls]]&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2025/RandVar.pdf 随机变量]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 第2章&#039;&#039;&#039; 或 &#039;&#039;&#039;[GS] Chapter 2, Sections 3.1~3.5, 3.7&#039;&#039;&#039;&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[MU] Chapter 2&#039;&#039;&#039;&lt;br /&gt;
#* the [http://tcs.nju.edu.cn/slides/prob2024/discrete-pmf.nb &#039;&#039;&#039;Mathematica notebook file&#039;&#039;&#039;] for the PMFs of basic discrete distributions&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2025)/Average-case analysis of QuickSort|Average-case analysis of &#039;&#039;&#039;&#039;&#039;QuickSort&#039;&#039;&#039;&#039;&#039;]]&lt;br /&gt;
#* [https://www.bilibili.com/video/BV1ta411A7fp/ 高尔顿板（Galton board）视频] 和 [https://en.wikipedia.org/wiki/Galton_board 维基百科页面]&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2025/Deviation.pdf 矩与偏差]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[MU] Chapter 3&#039;&#039;&#039;&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 章节 2.4, 4.2, 4.3, 5.1&#039;&#039;&#039; 或 &#039;&#039;&#039;[GS] Sections 3.3, 3.6, 7.3&#039;&#039;&#039;&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2025)/Two-point sampling|Two-point sampling]]&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2025)/Threshold of k-clique in random graph|Threshold of &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;-clique in random graph]]&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2025)/Weierstrass Approximation Theorem|Weierstrass approximation]]&lt;br /&gt;
&lt;br /&gt;
= Concepts =&lt;br /&gt;
* [https://plato.stanford.edu/entries/probability-interpret/ Interpretations of probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/History_of_probability History of probability]&lt;br /&gt;
* Example problems:&lt;br /&gt;
** [https://dornsifecms.usc.edu/assets/sites/520/docs/VonNeumann-ams12p36-38.pdf von Neumann&#039;s Bernoulli factory] and other [https://peteroupc.github.io/bernoulli.html Bernoulli factory algorithms]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Boy_or_Girl_paradox Boy or Girl paradox]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Monty_Hall_problem Monty Hall problem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bertrand_paradox_(probability) Bertrand paradox]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hard_spheres Hard spheres model] and [https://en.wikipedia.org/wiki/Ising_model Ising model]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/PageRank &#039;&#039;PageRank&#039;&#039;] and stationary [https://en.wikipedia.org/wiki/Random_walk random walk]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Diffusion_process Diffusion process] and [https://en.wikipedia.org/wiki/Diffusion_model diffusion model]&lt;br /&gt;
*[https://en.wikipedia.org/wiki/Probability_space Probability space]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Sample_space Sample space]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Event_(probability_theory) Event] and [https://en.wikipedia.org/wiki/Σ-algebra &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-algebra]&lt;br /&gt;
** Kolmogorov&#039;s [https://en.wikipedia.org/wiki/Probability_axioms axioms of probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Discrete_uniform_distribution Classical] and [https://en.wikipedia.org/wiki/Geometric_probability goemetric probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Boole%27s_inequality Union bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Inclusion%E2%80%93exclusion_principle Inclusion-Exclusion principle]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Boole%27s_inequality#Bonferroni_inequalities Bonferroni inequalities]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Conditional_probability Conditional probability]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chain_rule_(probability) Chain rule]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_total_probability Law of total probability]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bayes%27_theorem Bayes&#039; law]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Independence_(probability_theory) Independence] &lt;br /&gt;
** [https://en.wikipedia.org/wiki/Pairwise_independence Pairwise independence]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Random_variable Random variable]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cumulative_distribution_function Cumulative distribution function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_mass_function Probability mass function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_density_function Probability density function]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Multivariate_random_variable Random vector]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Joint_probability_distribution Joint probability distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Conditional_probability_distribution Conditional probability distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Marginal_distribution Marginal distribution]&lt;br /&gt;
* Some &#039;&#039;&#039;discrete&#039;&#039;&#039; probability distributions&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bernoulli_trial Bernoulli trial] and [https://en.wikipedia.org/wiki/Bernoulli_distribution Bernoulli distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Discrete_uniform_distribution Discrete uniform distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Binomial_distribution Binomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Geometric_distribution Geometric distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Negative_binomial_distribution Negative binomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hypergeometric_distribution Hypergeometric distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Poisson_distribution Poisson distribution]&lt;br /&gt;
** and [https://en.wikipedia.org/wiki/List_of_probability_distributions#Discrete_distributions others]&lt;br /&gt;
* Balls into bins model&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Multinomial_distribution Multinomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Birthday_problem Birthday problem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Coupon_collector%27s_problem Coupon collector]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Balls_into_bins_problem Occupancy problem]&lt;br /&gt;
* Random graphs&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Erd%C5%91s%E2%80%93R%C3%A9nyi_model Erdős–Rényi random graph model]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Galton%E2%80%93Watson_process Galton–Watson branching process]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Expected_value Expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_the_unconscious_statistician Law of the unconscious statistician, &#039;&#039;LOTUS&#039;&#039;]&lt;br /&gt;
** [https://dlsun.github.io/probability/linearity.html Linearity of expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Conditional_expectation Conditional expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_total_expectation Law of total expectation]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Markov%27s_inequality Markov&#039;s inequality]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Chebyshev%27s_inequality Chebyshev&#039;s inequality]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Moment_(mathematics) Moment] related&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Central_moment Central moment]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Variance Variance]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Standard_deviation Standard deviation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Covariance Covariance]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Correlation Correlation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Correlation_does_not_imply_causation Correlation does not imply causation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Skewness Skewness]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Kurtosis Kurtosis]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Moment_problem Moment problem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Measure_(mathematics) Measure]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Borel_set Borel set]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Lebesgue_integration Lebesgue integration]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Measurable_function Measurable function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cantor_set Cantor set]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Non-measurable_set Non-measurable set]&lt;br /&gt;
*[https://en.wikipedia.org/wiki/Probability_density_function Probability density function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_distribution#Absolutely_continuous_probability_distribution Continuous probability distribution]&lt;br /&gt;
**[https://en.wikipedia.org/wiki/Conditional_probability_distribution#Conditional_continuous_distributions Conditional continuous distributions]&lt;br /&gt;
**[https://en.wikipedia.org/wiki/Convolution_of_probability_distributions Convolution of probability distributions] and [https://en.wikipedia.org/wiki/List_of_convolutions_of_probability_distributions List of convolutions of probability distributions]&lt;br /&gt;
* Some &#039;&#039;&#039;continuous&#039;&#039;&#039; probability distributions&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Continuous_uniform_distribution Continuous uniform distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Exponential_distribution Exponential distribution] and [https://en.wikipedia.org/wiki/Poisson_point_process Poisson point process]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Normal_distribution Normal (Gaussion) distribution]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Gaussian_function Gaussian function] and [https://en.wikipedia.org/wiki/Gaussian_integral Gaussian integral]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Standard_normal_table Standard normal table]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/68%E2%80%9395%E2%80%9399.7_rule 68–95–99.7 rule]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Sum_of_normally_distributed_random_variables Sum of normally distributed random variables]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Multivariate_normal_distribution Multivariate normal distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chi-squared_distribution Chi-squared distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Gamma_distribution Gamma distribution] and [https://en.wikipedia.org/wiki/Gamma_function Gamma function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Beta_distribution Beta distribution] and [https://en.wikipedia.org/wiki/Beta_function Beta function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cauchy_distribution Cauchy distribution]&lt;br /&gt;
** and [https://en.wikipedia.org/wiki/List_of_probability_distributions#Absolutely_continuous_distributions others]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Inverse_transform_sampling Inverse transform sampling]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stochastic_dominance Stochastic dominance] and [https://en.wikipedia.org/wiki/Coupling_(probability) Coupling] &lt;br /&gt;
* [https://en.wikipedia.org/wiki/Moment-generating_function Moment-generating function]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Convergence_of_random_variables Convergence of random variables]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Pointwise_convergence Pointwise convergence]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Convergence_of_measures#Weak_convergence_of_measures Weak convergence of measures]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Convergence_in_measure Convergence in measure]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Skorokhod%27s_representation_theorem Skorokhod&#039;s representation theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Continuous_mapping_theorem Continuous mapping theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Borel%E2%80%93Cantelli_lemma Borel–Cantelli lemma] and [https://en.wikipedia.org/wiki/Zero%E2%80%93one_law zero-one laws]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Law_of_large_numbers Law of large numbers]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Central_limit_theorem Central limit theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/De_Moivre%E2%80%93Laplace_theorem De Moivre–Laplace theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Berry%E2%80%93Esseen_theorem Berry–Esseen theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Characteristic_function_(probability_theory) Characteristic function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Lévy%27s_continuity_theorem Lévy&#039;s continuity theorem]&lt;br /&gt;
* Concentration of measures:&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Moment-generating_function Moment-generating function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chernoff_bound Chernoff bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hoeffding%27s_inequality Hoeffding&#039;s bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Doob_martingale#McDiarmid&#039;s_inequality McDiarmid&#039;s inequality]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Doob_martingale Doob martingale] and [https://en.wikipedia.org/wiki/Azuma%27s_inequality Azuma&#039;s inequality]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Sub-Gaussian_distribution Sub-Gaussian distribution]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stochastic_process Stochastic process]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stopping_time Stopping time]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Martingale_(probability_theory) Martingale]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Optional_stopping_theorem Optional stopping theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Wald%27s_equation Wald&#039;s equation]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Markov_chain Markov chain]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Markov_property Markov property] and [https://en.wikipedia.org/wiki/Memorylessness memorylessness]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Stochastic_matrix Stochastic matrix]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Stationary_distribution Stationary distribution]&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/%E7%AC%AC%E4%BA%8C%E6%AC%A1%E4%BD%9C%E4%B8%9A%E6%8F%90%E4%BA%A4%E5%90%8D%E5%8D%95&amp;diff=13018</id>
		<title>概率论与数理统计 (Spring 2025)/第二次作业提交名单</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/%E7%AC%AC%E4%BA%8C%E6%AC%A1%E4%BD%9C%E4%B8%9A%E6%8F%90%E4%BA%A4%E5%90%8D%E5%8D%95&amp;diff=13018"/>
		<updated>2025-03-31T07:23:44Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: Created page with &amp;quot; 如有错漏请邮件联系助教. &amp;lt;center&amp;gt; {| class=&amp;quot;wikitable&amp;quot; |- ! 学号 !! 姓名 |- | 231840166 || 赵少凡  |- | 231200035 || 葛翰飞  |- | 231240011 || 陈力琰  |- | 231840058 || 王润希  |- | 231880123 || 陆博文  |- | 231240005 || 王甚易  |- | 231880394 || 翟笑晨  |- | 231240047 || 姜淮仁  |- | 231880152 || 李佳骏  |- | 231240015 || 邱文宣  |- | 231240002 || 余孟凡  |- | 231098046 || 李方泽  |- | 231240060 || 侯佳松  |- | 23124...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt; 如有错漏请邮件联系助教.&lt;br /&gt;
&amp;lt;center&amp;gt;&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! 学号 !! 姓名&lt;br /&gt;
|-&lt;br /&gt;
| 231840166 || 赵少凡 &lt;br /&gt;
|-&lt;br /&gt;
| 231200035 || 葛翰飞 &lt;br /&gt;
|-&lt;br /&gt;
| 231240011 || 陈力琰 &lt;br /&gt;
|-&lt;br /&gt;
| 231840058 || 王润希 &lt;br /&gt;
|-&lt;br /&gt;
| 231880123 || 陆博文 &lt;br /&gt;
|-&lt;br /&gt;
| 231240005 || 王甚易 &lt;br /&gt;
|-&lt;br /&gt;
| 231880394 || 翟笑晨 &lt;br /&gt;
|-&lt;br /&gt;
| 231240047 || 姜淮仁 &lt;br /&gt;
|-&lt;br /&gt;
| 231880152 || 李佳骏 &lt;br /&gt;
|-&lt;br /&gt;
| 231240015 || 邱文宣 &lt;br /&gt;
|-&lt;br /&gt;
| 231240002 || 余孟凡 &lt;br /&gt;
|-&lt;br /&gt;
| 231098046 || 李方泽 &lt;br /&gt;
|-&lt;br /&gt;
| 231240060 || 侯佳松 &lt;br /&gt;
|-&lt;br /&gt;
| 231240038 || 肖金宇 &lt;br /&gt;
|-&lt;br /&gt;
| 231240018 || 渠翔凯 &lt;br /&gt;
|-&lt;br /&gt;
| 231220005 || 樊书岩 &lt;br /&gt;
|-&lt;br /&gt;
| 231230102 || 庄铸锴 &lt;br /&gt;
|-&lt;br /&gt;
| 231240055 || 魏全烨 &lt;br /&gt;
|-&lt;br /&gt;
| 231240029 || 朱非凡 &lt;br /&gt;
|-&lt;br /&gt;
| 231220001 || 陈梓睿 &lt;br /&gt;
|-&lt;br /&gt;
| 231240021 || 陈奕帆 &lt;br /&gt;
|-&lt;br /&gt;
| 231240056 || 靳濡搏 &lt;br /&gt;
|-&lt;br /&gt;
| 231240026 || 蒙航 &lt;br /&gt;
|-&lt;br /&gt;
| 231880140 || 桂天麟 &lt;br /&gt;
|-&lt;br /&gt;
| 231240058 || 陈小川 &lt;br /&gt;
|-&lt;br /&gt;
| 231880320 || 尹铁越 &lt;br /&gt;
|-&lt;br /&gt;
| 231240007 || 高亮节 &lt;br /&gt;
|-&lt;br /&gt;
| 231840242 || 李昊凌 &lt;br /&gt;
|-&lt;br /&gt;
| 231240013 || 南晨曦 &lt;br /&gt;
|-&lt;br /&gt;
| 231240051 || 王昱霏 &lt;br /&gt;
|-&lt;br /&gt;
| 231240046 || 陈昊奕 &lt;br /&gt;
|-&lt;br /&gt;
| 231240024 || 陈越 &lt;br /&gt;
|-&lt;br /&gt;
| 231240004 || 何梓杨 &lt;br /&gt;
|-&lt;br /&gt;
| 231240019 || 邵与乔 &lt;br /&gt;
|-&lt;br /&gt;
| 231240009 || 陈心怡 &lt;br /&gt;
|-&lt;br /&gt;
| 231220002 || 潘谟天 &lt;br /&gt;
|-&lt;br /&gt;
| 231240016 || 徐冰冰 &lt;br /&gt;
|-&lt;br /&gt;
| 231220090 || 邓淇 &lt;br /&gt;
|-&lt;br /&gt;
| 231240053 || 王艺文 &lt;br /&gt;
|-&lt;br /&gt;
| 231240027 || 彭浩楠 &lt;br /&gt;
|-&lt;br /&gt;
| 231240023 || 叶宸见 &lt;br /&gt;
|-&lt;br /&gt;
| 231098142 || 陈逸帆 &lt;br /&gt;
|-&lt;br /&gt;
| 231240025 || 李熠辰 &lt;br /&gt;
|}&lt;br /&gt;
&amp;lt;/center&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)&amp;diff=13017</id>
		<title>概率论与数理统计 (Spring 2025)</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)&amp;diff=13017"/>
		<updated>2025-03-31T07:22:54Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: /* Assignments */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Infobox&lt;br /&gt;
|name         = Infobox&lt;br /&gt;
|bodystyle    = &lt;br /&gt;
|title        = &amp;lt;font size=3&amp;gt;&#039;&#039;&#039;概率论与数理统计&#039;&#039;&#039;&amp;lt;br&amp;gt;&lt;br /&gt;
&#039;&#039;&#039;Probability Theory&#039;&#039;&#039; &amp;lt;br&amp;gt; &amp;amp; &#039;&#039;&#039;Mathematical Statistics&#039;&#039;&#039;&amp;lt;/font&amp;gt;&lt;br /&gt;
|titlestyle   = &lt;br /&gt;
&lt;br /&gt;
|image        = &lt;br /&gt;
|imagestyle   = &lt;br /&gt;
|caption      = &lt;br /&gt;
|captionstyle = &lt;br /&gt;
|headerstyle  = background:#ccf;&lt;br /&gt;
|labelstyle   = background:#ddf;&lt;br /&gt;
|datastyle    = &lt;br /&gt;
&lt;br /&gt;
|header1 =Instructor&lt;br /&gt;
|label1  = &lt;br /&gt;
|data1   = &lt;br /&gt;
|header2 = &lt;br /&gt;
|label2  = &lt;br /&gt;
|data2   = &#039;&#039;&#039;尹一通&#039;&#039;&#039;&lt;br /&gt;
|header3 = &lt;br /&gt;
|label3  = Email&lt;br /&gt;
|data3   = yinyt@nju.edu.cn  &lt;br /&gt;
|header4 =&lt;br /&gt;
|label4  = office&lt;br /&gt;
|data4   = 计算机系 804&lt;br /&gt;
|header5 = &lt;br /&gt;
|label5  = &lt;br /&gt;
|data5   = &#039;&#039;&#039;刘景铖&#039;&#039;&#039;&lt;br /&gt;
|header6 = &lt;br /&gt;
|label6  = Email&lt;br /&gt;
|data6   = liu@nju.edu.cn  &lt;br /&gt;
|header7 =&lt;br /&gt;
|label7  = office&lt;br /&gt;
|data7   = 计算机系 516&lt;br /&gt;
|header8 = Class&lt;br /&gt;
|label8  = &lt;br /&gt;
|data8   = &lt;br /&gt;
|header9 =&lt;br /&gt;
|label9  = Class meeting&lt;br /&gt;
|data9   = Monday, 10am-12pm&amp;lt;br&amp;gt;&lt;br /&gt;
Wednesday (单), 10am-12pm&amp;lt;br&amp;gt;&lt;br /&gt;
仙Ⅰ-204&lt;br /&gt;
|header10=&lt;br /&gt;
|label10 = Office hour&lt;br /&gt;
|data10  = Friday, 2pm-3pm &amp;lt;br&amp;gt;计算机系 804（尹一通）&amp;lt;br&amp;gt;计算机系 516（刘景铖）&lt;br /&gt;
|header11= Textbook&lt;br /&gt;
|label11 = &lt;br /&gt;
|data11  = &lt;br /&gt;
|header12=&lt;br /&gt;
|label12 = &lt;br /&gt;
|data12  = [[File:概率导论.jpeg|border|100px]]&lt;br /&gt;
|header13=&lt;br /&gt;
|label13 = &lt;br /&gt;
|data13  = &#039;&#039;&#039;概率导论&#039;&#039;&#039;（第2版·修订版）&amp;lt;br&amp;gt; Dimitri P. Bertsekas and John N. Tsitsiklis&amp;lt;br&amp;gt; 郑忠国 童行伟 译；人民邮电出版社 (2022)&lt;br /&gt;
|header14=&lt;br /&gt;
|label14 = &lt;br /&gt;
|data14  = [[File:Grimmett_probability.jpg|border|100px]]&lt;br /&gt;
|header15=&lt;br /&gt;
|label15 = &lt;br /&gt;
|data15  = &#039;&#039;&#039;Probability and Random Processes&#039;&#039;&#039; (4E) &amp;lt;br&amp;gt; Geoffrey Grimmett and David Stirzaker &amp;lt;br&amp;gt;  Oxford University Press (2020)&lt;br /&gt;
|header16=&lt;br /&gt;
|label16 = &lt;br /&gt;
|data16  = [[File:Probability_and_Computing_2ed.jpg|border|100px]]&lt;br /&gt;
|header17=&lt;br /&gt;
|label17 = &lt;br /&gt;
|data17  = &#039;&#039;&#039;Probability and Computing&#039;&#039;&#039; (2E) &amp;lt;br&amp;gt; Michael Mitzenmacher and Eli Upfal &amp;lt;br&amp;gt;   Cambridge University Press (2017)&lt;br /&gt;
|belowstyle = background:#ddf;&lt;br /&gt;
|below = &lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
This is the webpage for the &#039;&#039;Probability Theory and Mathematical Statistics&#039;&#039; (概率论与数理统计) class of Spring 2025. Students who take this class should check this page periodically for content updates and new announcements. &lt;br /&gt;
&lt;br /&gt;
= Announcement =&lt;br /&gt;
* TBA&lt;br /&gt;
&lt;br /&gt;
= Course info =&lt;br /&gt;
* &#039;&#039;&#039;Instructor &#039;&#039;&#039;: &lt;br /&gt;
:* [http://tcs.nju.edu.cn/yinyt/ 尹一通]：[mailto:yinyt@nju.edu.cn &amp;lt;yinyt@nju.edu.cn&amp;gt;]，计算机系 804 &lt;br /&gt;
:* [https://liuexp.github.io 刘景铖]：[mailto:liu@nju.edu.cn &amp;lt;liu@nju.edu.cn&amp;gt;]，计算机系 516 &lt;br /&gt;
* &#039;&#039;&#039;Teaching assistant&#039;&#039;&#039;:&lt;br /&gt;
** [https://sites.google.com/view/xinyuanzhang 张昕渊]：[mailto:zhangxy@smail.nju.edu.cn &amp;lt;zhangxy@smail.nju.edu.cn&amp;gt;]，计算机系 410&lt;br /&gt;
** 邹宗瑞：[mailto:zou.zongrui@smail.nju.edu.cn &amp;lt;zou.zongrui@smail.nju.edu.cn&amp;gt;]，计算机系 410&lt;br /&gt;
* &#039;&#039;&#039;Class meeting&#039;&#039;&#039;:&lt;br /&gt;
** 周一：10am-12pm，仙Ⅰ-204&lt;br /&gt;
** 周三（单）：10am-12pm，仙Ⅰ-204&lt;br /&gt;
* &#039;&#039;&#039;Office hour&#039;&#039;&#039;: &lt;br /&gt;
:* 周五 2pm-3pm, 计算机系 804（尹一通）&lt;br /&gt;
:* 周五 2pm-3pm, 计算机系 516（刘景铖）&lt;br /&gt;
:* &#039;&#039;&#039;QQ群&#039;&#039;&#039;: 974742320（申请加入需提供姓名、院系、学号）&lt;br /&gt;
&lt;br /&gt;
= Syllabus =&lt;br /&gt;
课程内容分为三大部分：&lt;br /&gt;
* &#039;&#039;&#039;经典概率论&#039;&#039;&#039;：概率空间、随机变量及其数字特征、多维与连续随机变量、极限定理等内容&lt;br /&gt;
* &#039;&#039;&#039;概率与计算&#039;&#039;&#039;：测度集中现象 (concentration of measure)、概率法 (the probabilistic method)、离散随机过程的相关专题&lt;br /&gt;
* &#039;&#039;&#039;数理统计&#039;&#039;&#039;：参数估计、假设检验、贝叶斯估计、线性回归等统计推断等概念&lt;br /&gt;
&lt;br /&gt;
对于第一和第二部分，要求清楚掌握基本概念，深刻理解关键的现象与规律以及背后的原理，并可以灵活运用所学方法求解相关问题。对于第三部分，要求熟悉数理统计的若干基本概念，以及典型的统计模型和统计推断问题。&lt;br /&gt;
&lt;br /&gt;
经过本课程的训练，力求使学生能够熟悉掌握概率的语言，并会利用概率思维来理解客观世界并对其建模，以及驾驭概率的数学工具来分析和求解专业问题。&lt;br /&gt;
&lt;br /&gt;
=== 教材与参考书 Course Materials ===&lt;br /&gt;
* &#039;&#039;&#039;[BT]&#039;&#039;&#039; 概率导论（第2版·修订版），[美]伯特瑟卡斯（Dimitri P.Bertsekas）[美]齐齐克利斯（John N.Tsitsiklis）著，郑忠国 童行伟 译，人民邮电出版社（2022）。&lt;br /&gt;
* &#039;&#039;&#039;[GS]&#039;&#039;&#039; &#039;&#039;Probability and Random Processes&#039;&#039;, by Geoffrey Grimmett and David Stirzaker; Oxford University Press; 4th edition (2020).&lt;br /&gt;
* &#039;&#039;&#039;[MU]&#039;&#039;&#039; &#039;&#039;Probability and Computing: Randomization and Probabilistic Techniques in Algorithms and Data Analysis&#039;&#039;, by Michael Mitzenmacher, Eli Upfal; Cambridge University Press; 2nd edition (2017).&lt;br /&gt;
&lt;br /&gt;
=== 成绩 Grading Policy ===&lt;br /&gt;
* 课程成绩：本课程将会有若干次作业和一次期末考试。最终成绩将由平时作业成绩和期末考试成绩综合得出。&lt;br /&gt;
* 迟交：如果有特殊的理由，无法按时完成作业，请提前联系授课老师，给出正当理由。否则迟交的作业将不被接受。&lt;br /&gt;
&lt;br /&gt;
=== &amp;lt;font color=red&amp;gt; 学术诚信 Academic Integrity &amp;lt;/font&amp;gt;===&lt;br /&gt;
学术诚信是所有从事学术活动的学生和学者最基本的职业道德底线，本课程将不遗余力的维护学术诚信规范，违反这一底线的行为将不会被容忍。&lt;br /&gt;
&lt;br /&gt;
作业完成的原则：署你名字的工作必须是你个人的贡献。在完成作业的过程中，允许讨论，前提是讨论的所有参与者均处于同等完成度。但关键想法的执行、以及作业文本的写作必须独立完成，并在作业中致谢（acknowledge）所有参与讨论的人。符合规则的讨论与致谢将不会影响得分。不允许其他任何形式的合作——尤其是与已经完成作业的同学“讨论”。&lt;br /&gt;
&lt;br /&gt;
本课程将对剽窃行为采取零容忍的态度。在完成作业过程中，对他人工作（出版物、互联网资料、其他人的作业等）直接的文本抄袭和对关键思想、关键元素的抄袭，按照 [http://www.acm.org/publications/policies/plagiarism_policy ACM Policy on Plagiarism]的解释，都将视为剽窃。剽窃者成绩将被取消。如果发现互相抄袭行为，&amp;lt;font color=red&amp;gt; 抄袭和被抄袭双方的成绩都将被取消&amp;lt;/font&amp;gt;。因此请主动防止自己的作业被他人抄袭。&lt;br /&gt;
&lt;br /&gt;
学术诚信影响学生个人的品行，也关乎整个教育系统的正常运转。为了一点分数而做出学术不端的行为，不仅使自己沦为一个欺骗者，也使他人的诚实努力失去意义。让我们一起努力维护一个诚信的环境。&lt;br /&gt;
&lt;br /&gt;
= Assignments =&lt;br /&gt;
*[[概率论与数理统计 (Spring 2025)/Problem Set 1|Problem Set 1]]  请在 2025/3/10 上课之前(10am UTC+8)提交到 [mailto:pr_nju@163.com pr_nju@163.com] (文件名为&#039;&amp;lt;font color=red &amp;gt;学号_姓名_A1.pdf&amp;lt;/font&amp;gt;&#039;).&lt;br /&gt;
** [[概率论与数理统计 (Spring 2025)/第一次作业提交名单|第一次作业提交名单]]&lt;br /&gt;
&lt;br /&gt;
*[[概率论与数理统计 (Spring 2025)/Problem Set 2|Problem Set 2]]  请在 2025/3/31 上课之前(10am UTC+8)提交到 [mailto:pr_nju@163.com pr_nju@163.com] (文件名为&#039;&amp;lt;font color=red &amp;gt;学号_姓名_A2.pdf&amp;lt;/font&amp;gt;&#039;).&lt;br /&gt;
** [[概率论与数理统计 (Spring 2025)/第二次作业提交名单|第二次作业提交名单]]&lt;br /&gt;
&lt;br /&gt;
= Lectures =&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2025/Intro.pdf 课程简介]&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2025/ProbSpace.pdf 概率空间]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 第1章&#039;&#039;&#039; 或 &#039;&#039;&#039;[GS] Chapter 1&#039;&#039;&#039;&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2025)/Entropy and volume of Hamming balls|Entropy and volume of Hamming balls]]&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2025/RandVar.pdf 随机变量]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 第2章&#039;&#039;&#039; 或 &#039;&#039;&#039;[GS] Chapter 2, Sections 3.1~3.5, 3.7&#039;&#039;&#039;&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[MU] Chapter 2&#039;&#039;&#039;&lt;br /&gt;
#* the [http://tcs.nju.edu.cn/slides/prob2024/discrete-pmf.nb &#039;&#039;&#039;Mathematica notebook file&#039;&#039;&#039;] for the PMFs of basic discrete distributions&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2025)/Average-case analysis of QuickSort|Average-case analysis of &#039;&#039;&#039;&#039;&#039;QuickSort&#039;&#039;&#039;&#039;&#039;]]&lt;br /&gt;
#* [https://www.bilibili.com/video/BV1ta411A7fp/ 高尔顿板（Galton board）视频] 和 [https://en.wikipedia.org/wiki/Galton_board 维基百科页面]&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2025/Deviation.pdf 矩与偏差]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[MU] Chapter 3&#039;&#039;&#039;&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 章节 2.4, 4.2, 4.3, 5.1&#039;&#039;&#039; 或 &#039;&#039;&#039;[GS] Sections 3.3, 3.6, 7.3&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= Concepts =&lt;br /&gt;
* [https://plato.stanford.edu/entries/probability-interpret/ Interpretations of probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/History_of_probability History of probability]&lt;br /&gt;
* Example problems:&lt;br /&gt;
** [https://dornsifecms.usc.edu/assets/sites/520/docs/VonNeumann-ams12p36-38.pdf von Neumann&#039;s Bernoulli factory] and other [https://peteroupc.github.io/bernoulli.html Bernoulli factory algorithms]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Boy_or_Girl_paradox Boy or Girl paradox]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Monty_Hall_problem Monty Hall problem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bertrand_paradox_(probability) Bertrand paradox]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hard_spheres Hard spheres model] and [https://en.wikipedia.org/wiki/Ising_model Ising model]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/PageRank &#039;&#039;PageRank&#039;&#039;] and stationary [https://en.wikipedia.org/wiki/Random_walk random walk]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Diffusion_process Diffusion process] and [https://en.wikipedia.org/wiki/Diffusion_model diffusion model]&lt;br /&gt;
*[https://en.wikipedia.org/wiki/Probability_space Probability space]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Sample_space Sample space]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Event_(probability_theory) Event] and [https://en.wikipedia.org/wiki/Σ-algebra &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-algebra]&lt;br /&gt;
** Kolmogorov&#039;s [https://en.wikipedia.org/wiki/Probability_axioms axioms of probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Discrete_uniform_distribution Classical] and [https://en.wikipedia.org/wiki/Geometric_probability goemetric probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Boole%27s_inequality Union bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Inclusion%E2%80%93exclusion_principle Inclusion-Exclusion principle]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Boole%27s_inequality#Bonferroni_inequalities Bonferroni inequalities]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Conditional_probability Conditional probability]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chain_rule_(probability) Chain rule]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_total_probability Law of total probability]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bayes%27_theorem Bayes&#039; law]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Independence_(probability_theory) Independence] &lt;br /&gt;
** [https://en.wikipedia.org/wiki/Pairwise_independence Pairwise independence]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Random_variable Random variable]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cumulative_distribution_function Cumulative distribution function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_mass_function Probability mass function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_density_function Probability density function]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Multivariate_random_variable Random vector]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Joint_probability_distribution Joint probability distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Conditional_probability_distribution Conditional probability distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Marginal_distribution Marginal distribution]&lt;br /&gt;
* Some &#039;&#039;&#039;discrete&#039;&#039;&#039; probability distributions&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bernoulli_trial Bernoulli trial] and [https://en.wikipedia.org/wiki/Bernoulli_distribution Bernoulli distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Discrete_uniform_distribution Discrete uniform distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Binomial_distribution Binomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Geometric_distribution Geometric distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Negative_binomial_distribution Negative binomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hypergeometric_distribution Hypergeometric distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Poisson_distribution Poisson distribution]&lt;br /&gt;
** and [https://en.wikipedia.org/wiki/List_of_probability_distributions#Discrete_distributions others]&lt;br /&gt;
* Balls into bins model&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Multinomial_distribution Multinomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Birthday_problem Birthday problem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Coupon_collector%27s_problem Coupon collector]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Balls_into_bins_problem Occupancy problem]&lt;br /&gt;
* Random graphs&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Erd%C5%91s%E2%80%93R%C3%A9nyi_model Erdős–Rényi random graph model]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Galton%E2%80%93Watson_process Galton–Watson branching process]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Expected_value Expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_the_unconscious_statistician Law of the unconscious statistician, &#039;&#039;LOTUS&#039;&#039;]&lt;br /&gt;
** [https://dlsun.github.io/probability/linearity.html Linearity of expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Conditional_expectation Conditional expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_total_expectation Law of total expectation]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Markov%27s_inequality Markov&#039;s inequality]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Chebyshev%27s_inequality Chebyshev&#039;s inequality]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Moment_(mathematics) Moment] related&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Central_moment Central moment]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Variance Variance]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Standard_deviation Standard deviation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Covariance Covariance]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Correlation Correlation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Correlation_does_not_imply_causation Correlation does not imply causation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Skewness Skewness]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Kurtosis Kurtosis]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Moment_problem Moment problem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Measure_(mathematics) Measure]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Borel_set Borel set]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Lebesgue_integration Lebesgue integration]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Measurable_function Measurable function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cantor_set Cantor set]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Non-measurable_set Non-measurable set]&lt;br /&gt;
*[https://en.wikipedia.org/wiki/Probability_density_function Probability density function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_distribution#Absolutely_continuous_probability_distribution Continuous probability distribution]&lt;br /&gt;
**[https://en.wikipedia.org/wiki/Conditional_probability_distribution#Conditional_continuous_distributions Conditional continuous distributions]&lt;br /&gt;
**[https://en.wikipedia.org/wiki/Convolution_of_probability_distributions Convolution of probability distributions] and [https://en.wikipedia.org/wiki/List_of_convolutions_of_probability_distributions List of convolutions of probability distributions]&lt;br /&gt;
* Some &#039;&#039;&#039;continuous&#039;&#039;&#039; probability distributions&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Continuous_uniform_distribution Continuous uniform distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Exponential_distribution Exponential distribution] and [https://en.wikipedia.org/wiki/Poisson_point_process Poisson point process]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Normal_distribution Normal (Gaussion) distribution]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Gaussian_function Gaussian function] and [https://en.wikipedia.org/wiki/Gaussian_integral Gaussian integral]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Standard_normal_table Standard normal table]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/68%E2%80%9395%E2%80%9399.7_rule 68–95–99.7 rule]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Sum_of_normally_distributed_random_variables Sum of normally distributed random variables]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Multivariate_normal_distribution Multivariate normal distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chi-squared_distribution Chi-squared distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Gamma_distribution Gamma distribution] and [https://en.wikipedia.org/wiki/Gamma_function Gamma function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Beta_distribution Beta distribution] and [https://en.wikipedia.org/wiki/Beta_function Beta function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cauchy_distribution Cauchy distribution]&lt;br /&gt;
** and [https://en.wikipedia.org/wiki/List_of_probability_distributions#Absolutely_continuous_distributions others]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Inverse_transform_sampling Inverse transform sampling]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stochastic_dominance Stochastic dominance] and [https://en.wikipedia.org/wiki/Coupling_(probability) Coupling] &lt;br /&gt;
* [https://en.wikipedia.org/wiki/Moment-generating_function Moment-generating function]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Convergence_of_random_variables Convergence of random variables]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Pointwise_convergence Pointwise convergence]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Convergence_of_measures#Weak_convergence_of_measures Weak convergence of measures]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Convergence_in_measure Convergence in measure]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Skorokhod%27s_representation_theorem Skorokhod&#039;s representation theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Continuous_mapping_theorem Continuous mapping theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Borel%E2%80%93Cantelli_lemma Borel–Cantelli lemma] and [https://en.wikipedia.org/wiki/Zero%E2%80%93one_law zero-one laws]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Law_of_large_numbers Law of large numbers]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Central_limit_theorem Central limit theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/De_Moivre%E2%80%93Laplace_theorem De Moivre–Laplace theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Berry%E2%80%93Esseen_theorem Berry–Esseen theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Characteristic_function_(probability_theory) Characteristic function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Lévy%27s_continuity_theorem Lévy&#039;s continuity theorem]&lt;br /&gt;
* Concentration of measures:&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Moment-generating_function Moment-generating function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chernoff_bound Chernoff bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hoeffding%27s_inequality Hoeffding&#039;s bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Doob_martingale#McDiarmid&#039;s_inequality McDiarmid&#039;s inequality]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Doob_martingale Doob martingale] and [https://en.wikipedia.org/wiki/Azuma%27s_inequality Azuma&#039;s inequality]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Sub-Gaussian_distribution Sub-Gaussian distribution]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stochastic_process Stochastic process]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stopping_time Stopping time]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Martingale_(probability_theory) Martingale]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Optional_stopping_theorem Optional stopping theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Wald%27s_equation Wald&#039;s equation]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Markov_chain Markov chain]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Markov_property Markov property] and [https://en.wikipedia.org/wiki/Memorylessness memorylessness]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Stochastic_matrix Stochastic matrix]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Stationary_distribution Stationary distribution]&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)&amp;diff=12972</id>
		<title>概率论与数理统计 (Spring 2025)</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)&amp;diff=12972"/>
		<updated>2025-03-16T13:49:45Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: /* Assignments */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Infobox&lt;br /&gt;
|name         = Infobox&lt;br /&gt;
|bodystyle    = &lt;br /&gt;
|title        = &amp;lt;font size=3&amp;gt;&#039;&#039;&#039;概率论与数理统计&#039;&#039;&#039;&amp;lt;br&amp;gt;&lt;br /&gt;
&#039;&#039;&#039;Probability Theory&#039;&#039;&#039; &amp;lt;br&amp;gt; &amp;amp; &#039;&#039;&#039;Mathematical Statistics&#039;&#039;&#039;&amp;lt;/font&amp;gt;&lt;br /&gt;
|titlestyle   = &lt;br /&gt;
&lt;br /&gt;
|image        = &lt;br /&gt;
|imagestyle   = &lt;br /&gt;
|caption      = &lt;br /&gt;
|captionstyle = &lt;br /&gt;
|headerstyle  = background:#ccf;&lt;br /&gt;
|labelstyle   = background:#ddf;&lt;br /&gt;
|datastyle    = &lt;br /&gt;
&lt;br /&gt;
|header1 =Instructor&lt;br /&gt;
|label1  = &lt;br /&gt;
|data1   = &lt;br /&gt;
|header2 = &lt;br /&gt;
|label2  = &lt;br /&gt;
|data2   = &#039;&#039;&#039;尹一通&#039;&#039;&#039;&lt;br /&gt;
|header3 = &lt;br /&gt;
|label3  = Email&lt;br /&gt;
|data3   = yinyt@nju.edu.cn  &lt;br /&gt;
|header4 =&lt;br /&gt;
|label4  = office&lt;br /&gt;
|data4   = 计算机系 804&lt;br /&gt;
|header5 = &lt;br /&gt;
|label5  = &lt;br /&gt;
|data5   = &#039;&#039;&#039;刘景铖&#039;&#039;&#039;&lt;br /&gt;
|header6 = &lt;br /&gt;
|label6  = Email&lt;br /&gt;
|data6   = liu@nju.edu.cn  &lt;br /&gt;
|header7 =&lt;br /&gt;
|label7  = office&lt;br /&gt;
|data7   = 计算机系 516&lt;br /&gt;
|header8 = Class&lt;br /&gt;
|label8  = &lt;br /&gt;
|data8   = &lt;br /&gt;
|header9 =&lt;br /&gt;
|label9  = Class meeting&lt;br /&gt;
|data9   = Monday, 10am-12pm&amp;lt;br&amp;gt;&lt;br /&gt;
Wednesday (单), 10am-12pm&amp;lt;br&amp;gt;&lt;br /&gt;
仙Ⅰ-204&lt;br /&gt;
|header10=&lt;br /&gt;
|label10 = Office hour&lt;br /&gt;
|data10  = Friday, 2pm-3pm &amp;lt;br&amp;gt;计算机系 804（尹一通）&amp;lt;br&amp;gt;计算机系 516（刘景铖）&lt;br /&gt;
|header11= Textbook&lt;br /&gt;
|label11 = &lt;br /&gt;
|data11  = &lt;br /&gt;
|header12=&lt;br /&gt;
|label12 = &lt;br /&gt;
|data12  = [[File:概率导论.jpeg|border|100px]]&lt;br /&gt;
|header13=&lt;br /&gt;
|label13 = &lt;br /&gt;
|data13  = &#039;&#039;&#039;概率导论&#039;&#039;&#039;（第2版·修订版）&amp;lt;br&amp;gt; Dimitri P. Bertsekas and John N. Tsitsiklis&amp;lt;br&amp;gt; 郑忠国 童行伟 译；人民邮电出版社 (2022)&lt;br /&gt;
|header14=&lt;br /&gt;
|label14 = &lt;br /&gt;
|data14  = [[File:Grimmett_probability.jpg|border|100px]]&lt;br /&gt;
|header15=&lt;br /&gt;
|label15 = &lt;br /&gt;
|data15  = &#039;&#039;&#039;Probability and Random Processes&#039;&#039;&#039; (4E) &amp;lt;br&amp;gt; Geoffrey Grimmett and David Stirzaker &amp;lt;br&amp;gt;  Oxford University Press (2020)&lt;br /&gt;
|header16=&lt;br /&gt;
|label16 = &lt;br /&gt;
|data16  = [[File:Probability_and_Computing_2ed.jpg|border|100px]]&lt;br /&gt;
|header17=&lt;br /&gt;
|label17 = &lt;br /&gt;
|data17  = &#039;&#039;&#039;Probability and Computing&#039;&#039;&#039; (2E) &amp;lt;br&amp;gt; Michael Mitzenmacher and Eli Upfal &amp;lt;br&amp;gt;   Cambridge University Press (2017)&lt;br /&gt;
|belowstyle = background:#ddf;&lt;br /&gt;
|below = &lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
This is the webpage for the &#039;&#039;Probability Theory and Mathematical Statistics&#039;&#039; (概率论与数理统计) class of Spring 2025. Students who take this class should check this page periodically for content updates and new announcements. &lt;br /&gt;
&lt;br /&gt;
= Announcement =&lt;br /&gt;
* TBA&lt;br /&gt;
&lt;br /&gt;
= Course info =&lt;br /&gt;
* &#039;&#039;&#039;Instructor &#039;&#039;&#039;: &lt;br /&gt;
:* [http://tcs.nju.edu.cn/yinyt/ 尹一通]：[mailto:yinyt@nju.edu.cn &amp;lt;yinyt@nju.edu.cn&amp;gt;]，计算机系 804 &lt;br /&gt;
:* [https://liuexp.github.io 刘景铖]：[mailto:liu@nju.edu.cn &amp;lt;liu@nju.edu.cn&amp;gt;]，计算机系 516 &lt;br /&gt;
* &#039;&#039;&#039;Teaching assistant&#039;&#039;&#039;:&lt;br /&gt;
** [https://sites.google.com/view/xinyuanzhang 张昕渊]：[mailto:zhangxy@smail.nju.edu.cn &amp;lt;zhangxy@smail.nju.edu.cn&amp;gt;]，计算机系 410&lt;br /&gt;
** 邹宗瑞：[mailto:zou.zongrui@smail.nju.edu.cn &amp;lt;zou.zongrui@smail.nju.edu.cn&amp;gt;]，计算机系 410&lt;br /&gt;
* &#039;&#039;&#039;Class meeting&#039;&#039;&#039;:&lt;br /&gt;
** 周一：10am-12pm，仙Ⅰ-204&lt;br /&gt;
** 周三（单）：10am-12pm，仙Ⅰ-204&lt;br /&gt;
* &#039;&#039;&#039;Office hour&#039;&#039;&#039;: &lt;br /&gt;
:* 周五 2pm-3pm, 计算机系 804（尹一通）&lt;br /&gt;
:* 周五 2pm-3pm, 计算机系 516（刘景铖）&lt;br /&gt;
:* &#039;&#039;&#039;QQ群&#039;&#039;&#039;: 974742320（申请加入需提供姓名、院系、学号）&lt;br /&gt;
&lt;br /&gt;
= Syllabus =&lt;br /&gt;
课程内容分为三大部分：&lt;br /&gt;
* &#039;&#039;&#039;经典概率论&#039;&#039;&#039;：概率空间、随机变量及其数字特征、多维与连续随机变量、极限定理等内容&lt;br /&gt;
* &#039;&#039;&#039;概率与计算&#039;&#039;&#039;：测度集中现象 (concentration of measure)、概率法 (the probabilistic method)、离散随机过程的相关专题&lt;br /&gt;
* &#039;&#039;&#039;数理统计&#039;&#039;&#039;：参数估计、假设检验、贝叶斯估计、线性回归等统计推断等概念&lt;br /&gt;
&lt;br /&gt;
对于第一和第二部分，要求清楚掌握基本概念，深刻理解关键的现象与规律以及背后的原理，并可以灵活运用所学方法求解相关问题。对于第三部分，要求熟悉数理统计的若干基本概念，以及典型的统计模型和统计推断问题。&lt;br /&gt;
&lt;br /&gt;
经过本课程的训练，力求使学生能够熟悉掌握概率的语言，并会利用概率思维来理解客观世界并对其建模，以及驾驭概率的数学工具来分析和求解专业问题。&lt;br /&gt;
&lt;br /&gt;
=== 教材与参考书 Course Materials ===&lt;br /&gt;
* &#039;&#039;&#039;[BT]&#039;&#039;&#039; 概率导论（第2版·修订版），[美]伯特瑟卡斯（Dimitri P.Bertsekas）[美]齐齐克利斯（John N.Tsitsiklis）著，郑忠国 童行伟 译，人民邮电出版社（2022）。&lt;br /&gt;
* &#039;&#039;&#039;[GS]&#039;&#039;&#039; &#039;&#039;Probability and Random Processes&#039;&#039;, by Geoffrey Grimmett and David Stirzaker; Oxford University Press; 4th edition (2020).&lt;br /&gt;
* &#039;&#039;&#039;[MU]&#039;&#039;&#039; &#039;&#039;Probability and Computing: Randomization and Probabilistic Techniques in Algorithms and Data Analysis&#039;&#039;, by Michael Mitzenmacher, Eli Upfal; Cambridge University Press; 2nd edition (2017).&lt;br /&gt;
&lt;br /&gt;
=== 成绩 Grading Policy ===&lt;br /&gt;
* 课程成绩：本课程将会有若干次作业和一次期末考试。最终成绩将由平时作业成绩和期末考试成绩综合得出。&lt;br /&gt;
* 迟交：如果有特殊的理由，无法按时完成作业，请提前联系授课老师，给出正当理由。否则迟交的作业将不被接受。&lt;br /&gt;
&lt;br /&gt;
=== &amp;lt;font color=red&amp;gt; 学术诚信 Academic Integrity &amp;lt;/font&amp;gt;===&lt;br /&gt;
学术诚信是所有从事学术活动的学生和学者最基本的职业道德底线，本课程将不遗余力的维护学术诚信规范，违反这一底线的行为将不会被容忍。&lt;br /&gt;
&lt;br /&gt;
作业完成的原则：署你名字的工作必须是你个人的贡献。在完成作业的过程中，允许讨论，前提是讨论的所有参与者均处于同等完成度。但关键想法的执行、以及作业文本的写作必须独立完成，并在作业中致谢（acknowledge）所有参与讨论的人。符合规则的讨论与致谢将不会影响得分。不允许其他任何形式的合作——尤其是与已经完成作业的同学“讨论”。&lt;br /&gt;
&lt;br /&gt;
本课程将对剽窃行为采取零容忍的态度。在完成作业过程中，对他人工作（出版物、互联网资料、其他人的作业等）直接的文本抄袭和对关键思想、关键元素的抄袭，按照 [http://www.acm.org/publications/policies/plagiarism_policy ACM Policy on Plagiarism]的解释，都将视为剽窃。剽窃者成绩将被取消。如果发现互相抄袭行为，&amp;lt;font color=red&amp;gt; 抄袭和被抄袭双方的成绩都将被取消&amp;lt;/font&amp;gt;。因此请主动防止自己的作业被他人抄袭。&lt;br /&gt;
&lt;br /&gt;
学术诚信影响学生个人的品行，也关乎整个教育系统的正常运转。为了一点分数而做出学术不端的行为，不仅使自己沦为一个欺骗者，也使他人的诚实努力失去意义。让我们一起努力维护一个诚信的环境。&lt;br /&gt;
&lt;br /&gt;
= Assignments =&lt;br /&gt;
*[[概率论与数理统计 (Spring 2025)/Problem Set 1|Problem Set 1]]  请在 2025/3/10 上课之前(10am UTC+8)提交到 [mailto:pr_nju@163.com pr_nju@163.com] (文件名为&#039;&amp;lt;font color=red &amp;gt;学号_姓名_A1.pdf&amp;lt;/font&amp;gt;&#039;).&lt;br /&gt;
** [[概率论与数理统计 (Spring 2025)/第一次作业提交名单|第一次作业提交名单]]&lt;br /&gt;
&lt;br /&gt;
*[[概率论与数理统计 (Spring 2025)/Problem Set 2|Problem Set 2]]  请在 2025/3/31 上课之前(10am UTC+8)提交到 [mailto:pr_nju@163.com pr_nju@163.com] (文件名为&#039;&amp;lt;font color=red &amp;gt;学号_姓名_A2.pdf&amp;lt;/font&amp;gt;&#039;).&lt;br /&gt;
&lt;br /&gt;
= Lectures =&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2025/Intro.pdf 课程简介]&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2025/ProbSpace.pdf 概率空间]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 第1章&#039;&#039;&#039; 或 &#039;&#039;&#039;[GS] Chapter 1&#039;&#039;&#039;&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2025)/Entropy and volume of Hamming balls|Entropy and volume of Hamming balls]]&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2025/RandVar.pdf 随机变量]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 第2章&#039;&#039;&#039; 或 &#039;&#039;&#039;[GS] Chapter 2, Sections 3.1~3.5, 3.7&#039;&#039;&#039;&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[MU] Chapter 2&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= Concepts =&lt;br /&gt;
* [https://plato.stanford.edu/entries/probability-interpret/ Interpretations of probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/History_of_probability History of probability]&lt;br /&gt;
* Example problems:&lt;br /&gt;
** [https://dornsifecms.usc.edu/assets/sites/520/docs/VonNeumann-ams12p36-38.pdf von Neumann&#039;s Bernoulli factory] and other [https://peteroupc.github.io/bernoulli.html Bernoulli factory algorithms]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Boy_or_Girl_paradox Boy or Girl paradox]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Monty_Hall_problem Monty Hall problem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bertrand_paradox_(probability) Bertrand paradox]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hard_spheres Hard spheres model] and [https://en.wikipedia.org/wiki/Ising_model Ising model]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/PageRank &#039;&#039;PageRank&#039;&#039;] and stationary [https://en.wikipedia.org/wiki/Random_walk random walk]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Diffusion_process Diffusion process] and [https://en.wikipedia.org/wiki/Diffusion_model diffusion model]&lt;br /&gt;
*[https://en.wikipedia.org/wiki/Probability_space Probability space]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Sample_space Sample space]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Event_(probability_theory) Event] and [https://en.wikipedia.org/wiki/Σ-algebra &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-algebra]&lt;br /&gt;
** Kolmogorov&#039;s [https://en.wikipedia.org/wiki/Probability_axioms axioms of probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Discrete_uniform_distribution Classical] and [https://en.wikipedia.org/wiki/Geometric_probability goemetric probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Boole%27s_inequality Union bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Inclusion%E2%80%93exclusion_principle Inclusion-Exclusion principle]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Boole%27s_inequality#Bonferroni_inequalities Bonferroni inequalities]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Conditional_probability Conditional probability]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chain_rule_(probability) Chain rule]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_total_probability Law of total probability]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bayes%27_theorem Bayes&#039; law]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Independence_(probability_theory) Independence] &lt;br /&gt;
** [https://en.wikipedia.org/wiki/Pairwise_independence Pairwise independence]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Random_variable Random variable]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cumulative_distribution_function Cumulative distribution function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_mass_function Probability mass function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_density_function Probability density function]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Multivariate_random_variable Random vector]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Joint_probability_distribution Joint probability distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Conditional_probability_distribution Conditional probability distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Marginal_distribution Marginal distribution]&lt;br /&gt;
* Some &#039;&#039;&#039;discrete&#039;&#039;&#039; probability distributions&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bernoulli_trial Bernoulli trial] and [https://en.wikipedia.org/wiki/Bernoulli_distribution Bernoulli distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Discrete_uniform_distribution Discrete uniform distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Binomial_distribution Binomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Geometric_distribution Geometric distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Negative_binomial_distribution Negative binomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hypergeometric_distribution Hypergeometric distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Poisson_distribution Poisson distribution]&lt;br /&gt;
** and [https://en.wikipedia.org/wiki/List_of_probability_distributions#Discrete_distributions others]&lt;br /&gt;
* Balls into bins model&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Multinomial_distribution Multinomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Birthday_problem Birthday problem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Coupon_collector%27s_problem Coupon collector]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Balls_into_bins_problem Occupancy problem]&lt;br /&gt;
* Random graphs&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Erd%C5%91s%E2%80%93R%C3%A9nyi_model Erdős–Rényi random graph model]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Galton%E2%80%93Watson_process Galton–Watson branching process]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Expected_value Expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_the_unconscious_statistician Law of the unconscious statistician, &#039;&#039;LOTUS&#039;&#039;]&lt;br /&gt;
** [https://dlsun.github.io/probability/linearity.html Linearity of expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Conditional_expectation Conditional expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_total_expectation Law of total expectation]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Markov%27s_inequality Markov&#039;s inequality]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Chebyshev%27s_inequality Chebyshev&#039;s inequality]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Moment_(mathematics) Moment] related&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Central_moment Central moment]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Variance Variance]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Standard_deviation Standard deviation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Covariance Covariance]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Correlation Correlation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Correlation_does_not_imply_causation Correlation does not imply causation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Skewness Skewness]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Kurtosis Kurtosis]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Moment_problem Moment problem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Measure_(mathematics) Measure]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Borel_set Borel set]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Lebesgue_integration Lebesgue integration]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Measurable_function Measurable function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cantor_set Cantor set]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Non-measurable_set Non-measurable set]&lt;br /&gt;
*[https://en.wikipedia.org/wiki/Probability_density_function Probability density function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_distribution#Absolutely_continuous_probability_distribution Continuous probability distribution]&lt;br /&gt;
**[https://en.wikipedia.org/wiki/Conditional_probability_distribution#Conditional_continuous_distributions Conditional continuous distributions]&lt;br /&gt;
**[https://en.wikipedia.org/wiki/Convolution_of_probability_distributions Convolution of probability distributions] and [https://en.wikipedia.org/wiki/List_of_convolutions_of_probability_distributions List of convolutions of probability distributions]&lt;br /&gt;
* Some &#039;&#039;&#039;continuous&#039;&#039;&#039; probability distributions&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Continuous_uniform_distribution Continuous uniform distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Exponential_distribution Exponential distribution] and [https://en.wikipedia.org/wiki/Poisson_point_process Poisson point process]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Normal_distribution Normal (Gaussion) distribution]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Gaussian_function Gaussian function] and [https://en.wikipedia.org/wiki/Gaussian_integral Gaussian integral]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Standard_normal_table Standard normal table]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/68%E2%80%9395%E2%80%9399.7_rule 68–95–99.7 rule]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Sum_of_normally_distributed_random_variables Sum of normally distributed random variables]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Multivariate_normal_distribution Multivariate normal distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chi-squared_distribution Chi-squared distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Gamma_distribution Gamma distribution] and [https://en.wikipedia.org/wiki/Gamma_function Gamma function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Beta_distribution Beta distribution] and [https://en.wikipedia.org/wiki/Beta_function Beta function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cauchy_distribution Cauchy distribution]&lt;br /&gt;
** and [https://en.wikipedia.org/wiki/List_of_probability_distributions#Absolutely_continuous_distributions others]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Inverse_transform_sampling Inverse transform sampling]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stochastic_dominance Stochastic dominance] and [https://en.wikipedia.org/wiki/Coupling_(probability) Coupling] &lt;br /&gt;
* [https://en.wikipedia.org/wiki/Moment-generating_function Moment-generating function]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Convergence_of_random_variables Convergence of random variables]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Pointwise_convergence Pointwise convergence]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Convergence_of_measures#Weak_convergence_of_measures Weak convergence of measures]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Convergence_in_measure Convergence in measure]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Skorokhod%27s_representation_theorem Skorokhod&#039;s representation theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Continuous_mapping_theorem Continuous mapping theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Borel%E2%80%93Cantelli_lemma Borel–Cantelli lemma] and [https://en.wikipedia.org/wiki/Zero%E2%80%93one_law zero-one laws]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Law_of_large_numbers Law of large numbers]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Central_limit_theorem Central limit theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/De_Moivre%E2%80%93Laplace_theorem De Moivre–Laplace theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Berry%E2%80%93Esseen_theorem Berry–Esseen theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Characteristic_function_(probability_theory) Characteristic function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Lévy%27s_continuity_theorem Lévy&#039;s continuity theorem]&lt;br /&gt;
* Concentration of measures:&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Moment-generating_function Moment-generating function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chernoff_bound Chernoff bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hoeffding%27s_inequality Hoeffding&#039;s bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Doob_martingale#McDiarmid&#039;s_inequality McDiarmid&#039;s inequality]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Doob_martingale Doob martingale] and [https://en.wikipedia.org/wiki/Azuma%27s_inequality Azuma&#039;s inequality]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Sub-Gaussian_distribution Sub-Gaussian distribution]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stochastic_process Stochastic process]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stopping_time Stopping time]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Martingale_(probability_theory) Martingale]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Optional_stopping_theorem Optional stopping theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Wald%27s_equation Wald&#039;s equation]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Markov_chain Markov chain]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Markov_property Markov property] and [https://en.wikipedia.org/wiki/Memorylessness memorylessness]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Stochastic_matrix Stochastic matrix]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Stationary_distribution Stationary distribution]&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12969</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 2</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12969"/>
		<updated>2025-03-14T08:10:13Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: /* Problem 1 (Warm-up problems, 16 points) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 2==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Warm-up problems, 16 points) ==&lt;br /&gt;
* [&#039;&#039;&#039;Cumulative distribution function (CDF) (I)&#039;&#039;&#039;] Express the distribution functions of [math]X^+ = \max\{0,X\}[/math], [math]X^- = -\min\{0,X\}[/math], [math]|X|=X^+ + X^-[/math], [math]-X[/math], in terms of the distribution function [math]F[/math] of the random variable [math]X[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Cumulative distribution function (CDF) (II)&#039;&#039;&#039;] Let [math]X[/math] be a random variable with distribution function [math]\max(0,\min(1,x))[/math]. Let [math]F[/math] be a distribution function which is continuous and strictly increasing. Show that [math]Y=F^{-1}(X)[/math] is a random variable with distribution function [math]F[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Probability density function (PDF)&#039;&#039;&#039;] We toss &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; coins, and each one shows heads with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;, independently of each of the others. Each coin which shows head is tossed again. (If the coin shows tail, it won&#039;t be tossed again.) Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of heads resulting from the &amp;lt;strong&amp;gt;second&amp;lt;/strong&amp;gt; round of tosses, and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the number of heads resulting from &amp;lt;strong&amp;gt;all&amp;lt;/strong&amp;gt; tosses, which includes the first and (possible) second round of each toss.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the PDF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{E}[Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Let &amp;lt;math&amp;gt;p_X&amp;lt;/math&amp;gt; be the PDF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, show that [math]p_X(k-1)p_X(k+1)\leq [p_X(k)]^2[/math] for &amp;lt;math&amp;gt;1\leq k \leq n-1&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
* [&#039;&#039;&#039;Independence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;1\leq r\leq n&amp;lt;/math&amp;gt; be independent random variables which are symmetric about &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;; that is, &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-X_r&amp;lt;/math&amp;gt; have the same distributions. Show that, for all &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}[S_n \geq x] = \mathbf{Pr}[S_n \leq -x]&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;S_n = \sum_{r=1}^n X_r&amp;lt;/math&amp;gt;. Is the conclusion true without the assumtion of independence?&lt;br /&gt;
* [&#039;&#039;&#039;Dependence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be discrete random variables with joint mass function &amp;lt;math&amp;gt;f(x,y) = \frac{C}{(x+y-1)(x+y)(x+y+1)}&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;x,y \in \mathbb{N}_+&amp;lt;/math&amp;gt; (in other words, &amp;lt;math&amp;gt;x,y = 1,2,3,\cdots&amp;lt;/math&amp;gt;). Find (1) the value of &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;, (2) marginal mass function of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and (3) &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt;.&lt;br /&gt;
* [&#039;&#039;&#039;Expectation&#039;&#039;&#039;] It is required to place in order &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; books &amp;lt;math&amp;gt;B_1, B_2, \cdots, B_n&amp;lt;/math&amp;gt; on a library shelf in such way that readers searching from left to right waste as little time as possible on average. Assuming that a random reader requires book &amp;lt;math&amp;gt;B_i&amp;lt;/math&amp;gt; with probability &amp;lt;math&amp;gt;p_i&amp;lt;/math&amp;gt;, find the ordering of the books which minimizes the expected number of titles examined by a random reader before discovery of the required book.&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Law of total expectation&amp;lt;/strong&amp;gt;] Let [math]X \sim \mathrm{Geom}(p)[/math] for some parameter [math]p \in (0,1)[/math]. Calculate [math]\mathbf{E}[X][/math] using the law of total expectation.&lt;br /&gt;
&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Entropy of discrete random variable&amp;lt;/strong&amp;gt;] Let [math]X[/math] be a discrete random variable with range of values [math][N] = \{1,2,\ldots,N\}[/math] and probability mass function [math]p[/math]. Define [math]H(X) = -\sum_{n \ge 1} p(n) \log p(n)[/math] with convention [math]0\log 0 = 0[/math]. Prove that [math]H(X) \le \log N[/math] using Jensen&#039;s inequality.&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Discrete random variable, 14 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (I)&amp;lt;/strong&amp;gt;] Every package of some intrinsically dull commodity includes a small and exciting plastic object. There are [math]c[/math] different types of object, and each package is equally likely to contain any given type. You buy one package each day.&amp;lt;/p&amp;gt;&lt;br /&gt;
    &amp;lt;ol&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse between the acquisitions of the [math]j[/math]-th new type of object and the [math](j + 1)[/math]-th new type.&amp;lt;/li&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse before you have a full set of objects.&amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (II)&amp;lt;/strong&amp;gt;] Prove that geometric distribution is the only discrete memoryless distribution with range values &amp;lt;math&amp;gt;\mathbb{N}_+&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;n_1,n_2 \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;0 \le p \le 1&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Bin}(n_1,p),Y \sim \mathrm{Bin}(n_2,p)&amp;lt;/math&amp;gt; be independent random variables. Prove that &amp;lt;math&amp;gt;X+Y \sim \mathrm{Bin}(n_1+n_2,p)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Negative binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; follows the negative binomial distribution with parameter &amp;lt;math&amp;gt;r \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt;. Calculate &amp;lt;math&amp;gt;\mathbf{Var}[X] = \mathbf{E}[X^2] - \left(\mathbf{E}[X]\right)^2&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Hypergeometric distribution&amp;lt;/strong&amp;gt;] An urn contains [math]N[/math] balls, [math]b[/math] of which are blue and [math]r = N -b[/math] of which are red. A random sample of [math]n[/math] balls is drawn &amp;lt;strong&amp;gt;without replacement&amp;lt;/strong&amp;gt; (无放回) from the urn. Let [math]B[/math] the number of blue balls in this sample. Show that if [math]N, b[/math], and [math]r[/math] approach [math]+\infty[/math] in such a way that [math]b/N \rightarrow  p[/math] and [math]r/N \rightarrow 1 - p[/math], then [math]\mathbf{Pr}(B = k) \rightarrow {n\choose k}p^k(1-p)^{n-k}[/math] for [math]0\leq k \leq n[/math].&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Poisson distribution&amp;lt;/strong&amp;gt;] In your pocket is a random number &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; of coins, where &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; has the Poisson distribution with parameter &amp;lt;math&amp;gt;\lambda&amp;lt;/math&amp;gt;. You toss each coin once, with heads showing with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; each time. Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the (random) number of heads outcomes and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the (also random) number of tails.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the joint mass function of &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find PDF of the marginal distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; in &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;. Are &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; independent?&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Conditional distribution&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;\lambda,\mu &amp;gt; 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;n \in \mathbb{N}&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Pois}(\lambda), Y \sim \mathrm{Pois}(\mu)&amp;lt;/math&amp;gt; be independent random variables. Find out the conditional distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, given &amp;lt;math&amp;gt;X+Y = n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Linearity of Expectation, 12 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Streak&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Suppose we flip a fair coin &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; times independently to obtain a sequence of flips &amp;lt;math&amp;gt;X_1, X_2, \ldots , X_n&amp;lt;/math&amp;gt;. A streak of flips is a consecutive subsequence of flips that are all the same. For example, if &amp;lt;math&amp;gt;X_3&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;X_4&amp;lt;/math&amp;gt;, and &amp;lt;math&amp;gt;X_5&amp;lt;/math&amp;gt; are all heads, there is a streak of length &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; starting at the third lip. (If &amp;lt;math&amp;gt;X_6&amp;lt;/math&amp;gt; is also heads, then there is also a streak of length &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; starting at the third lip.) Find the expected number of streaks of length &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; for some integer &amp;lt;math&amp;gt;k \ge 1&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Number of cycles&amp;lt;/strong&amp;gt;] &lt;br /&gt;
At a banquet, there are &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people who shake hands according to the following process: In each round, two idle hands are randomly selected and shaken (&amp;lt;strong&amp;gt;these two hands are no longer idle&amp;lt;/strong&amp;gt;). After &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds, there will be no idle hands left, and the &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people will form several cycles. For example, when &amp;lt;math&amp;gt;n=3&amp;lt;/math&amp;gt;, the following situation may occur: the left and right hands of the first person are held together, the left hand of the second person and the right hand of the third person are held together, and the right hand of the second person and the left hand of the third person are held together. In this case, three people form two cycles. How many cycles are expected to be formed after &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds?&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Number of operations&amp;lt;/strong&amp;gt;]&lt;br /&gt;
We have a directed graph &amp;lt;math&amp;gt;G=(V,E)&amp;lt;/math&amp;gt; without self-loops and multi-edges.&lt;br /&gt;
&lt;br /&gt;
Until the graph becomes empty, repeat the following operation:&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
Choose one (unerased) vertex uniformly at random (independently from the previous choices). Then, erase that vertex and all vertices that are reachable from the chosen vertex by traversing some edges. Erasing a vertex will also erase the edges incident to it.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;R_v&amp;lt;/math&amp;gt; denotes the number of vertices from which vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt; is reachable. Prove that the expected value of the number of operations equals to &amp;lt;math&amp;gt;\sum_{v \in V} \frac{1}{R_v}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Expected Mex&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X_1,X_2,\ldots,X_{100} \sim \mathrm{Geo}(1/2)&amp;lt;/math&amp;gt; be independent random variables. Compute &amp;lt;math&amp;gt;\mathbf{E}[\mathrm{mex}(X_1,X_2,\ldots,X_{100})]&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\mathrm{mex}(a_1,a_2,\ldots,a_n)&amp;lt;/math&amp;gt; is the smallest positive integer that does not appear in &amp;lt;math&amp;gt;a_1,a_2,\ldots,a_n&amp;lt;/math&amp;gt;. Your answer is considered correct if the absolute error does not exceed &amp;lt;math&amp;gt;10^{-6}&amp;lt;/math&amp;gt;. (Hint: Although mathematics will help you arrive at elegant and efficient methods, the use of a computer and programming skills will be required.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 4 (Probability meets graph theory) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random social networks&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G = (V, E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph without isolating vertex. &lt;br /&gt;
Let &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; be the degree of vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Let &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be a uniformly chosen vertex, and &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; a uniformly chosen neighbor of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Show that &amp;lt;math&amp;gt;\mathbf{E}[d_Z] \geq \mathbf{E}[d_Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Interpret this inequality in the context of social networks, in which the vertices represent people, and the edges represent friendship.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Turán&#039;s Theorem&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G=(V,E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph, and write &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; for the degree of the vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Use probablistic method to prove that &amp;lt;math&amp;gt;\alpha(G) \ge \sum_{v \in V} \frac{1}{d_v + 1}&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\alpha(G)&amp;lt;/math&amp;gt; is the size of a maximum independent set. (Hint: Consider the following random procedure for generating an independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; from a graph with vertex set &amp;lt;math&amp;gt;V&amp;lt;/math&amp;gt;: First, generate a random permutation of the vertices, denoted as &amp;lt;math&amp;gt;v_1,v_2,\ldots,v_n&amp;lt;/math&amp;gt;. Then, construct the independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; as follows: For each vertex &amp;lt;math&amp;gt;v_i \in V&amp;lt;/math&amp;gt;, add &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; if and only if none of its predecessors in the permutation, i.e., &amp;lt;math&amp;gt;v_1,\ldots,v_{i-1}&amp;lt;/math&amp;gt;, are neighbors of &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Tournament&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Prove that there is a [https://en.wikipedia.org/wiki/Tournament_(graph_theory) tournament] &amp;lt;math&amp;gt;T&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; players and at least &amp;lt;math&amp;gt;n!/2^{n-1}&amp;lt;/math&amp;gt;Hamiltonian paths.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 5 (1D random walk, 8 points) ==&lt;br /&gt;
Let &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt; be a constant, and &amp;lt;math&amp;gt;\{X_n\}_{n \ge 1}&amp;lt;/math&amp;gt; be independent Bernoulli trials with successful probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;.&lt;br /&gt;
Define &amp;lt;math&amp;gt;S_n = 2\sum_{i=1}^n X_i - n&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;S_0 = 0&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Range of random walk&amp;lt;/strong&amp;gt;] The range &amp;lt;math&amp;gt;R_n&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;S_0, S_1, \ldots, S_n&amp;lt;/math&amp;gt; is defined as the number of distinct values taken by the sequence. Show that &amp;lt;math&amp;gt;\mathbf{Pr}\left(R_n = R_{n-1}+1\right) = \mathbf{Pr}\left(\forall 1 \le i \le n, S_i \neq 0\right)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, and deduce that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n]\to \mathbf{Pr}(\forall i \ge 1, S_i \neq 0)&amp;lt;/math&amp;gt;. Hence show that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n] \to |2p-1|&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Symmetric 1D random walk (III)&amp;lt;/strong&amp;gt;] Suppose &amp;lt;math&amp;gt;p = \frac{1}{2}&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{E}[|S_n|] = \Theta(\sqrt{n})&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12968</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 2</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12968"/>
		<updated>2025-03-14T08:06:44Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: /* Problem 1 (Warm-up problems, 16 points) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 2==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Warm-up problems, 16 points) ==&lt;br /&gt;
* [&#039;&#039;&#039;Cumulative distribution function (CDF) (I)&#039;&#039;&#039;] Express the distribution functions of [math]X^+ = \max\{0,X\}[/math], [math]X^- = -\min\{0,X\}[/math], [math]|X|=X^+ + X^-[/math], [math]-X[/math], in terms of the distribution function [math]F[/math] of the random variable [math]X[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Cumulative distribution function (CDF) (II)&#039;&#039;&#039;] Let [math]X[/math] be a random variable with distribution function [math]\max(0,\min(1,x))[/math]. Let [math]F[/math] be a distribution function which is continuous and strictly increasing. Show that [math]Y=F^{-1}(X)[/math] is a random variable with distribution function [math]F[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Probability density function (PDF)&#039;&#039;&#039;] We toss &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; coins, and each one shows heads with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;, independently of each of the others. Each coin which shows head is tossed again. (If the coin shows tail, it won&#039;t be tossed again.) Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of heads resulting from the &amp;lt;strong&amp;gt;second&amp;lt;/strong&amp;gt; round of tosses, and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the number of heads resulting from &amp;lt;strong&amp;gt;all&amp;lt;/strong&amp;gt; tosses, which includes the first and (possible) second round of each toss.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the PDF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{E}[Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Let &amp;lt;math&amp;gt;p_X&amp;lt;/math&amp;gt; be the PMF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, show that [math]p_X(k-1)p_X(k+1)\leq [p_X(k)]^2[/math] for &amp;lt;math&amp;gt;1\leq k \leq n-1&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
* [&#039;&#039;&#039;Independence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;1\leq r\leq n&amp;lt;/math&amp;gt; be independent random variables which are symmetric about &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;; that is, &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-X_r&amp;lt;/math&amp;gt; have the same distributions. Show that, for all &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}[S_n \geq x] = \mathbf{Pr}[S_n \leq -x]&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;S_n = \sum_{r=1}^n X_r&amp;lt;/math&amp;gt;. Is the conclusion true without the assumtion of independence?&lt;br /&gt;
* [&#039;&#039;&#039;Dependence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be discrete random variables with joint mass function &amp;lt;math&amp;gt;f(x,y) = \frac{C}{(x+y-1)(x+y)(x+y+1)}&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;x,y \in \mathbb{N}_+&amp;lt;/math&amp;gt; (in other words, &amp;lt;math&amp;gt;x,y = 1,2,3,\cdots&amp;lt;/math&amp;gt;). Find (1) the value of &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;, (2) marginal mass function of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and (3) &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt;.&lt;br /&gt;
* [&#039;&#039;&#039;Expectation&#039;&#039;&#039;] It is required to place in order &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; books &amp;lt;math&amp;gt;B_1, B_2, \cdots, B_n&amp;lt;/math&amp;gt; on a library shelf in such way that readers searching from left to right waste as little time as possible on average. Assuming that a random reader requires book &amp;lt;math&amp;gt;B_i&amp;lt;/math&amp;gt; with probability &amp;lt;math&amp;gt;p_i&amp;lt;/math&amp;gt;, find the ordering of the books which minimizes the expected number of titles examined by a random reader before discovery of the required book.&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Law of total expectation&amp;lt;/strong&amp;gt;] Let [math]X \sim \mathrm{Geom}(p)[/math] for some parameter [math]p \in (0,1)[/math]. Calculate [math]\mathbf{E}[X][/math] using the law of total expectation.&lt;br /&gt;
&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Entropy of discrete random variable&amp;lt;/strong&amp;gt;] Let [math]X[/math] be a discrete random variable with range of values [math][N] = \{1,2,\ldots,N\}[/math] and probability mass function [math]p[/math]. Define [math]H(X) = -\sum_{n \ge 1} p(n) \log p(n)[/math] with convention [math]0\log 0 = 0[/math]. Prove that [math]H(X) \le \log N[/math] using Jensen&#039;s inequality.&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Discrete random variable, 14 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (I)&amp;lt;/strong&amp;gt;] Every package of some intrinsically dull commodity includes a small and exciting plastic object. There are [math]c[/math] different types of object, and each package is equally likely to contain any given type. You buy one package each day.&amp;lt;/p&amp;gt;&lt;br /&gt;
    &amp;lt;ol&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse between the acquisitions of the [math]j[/math]-th new type of object and the [math](j + 1)[/math]-th new type.&amp;lt;/li&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse before you have a full set of objects.&amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (II)&amp;lt;/strong&amp;gt;] Prove that geometric distribution is the only discrete memoryless distribution with range values &amp;lt;math&amp;gt;\mathbb{N}_+&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;n_1,n_2 \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;0 \le p \le 1&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Bin}(n_1,p),Y \sim \mathrm{Bin}(n_2,p)&amp;lt;/math&amp;gt; be independent random variables. Prove that &amp;lt;math&amp;gt;X+Y \sim \mathrm{Bin}(n_1+n_2,p)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Negative binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; follows the negative binomial distribution with parameter &amp;lt;math&amp;gt;r \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt;. Calculate &amp;lt;math&amp;gt;\mathbf{Var}[X] = \mathbf{E}[X^2] - \left(\mathbf{E}[X]\right)^2&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Hypergeometric distribution&amp;lt;/strong&amp;gt;] An urn contains [math]N[/math] balls, [math]b[/math] of which are blue and [math]r = N -b[/math] of which are red. A random sample of [math]n[/math] balls is drawn &amp;lt;strong&amp;gt;without replacement&amp;lt;/strong&amp;gt; (无放回) from the urn. Let [math]B[/math] the number of blue balls in this sample. Show that if [math]N, b[/math], and [math]r[/math] approach [math]+\infty[/math] in such a way that [math]b/N \rightarrow  p[/math] and [math]r/N \rightarrow 1 - p[/math], then [math]\mathbf{Pr}(B = k) \rightarrow {n\choose k}p^k(1-p)^{n-k}[/math] for [math]0\leq k \leq n[/math].&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Poisson distribution&amp;lt;/strong&amp;gt;] In your pocket is a random number &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; of coins, where &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; has the Poisson distribution with parameter &amp;lt;math&amp;gt;\lambda&amp;lt;/math&amp;gt;. You toss each coin once, with heads showing with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; each time. Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the (random) number of heads outcomes and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the (also random) number of tails.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the joint mass function of &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find PDF of the marginal distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; in &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;. Are &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; independent?&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Conditional distribution&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;\lambda,\mu &amp;gt; 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;n \in \mathbb{N}&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Pois}(\lambda), Y \sim \mathrm{Pois}(\mu)&amp;lt;/math&amp;gt; be independent random variables. Find out the conditional distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, given &amp;lt;math&amp;gt;X+Y = n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Linearity of Expectation, 12 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Streak&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Suppose we flip a fair coin &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; times independently to obtain a sequence of flips &amp;lt;math&amp;gt;X_1, X_2, \ldots , X_n&amp;lt;/math&amp;gt;. A streak of flips is a consecutive subsequence of flips that are all the same. For example, if &amp;lt;math&amp;gt;X_3&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;X_4&amp;lt;/math&amp;gt;, and &amp;lt;math&amp;gt;X_5&amp;lt;/math&amp;gt; are all heads, there is a streak of length &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; starting at the third lip. (If &amp;lt;math&amp;gt;X_6&amp;lt;/math&amp;gt; is also heads, then there is also a streak of length &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; starting at the third lip.) Find the expected number of streaks of length &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; for some integer &amp;lt;math&amp;gt;k \ge 1&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Number of cycles&amp;lt;/strong&amp;gt;] &lt;br /&gt;
At a banquet, there are &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people who shake hands according to the following process: In each round, two idle hands are randomly selected and shaken (&amp;lt;strong&amp;gt;these two hands are no longer idle&amp;lt;/strong&amp;gt;). After &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds, there will be no idle hands left, and the &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people will form several cycles. For example, when &amp;lt;math&amp;gt;n=3&amp;lt;/math&amp;gt;, the following situation may occur: the left and right hands of the first person are held together, the left hand of the second person and the right hand of the third person are held together, and the right hand of the second person and the left hand of the third person are held together. In this case, three people form two cycles. How many cycles are expected to be formed after &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds?&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Number of operations&amp;lt;/strong&amp;gt;]&lt;br /&gt;
We have a directed graph &amp;lt;math&amp;gt;G=(V,E)&amp;lt;/math&amp;gt; without self-loops and multi-edges.&lt;br /&gt;
&lt;br /&gt;
Until the graph becomes empty, repeat the following operation:&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
Choose one (unerased) vertex uniformly at random (independently from the previous choices). Then, erase that vertex and all vertices that are reachable from the chosen vertex by traversing some edges. Erasing a vertex will also erase the edges incident to it.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;R_v&amp;lt;/math&amp;gt; denotes the number of vertices from which vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt; is reachable. Prove that the expected value of the number of operations equals to &amp;lt;math&amp;gt;\sum_{v \in V} \frac{1}{R_v}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Expected Mex&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X_1,X_2,\ldots,X_{100} \sim \mathrm{Geo}(1/2)&amp;lt;/math&amp;gt; be independent random variables. Compute &amp;lt;math&amp;gt;\mathbf{E}[\mathrm{mex}(X_1,X_2,\ldots,X_{100})]&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\mathrm{mex}(a_1,a_2,\ldots,a_n)&amp;lt;/math&amp;gt; is the smallest positive integer that does not appear in &amp;lt;math&amp;gt;a_1,a_2,\ldots,a_n&amp;lt;/math&amp;gt;. Your answer is considered correct if the absolute error does not exceed &amp;lt;math&amp;gt;10^{-6}&amp;lt;/math&amp;gt;. (Hint: Although mathematics will help you arrive at elegant and efficient methods, the use of a computer and programming skills will be required.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 4 (Probability meets graph theory) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random social networks&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G = (V, E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph without isolating vertex. &lt;br /&gt;
Let &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; be the degree of vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Let &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be a uniformly chosen vertex, and &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; a uniformly chosen neighbor of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Show that &amp;lt;math&amp;gt;\mathbf{E}[d_Z] \geq \mathbf{E}[d_Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Interpret this inequality in the context of social networks, in which the vertices represent people, and the edges represent friendship.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Turán&#039;s Theorem&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G=(V,E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph, and write &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; for the degree of the vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Use probablistic method to prove that &amp;lt;math&amp;gt;\alpha(G) \ge \sum_{v \in V} \frac{1}{d_v + 1}&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\alpha(G)&amp;lt;/math&amp;gt; is the size of a maximum independent set. (Hint: Consider the following random procedure for generating an independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; from a graph with vertex set &amp;lt;math&amp;gt;V&amp;lt;/math&amp;gt;: First, generate a random permutation of the vertices, denoted as &amp;lt;math&amp;gt;v_1,v_2,\ldots,v_n&amp;lt;/math&amp;gt;. Then, construct the independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; as follows: For each vertex &amp;lt;math&amp;gt;v_i \in V&amp;lt;/math&amp;gt;, add &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; if and only if none of its predecessors in the permutation, i.e., &amp;lt;math&amp;gt;v_1,\ldots,v_{i-1}&amp;lt;/math&amp;gt;, are neighbors of &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Tournament&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Prove that there is a [https://en.wikipedia.org/wiki/Tournament_(graph_theory) tournament] &amp;lt;math&amp;gt;T&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; players and at least &amp;lt;math&amp;gt;n!/2^{n-1}&amp;lt;/math&amp;gt;Hamiltonian paths.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 5 (1D random walk, 8 points) ==&lt;br /&gt;
Let &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt; be a constant, and &amp;lt;math&amp;gt;\{X_n\}_{n \ge 1}&amp;lt;/math&amp;gt; be independent Bernoulli trials with successful probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;.&lt;br /&gt;
Define &amp;lt;math&amp;gt;S_n = 2\sum_{i=1}^n X_i - n&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;S_0 = 0&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Range of random walk&amp;lt;/strong&amp;gt;] The range &amp;lt;math&amp;gt;R_n&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;S_0, S_1, \ldots, S_n&amp;lt;/math&amp;gt; is defined as the number of distinct values taken by the sequence. Show that &amp;lt;math&amp;gt;\mathbf{Pr}\left(R_n = R_{n-1}+1\right) = \mathbf{Pr}\left(\forall 1 \le i \le n, S_i \neq 0\right)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, and deduce that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n]\to \mathbf{Pr}(\forall i \ge 1, S_i \neq 0)&amp;lt;/math&amp;gt;. Hence show that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n] \to |2p-1|&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Symmetric 1D random walk (III)&amp;lt;/strong&amp;gt;] Suppose &amp;lt;math&amp;gt;p = \frac{1}{2}&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{E}[|S_n|] = \Theta(\sqrt{n})&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12967</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 2</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12967"/>
		<updated>2025-03-13T15:49:20Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: /* Problem 1 (Warm-up problems, 16 points) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 2==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Warm-up problems, 16 points) ==&lt;br /&gt;
* [&#039;&#039;&#039;Cumulative distribution function (CDF) (I)&#039;&#039;&#039;] Express the distribution functions of [math]X^+ = \max\{0,X\}[/math], [math]X^- = -\min\{0,X\}[/math], [math]|X|=X^+ + X^-[/math], [math]-X[/math], in terms of the distribution function [math]F[/math] of the random variable [math]X[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Cumulative distribution function (CDF) (II)&#039;&#039;&#039;] Let [math]X[/math] be a random variable with distribution function [math]\max(0,\min(1,x))[/math]. Let [math]F[/math] be a distribution function which is continuous and strictly increasing. Show that [math]Y=F^{-1}(X)[/math] be a random variable with distribution function [math]F[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Probability density function (PDF)&#039;&#039;&#039;] We toss &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; coins, and each one shows heads with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;, independently of each of the others. Each coin which shows head is tossed again. (If the coin shows tail, it won&#039;t be tossed again.) Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of heads resulting from the &amp;lt;strong&amp;gt;second&amp;lt;/strong&amp;gt; round of tosses, and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the number of heads resulting from &amp;lt;strong&amp;gt;all&amp;lt;/strong&amp;gt; tosses, which includes the first and (possible) second round of each toss.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the PDF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{E}[Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Let &amp;lt;math&amp;gt;p_X&amp;lt;/math&amp;gt; be the PMF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, show that [math]p_X(k-1)p_X(k+1)\leq [p_X(k)]^2[/math] for &amp;lt;math&amp;gt;1\leq k \leq n-1&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
* [&#039;&#039;&#039;Independence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;1\leq r\leq n&amp;lt;/math&amp;gt; be independent random variables which are symmetric about &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;; that is, &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-X_r&amp;lt;/math&amp;gt; have the same distributions. Show that, for all &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}[S_n \geq x] = \mathbf{Pr}[S_n \leq -x]&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;S_n = \sum_{r=1}^n X_r&amp;lt;/math&amp;gt;. Is the conclusion true without the assumtion of independence?&lt;br /&gt;
* [&#039;&#039;&#039;Dependence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be discrete random variables with joint mass function &amp;lt;math&amp;gt;f(x,y) = \frac{C}{(x+y-1)(x+y)(x+y+1)}&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;x,y \in \mathbb{N}_+&amp;lt;/math&amp;gt; (in other words, &amp;lt;math&amp;gt;x,y = 1,2,3,\cdots&amp;lt;/math&amp;gt;). Find (1) the value of &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;, (2) marginal mass function of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and (3) &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt;.&lt;br /&gt;
* [&#039;&#039;&#039;Expectation&#039;&#039;&#039;] It is required to place in order &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; books &amp;lt;math&amp;gt;B_1, B_2, \cdots, B_n&amp;lt;/math&amp;gt; on a library shelf in such way that readers searching from left to right waste as little time as possible on average. Assuming that a random reader requires book &amp;lt;math&amp;gt;B_i&amp;lt;/math&amp;gt; with probability &amp;lt;math&amp;gt;p_i&amp;lt;/math&amp;gt;, find the ordering of the books which minimizes the expected number of titles examined by a random reader before discovery of the required book.&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Law of total expectation&amp;lt;/strong&amp;gt;] Let [math]X \sim \mathrm{Geom}(p)[/math] for some parameter [math]p \in (0,1)[/math]. Calculate [math]\mathbf{E}[X][/math] using the law of total expectation.&lt;br /&gt;
&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Entropy of discrete random variable&amp;lt;/strong&amp;gt;] Let [math]X[/math] be a discrete random variable with range of values [math][N] = \{1,2,\ldots,N\}[/math] and probability mass function [math]p[/math]. Define [math]H(X) = -\sum_{n \ge 1} p(n) \log p(n)[/math] with convention [math]0\log 0 = 0[/math]. Prove that [math]H(X) \le \log N[/math] using Jensen&#039;s inequality.&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Discrete random variable, 14 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (I)&amp;lt;/strong&amp;gt;] Every package of some intrinsically dull commodity includes a small and exciting plastic object. There are [math]c[/math] different types of object, and each package is equally likely to contain any given type. You buy one package each day.&amp;lt;/p&amp;gt;&lt;br /&gt;
    &amp;lt;ol&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse between the acquisitions of the [math]j[/math]-th new type of object and the [math](j + 1)[/math]-th new type.&amp;lt;/li&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse before you have a full set of objects.&amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (II)&amp;lt;/strong&amp;gt;] Prove that geometric distribution is the only discrete memoryless distribution with range values &amp;lt;math&amp;gt;\mathbb{N}_+&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;n_1,n_2 \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;0 \le p \le 1&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Bin}(n_1,p),Y \sim \mathrm{Bin}(n_2,p)&amp;lt;/math&amp;gt; be independent random variables. Prove that &amp;lt;math&amp;gt;X+Y \sim \mathrm{Bin}(n_1+n_2,p)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Negative binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; follows the negative binomial distribution with parameter &amp;lt;math&amp;gt;r \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt;. Calculate &amp;lt;math&amp;gt;\mathbf{Var}[X] = \mathbf{E}[X^2] - \left(\mathbf{E}[X]\right)^2&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Hypergeometric distribution&amp;lt;/strong&amp;gt;] An urn contains [math]N[/math] balls, [math]b[/math] of which are blue and [math]r = N -b[/math] of which are red. A random sample of [math]n[/math] balls is drawn &amp;lt;strong&amp;gt;without replacement&amp;lt;/strong&amp;gt; (无放回) from the urn. Let [math]B[/math] the number of blue balls in this sample. Show that if [math]N, b[/math], and [math]r[/math] approach [math]+\infty[/math] in such a way that [math]b/N \rightarrow  p[/math] and [math]r/N \rightarrow 1 - p[/math], then [math]\mathbf{Pr}(B = k) \rightarrow {n\choose k}p^k(1-p)^{n-k}[/math] for [math]0\leq k \leq n[/math].&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Poisson distribution&amp;lt;/strong&amp;gt;] In your pocket is a random number &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; of coins, where &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; has the Poisson distribution with parameter &amp;lt;math&amp;gt;\lambda&amp;lt;/math&amp;gt;. You toss each coin once, with heads showing with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; each time. Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the (random) number of heads outcomes and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the (also random) number of tails.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the joint mass function of &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find PDF of the marginal distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; in &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;. Are &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; independent?&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Conditional distribution&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;\lambda,\mu &amp;gt; 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;n \in \mathbb{N}&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Pois}(\lambda), Y \sim \mathrm{Pois}(\mu)&amp;lt;/math&amp;gt; be independent random variables. Find out the conditional distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, given &amp;lt;math&amp;gt;X+Y = n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Linearity of Expectation, 12 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Streak&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Suppose we flip a fair coin &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; times independently to obtain a sequence of flips &amp;lt;math&amp;gt;X_1, X_2, \ldots , X_n&amp;lt;/math&amp;gt;. A streak of flips is a consecutive subsequence of flips that are all the same. For example, if &amp;lt;math&amp;gt;X_3&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;X_4&amp;lt;/math&amp;gt;, and &amp;lt;math&amp;gt;X_5&amp;lt;/math&amp;gt; are all heads, there is a streak of length &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; starting at the third lip. (If &amp;lt;math&amp;gt;X_6&amp;lt;/math&amp;gt; is also heads, then there is also a streak of length &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; starting at the third lip.) Find the expected number of streaks of length &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; for some integer &amp;lt;math&amp;gt;k \ge 1&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Number of cycles&amp;lt;/strong&amp;gt;] &lt;br /&gt;
At a banquet, there are &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people who shake hands according to the following process: In each round, two idle hands are randomly selected and shaken (&amp;lt;strong&amp;gt;these two hands are no longer idle&amp;lt;/strong&amp;gt;). After &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds, there will be no idle hands left, and the &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people will form several cycles. For example, when &amp;lt;math&amp;gt;n=3&amp;lt;/math&amp;gt;, the following situation may occur: the left and right hands of the first person are held together, the left hand of the second person and the right hand of the third person are held together, and the right hand of the second person and the left hand of the third person are held together. In this case, three people form two cycles. How many cycles are expected to be formed after &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds?&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Number of operations&amp;lt;/strong&amp;gt;]&lt;br /&gt;
We have a directed graph &amp;lt;math&amp;gt;G=(V,E)&amp;lt;/math&amp;gt; without self-loops and multi-edges.&lt;br /&gt;
&lt;br /&gt;
Until the graph becomes empty, repeat the following operation:&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
Choose one (unerased) vertex uniformly at random (independently from the previous choices). Then, erase that vertex and all vertices that are reachable from the chosen vertex by traversing some edges. Erasing a vertex will also erase the edges incident to it.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;R_v&amp;lt;/math&amp;gt; denotes the number of vertices from which vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt; is reachable. Prove that the expected value of the number of operations equals to &amp;lt;math&amp;gt;\sum_{v \in V} \frac{1}{R_v}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Expected Mex&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X_1,X_2,\ldots,X_{100} \sim \mathrm{Geo}(1/2)&amp;lt;/math&amp;gt; be independent random variables. Compute &amp;lt;math&amp;gt;\mathbf{E}[\mathrm{mex}(X_1,X_2,\ldots,X_{100})]&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\mathrm{mex}(a_1,a_2,\ldots,a_n)&amp;lt;/math&amp;gt; is the smallest positive integer that does not appear in &amp;lt;math&amp;gt;a_1,a_2,\ldots,a_n&amp;lt;/math&amp;gt;. Your answer is considered correct if the absolute error does not exceed &amp;lt;math&amp;gt;10^{-6}&amp;lt;/math&amp;gt;. (Hint: Although mathematics will help you arrive at elegant and efficient methods, the use of a computer and programming skills will be required.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 4 (Probability meets graph theory) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random social networks&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G = (V, E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph without isolating vertex. &lt;br /&gt;
Let &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; be the degree of vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Let &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be a uniformly chosen vertex, and &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; a uniformly chosen neighbor of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Show that &amp;lt;math&amp;gt;\mathbf{E}[d_Z] \geq \mathbf{E}[d_Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Interpret this inequality in the context of social networks, in which the vertices represent people, and the edges represent friendship.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Turán&#039;s Theorem&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G=(V,E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph, and write &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; for the degree of the vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Use probablistic method to prove that &amp;lt;math&amp;gt;\alpha(G) \ge \sum_{v \in V} \frac{1}{d_v + 1}&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\alpha(G)&amp;lt;/math&amp;gt; is the size of a maximum independent set. (Hint: Consider the following random procedure for generating an independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; from a graph with vertex set &amp;lt;math&amp;gt;V&amp;lt;/math&amp;gt;: First, generate a random permutation of the vertices, denoted as &amp;lt;math&amp;gt;v_1,v_2,\ldots,v_n&amp;lt;/math&amp;gt;. Then, construct the independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; as follows: For each vertex &amp;lt;math&amp;gt;v_i \in V&amp;lt;/math&amp;gt;, add &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; if and only if none of its predecessors in the permutation, i.e., &amp;lt;math&amp;gt;v_1,\ldots,v_{i-1}&amp;lt;/math&amp;gt;, are neighbors of &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Tournament&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Prove that there is a [https://en.wikipedia.org/wiki/Tournament_(graph_theory) tournament] &amp;lt;math&amp;gt;T&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; players and at least &amp;lt;math&amp;gt;n!/2^{n-1}&amp;lt;/math&amp;gt;Hamiltonian paths.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 5 (1D random walk, 8 points) ==&lt;br /&gt;
Let &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt; be a constant, and &amp;lt;math&amp;gt;\{X_n\}_{n \ge 1}&amp;lt;/math&amp;gt; be independent Bernoulli trials with successful probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;.&lt;br /&gt;
Define &amp;lt;math&amp;gt;S_n = 2\sum_{i=1}^n X_i - n&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;S_0 = 0&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Range of random walk&amp;lt;/strong&amp;gt;] The range &amp;lt;math&amp;gt;R_n&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;S_0, S_1, \ldots, S_n&amp;lt;/math&amp;gt; is defined as the number of distinct values taken by the sequence. Show that &amp;lt;math&amp;gt;\mathbf{Pr}\left(R_n = R_{n-1}+1\right) = \mathbf{Pr}\left(\forall 1 \le i \le n, S_i \neq 0\right)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, and deduce that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n]\to \mathbf{Pr}(\forall i \ge 1, S_i \neq 0)&amp;lt;/math&amp;gt;. Hence show that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n] \to |2p-1|&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Symmetric 1D random walk (III)&amp;lt;/strong&amp;gt;] Suppose &amp;lt;math&amp;gt;p = \frac{1}{2}&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{E}[|S_n|] = \Theta(\sqrt{n})&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12966</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 2</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12966"/>
		<updated>2025-03-13T15:47:13Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: /* Problem 1 (Warm-up problems, 16 points) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 2==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Warm-up problems, 16 points) ==&lt;br /&gt;
* [&#039;&#039;&#039;Cumulative distribution function (CDF) (I)&#039;&#039;&#039;] Express the distribution functions of [math]X^+ = \max\{0,X\}[/math], [math]X^- = -\min\{0,X\}[/math], [math]|X|=X^+ + X^-[/math], [math]-X[/math], in terms of the distribution function [math]F[/math] of the random variable [math]X[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Cumulative distribution function (CDF) (II)&#039;&#039;&#039;] Let [math]X[/math] be a random variable with distribution function [math]\max(0,\min(1,x))[/math]. Let [math]F[/math] be a distribution function which is continuous and strictly increasing. Show that [math]Y=F^{-1}(X)[/math] be a random variable with distribution function [math]F[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Probability density function (PDF)&#039;&#039;&#039;] We toss &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; coins, and each one shows heads with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;, independently of each of the others. Each coin which shows head is tossed again. (If the coin shows tail, it won&#039;t be tossed again.) Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of heads resulting from the &amp;lt;strong&amp;gt;second&amp;lt;/strong&amp;gt; round of tosses, and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the number of heads resulting from &amp;lt;strong&amp;gt;all&amp;lt;/strong&amp;gt; tosses, which includes the first and (possible) second round of each toss.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the PDF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{E}[Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Let &amp;lt;math&amp;gt;p_X&amp;lt;/math&amp;gt; be the PMF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, show that [math]p_X(k-1)p_X(k+1)\leq [p_X(k)]^2[/math] for &amp;lt;math&amp;gt;1\leq k \leq n-1&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
* [&#039;&#039;&#039;Independence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;1\leq r\leq n&amp;lt;/math&amp;gt; be independent random variables which are symmetric about &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;; that is, &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-X_r&amp;lt;/math&amp;gt; have the same distributions. Show that, for all &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}[S_n \geq x] = \mathbf{Pr}[S_n \leq -x]&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;S_n = \sum_{r=1}^n X_r&amp;lt;/math&amp;gt;. Is the conclusion true without the assumtion of independence?&lt;br /&gt;
* [&#039;&#039;&#039;Dependence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be discrete random variables with joint mass function &amp;lt;math&amp;gt;f(x,y) = \frac{C}{(x+y-1)(x+y)(x+y+1)}&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;x,y \in \mathbb{N}_+&amp;lt;/math&amp;gt; (in other words, &amp;lt;math&amp;gt;x,y = 1,2,3,\cdots&amp;lt;/math&amp;gt;). Find (1) the value of &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;, (2) marginal mass function of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and (3) &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt;.&lt;br /&gt;
* [&#039;&#039;&#039;Expectation&#039;&#039;&#039;] It is required to place in order &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; books &amp;lt;math&amp;gt;B_1, B_2, \cdots, B_n&amp;lt;/math&amp;gt; on a library shelf in such way that readers searching from left to right waste as little time as possible on average. Assuming that a random reader requires book &amp;lt;math&amp;gt;B_i&amp;lt;/math&amp;gt; with probability &amp;lt;math&amp;gt;p_i&amp;lt;/math&amp;gt;, find the ordering of the books which minimizes the number of titles examined by a random reader before discovery of the required book.&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Law of total expectation&amp;lt;/strong&amp;gt;] Let [math]X \sim \mathrm{Geom}(p)[/math] for some parameter [math]p \in (0,1)[/math]. Calculate [math]\mathbf{E}[X][/math] using the law of total expectation.&lt;br /&gt;
&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Entropy of discrete random variable&amp;lt;/strong&amp;gt;] Let [math]X[/math] be a discrete random variable with range of values [math][N] = \{1,2,\ldots,N\}[/math] and probability mass function [math]p[/math]. Define [math]H(X) = -\sum_{n \ge 1} p(n) \log p(n)[/math] with convention [math]0\log 0 = 0[/math]. Prove that [math]H(X) \le \log N[/math] using Jensen&#039;s inequality.&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Discrete random variable, 14 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (I)&amp;lt;/strong&amp;gt;] Every package of some intrinsically dull commodity includes a small and exciting plastic object. There are [math]c[/math] different types of object, and each package is equally likely to contain any given type. You buy one package each day.&amp;lt;/p&amp;gt;&lt;br /&gt;
    &amp;lt;ol&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse between the acquisitions of the [math]j[/math]-th new type of object and the [math](j + 1)[/math]-th new type.&amp;lt;/li&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse before you have a full set of objects.&amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (II)&amp;lt;/strong&amp;gt;] Prove that geometric distribution is the only discrete memoryless distribution with range values &amp;lt;math&amp;gt;\mathbb{N}_+&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;n_1,n_2 \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;0 \le p \le 1&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Bin}(n_1,p),Y \sim \mathrm{Bin}(n_2,p)&amp;lt;/math&amp;gt; be independent random variables. Prove that &amp;lt;math&amp;gt;X+Y \sim \mathrm{Bin}(n_1+n_2,p)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Negative binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; follows the negative binomial distribution with parameter &amp;lt;math&amp;gt;r \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt;. Calculate &amp;lt;math&amp;gt;\mathbf{Var}[X] = \mathbf{E}[X^2] - \left(\mathbf{E}[X]\right)^2&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Hypergeometric distribution&amp;lt;/strong&amp;gt;] An urn contains [math]N[/math] balls, [math]b[/math] of which are blue and [math]r = N -b[/math] of which are red. A random sample of [math]n[/math] balls is drawn &amp;lt;strong&amp;gt;without replacement&amp;lt;/strong&amp;gt; (无放回) from the urn. Let [math]B[/math] the number of blue balls in this sample. Show that if [math]N, b[/math], and [math]r[/math] approach [math]+\infty[/math] in such a way that [math]b/N \rightarrow  p[/math] and [math]r/N \rightarrow 1 - p[/math], then [math]\mathbf{Pr}(B = k) \rightarrow {n\choose k}p^k(1-p)^{n-k}[/math] for [math]0\leq k \leq n[/math].&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Poisson distribution&amp;lt;/strong&amp;gt;] In your pocket is a random number &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; of coins, where &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; has the Poisson distribution with parameter &amp;lt;math&amp;gt;\lambda&amp;lt;/math&amp;gt;. You toss each coin once, with heads showing with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; each time. Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the (random) number of heads outcomes and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the (also random) number of tails.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the joint mass function of &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find PDF of the marginal distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; in &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;. Are &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; independent?&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Conditional distribution&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;\lambda,\mu &amp;gt; 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;n \in \mathbb{N}&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Pois}(\lambda), Y \sim \mathrm{Pois}(\mu)&amp;lt;/math&amp;gt; be independent random variables. Find out the conditional distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, given &amp;lt;math&amp;gt;X+Y = n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Linearity of Expectation, 12 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Streak&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Suppose we flip a fair coin &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; times independently to obtain a sequence of flips &amp;lt;math&amp;gt;X_1, X_2, \ldots , X_n&amp;lt;/math&amp;gt;. A streak of flips is a consecutive subsequence of flips that are all the same. For example, if &amp;lt;math&amp;gt;X_3&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;X_4&amp;lt;/math&amp;gt;, and &amp;lt;math&amp;gt;X_5&amp;lt;/math&amp;gt; are all heads, there is a streak of length &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; starting at the third lip. (If &amp;lt;math&amp;gt;X_6&amp;lt;/math&amp;gt; is also heads, then there is also a streak of length &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; starting at the third lip.) Find the expected number of streaks of length &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; for some integer &amp;lt;math&amp;gt;k \ge 1&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Number of cycles&amp;lt;/strong&amp;gt;] &lt;br /&gt;
At a banquet, there are &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people who shake hands according to the following process: In each round, two idle hands are randomly selected and shaken (&amp;lt;strong&amp;gt;these two hands are no longer idle&amp;lt;/strong&amp;gt;). After &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds, there will be no idle hands left, and the &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people will form several cycles. For example, when &amp;lt;math&amp;gt;n=3&amp;lt;/math&amp;gt;, the following situation may occur: the left and right hands of the first person are held together, the left hand of the second person and the right hand of the third person are held together, and the right hand of the second person and the left hand of the third person are held together. In this case, three people form two cycles. How many cycles are expected to be formed after &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds?&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Number of operations&amp;lt;/strong&amp;gt;]&lt;br /&gt;
We have a directed graph &amp;lt;math&amp;gt;G=(V,E)&amp;lt;/math&amp;gt; without self-loops and multi-edges.&lt;br /&gt;
&lt;br /&gt;
Until the graph becomes empty, repeat the following operation:&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
Choose one (unerased) vertex uniformly at random (independently from the previous choices). Then, erase that vertex and all vertices that are reachable from the chosen vertex by traversing some edges. Erasing a vertex will also erase the edges incident to it.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;R_v&amp;lt;/math&amp;gt; denotes the number of vertices from which vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt; is reachable. Prove that the expected value of the number of operations equals to &amp;lt;math&amp;gt;\sum_{v \in V} \frac{1}{R_v}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Expected Mex&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X_1,X_2,\ldots,X_{100} \sim \mathrm{Geo}(1/2)&amp;lt;/math&amp;gt; be independent random variables. Compute &amp;lt;math&amp;gt;\mathbf{E}[\mathrm{mex}(X_1,X_2,\ldots,X_{100})]&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\mathrm{mex}(a_1,a_2,\ldots,a_n)&amp;lt;/math&amp;gt; is the smallest positive integer that does not appear in &amp;lt;math&amp;gt;a_1,a_2,\ldots,a_n&amp;lt;/math&amp;gt;. Your answer is considered correct if the absolute error does not exceed &amp;lt;math&amp;gt;10^{-6}&amp;lt;/math&amp;gt;. (Hint: Although mathematics will help you arrive at elegant and efficient methods, the use of a computer and programming skills will be required.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 4 (Probability meets graph theory) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random social networks&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G = (V, E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph without isolating vertex. &lt;br /&gt;
Let &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; be the degree of vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Let &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be a uniformly chosen vertex, and &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; a uniformly chosen neighbor of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Show that &amp;lt;math&amp;gt;\mathbf{E}[d_Z] \geq \mathbf{E}[d_Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Interpret this inequality in the context of social networks, in which the vertices represent people, and the edges represent friendship.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Turán&#039;s Theorem&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G=(V,E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph, and write &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; for the degree of the vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Use probablistic method to prove that &amp;lt;math&amp;gt;\alpha(G) \ge \sum_{v \in V} \frac{1}{d_v + 1}&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\alpha(G)&amp;lt;/math&amp;gt; is the size of a maximum independent set. (Hint: Consider the following random procedure for generating an independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; from a graph with vertex set &amp;lt;math&amp;gt;V&amp;lt;/math&amp;gt;: First, generate a random permutation of the vertices, denoted as &amp;lt;math&amp;gt;v_1,v_2,\ldots,v_n&amp;lt;/math&amp;gt;. Then, construct the independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; as follows: For each vertex &amp;lt;math&amp;gt;v_i \in V&amp;lt;/math&amp;gt;, add &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; if and only if none of its predecessors in the permutation, i.e., &amp;lt;math&amp;gt;v_1,\ldots,v_{i-1}&amp;lt;/math&amp;gt;, are neighbors of &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Tournament&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Prove that there is a [https://en.wikipedia.org/wiki/Tournament_(graph_theory) tournament] &amp;lt;math&amp;gt;T&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; players and at least &amp;lt;math&amp;gt;n!/2^{n-1}&amp;lt;/math&amp;gt;Hamiltonian paths.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 5 (1D random walk, 8 points) ==&lt;br /&gt;
Let &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt; be a constant, and &amp;lt;math&amp;gt;\{X_n\}_{n \ge 1}&amp;lt;/math&amp;gt; be independent Bernoulli trials with successful probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;.&lt;br /&gt;
Define &amp;lt;math&amp;gt;S_n = 2\sum_{i=1}^n X_i - n&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;S_0 = 0&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Range of random walk&amp;lt;/strong&amp;gt;] The range &amp;lt;math&amp;gt;R_n&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;S_0, S_1, \ldots, S_n&amp;lt;/math&amp;gt; is defined as the number of distinct values taken by the sequence. Show that &amp;lt;math&amp;gt;\mathbf{Pr}\left(R_n = R_{n-1}+1\right) = \mathbf{Pr}\left(\forall 1 \le i \le n, S_i \neq 0\right)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, and deduce that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n]\to \mathbf{Pr}(\forall i \ge 1, S_i \neq 0)&amp;lt;/math&amp;gt;. Hence show that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n] \to |2p-1|&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Symmetric 1D random walk (III)&amp;lt;/strong&amp;gt;] Suppose &amp;lt;math&amp;gt;p = \frac{1}{2}&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{E}[|S_n|] = \Theta(\sqrt{n})&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12965</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 2</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12965"/>
		<updated>2025-03-13T15:43:43Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: /* Problem 2 (Discrete random variable) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 2==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Warm-up problems, 16 points) ==&lt;br /&gt;
* [&#039;&#039;&#039;Cumulative distribution function (CDF) (I)&#039;&#039;&#039;] Express the distribution functions of [math]X^+ = \max\{0,X\}[/math], [math]X^- = -\min\{0,X\}[/math], [math]|X|=X^+ + X^-[/math], [math]-X[/math], in terms of the distribution function [math]F[/math] of the random variable [math]X[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Cumulative distribution function (CDF) (II)&#039;&#039;&#039;] Let [math]X[/math] be a random variable with distribution function [math]\max(0,\min(1,x))[/math]. Let [math]F[/math] be a distribution function which is continuous and strictly increasing. Show that [math]Y=F^{-1}(X)[/math] be a random variable with distribution function [math]F[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Probability density function (PDF)&#039;&#039;&#039;] We toss &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; coins, and each one shows heads with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;, independently of each of the others. Each coin which shows head is tossed again. (If the coin shows tail, it won&#039;t be tossed again.) Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of heads resulting from the &amp;lt;strong&amp;gt;second&amp;lt;/strong&amp;gt; round of tosses, and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the number of heads resulting from &amp;lt;strong&amp;gt;all&amp;lt;/strong&amp;gt; tosses, which includes the first and (possible) second round of each toss.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the PDF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{E}[Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Let &amp;lt;math&amp;gt;p_X&amp;lt;/math&amp;gt; be the PMF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, show that [math]p_X(k-1)p_X(k+1)\leq [p_X(k)]^2[/math] for &amp;lt;math&amp;gt;1\leq k \leq n-1&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
* [&#039;&#039;&#039;Independence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;1\leq r\leq n&amp;lt;/math&amp;gt; be independent random variables which are symmetric about &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;; that is, &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-X_r&amp;lt;/math&amp;gt; have the same distributions. Show that, for all &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}[S_n \geq x] = \mathbf{Pr}[S_n \leq -x]&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;S_n = \sum_{r=1}^n X_r&amp;lt;/math&amp;gt;. Is the conclusion true without the assumtion of independence?&lt;br /&gt;
* [&#039;&#039;&#039;Dependence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be discrete random variables with joint mass function &amp;lt;math&amp;gt;f(x,y) = \frac{C}{(x+y-1)(x+y)(x+y+1)}&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;x,y \in \mathbb{N}_+&amp;lt;/math&amp;gt; (in other words, &amp;lt;math&amp;gt;x,y = 1,2,3,\cdots&amp;lt;/math&amp;gt;). Find (1) the value of &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;, (2) marginal mass function of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and (3) &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt;.&lt;br /&gt;
* [&#039;&#039;&#039;Expectation&#039;&#039;&#039;] It is required to place in order &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; books &amp;lt;math&amp;gt;B_1, B_2, \cdots, B_n&amp;lt;/math&amp;gt; on a library shelf in such way that readers searching from left to right waste as little time as possible on average. Assuming that a random reader requires book &amp;lt;math&amp;gt;B_i&amp;lt;/math&amp;gt; with probability &amp;lt;math&amp;gt;p_i&amp;lt;/math&amp;gt;, find the ordering of the books which minimizes the number of titles examined by a random reader before discovery of the required book.&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Law of total expectation&amp;lt;/strong&amp;gt;] Let [math]X \sim \mathrm{Geom}(p)[/math] for some parameter [math]p \in (0,1)[/math]. Calculate [math]\mathbf{E}[X][/math] using the law of total expectation.&lt;br /&gt;
&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Random number of random variables&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\{X_n\}_{n \ge 1}&amp;lt;/math&amp;gt; be identically distributed random variable and &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; be a random variable taking values in the non-negative integers and independent of the &amp;lt;math&amp;gt;X_n&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;n \ge 1&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{E}\left[\sum_{i=1}^N X_i\right] = \mathbf{E}[N] \mathbf{E}[X_1]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Discrete random variable, 14 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (I)&amp;lt;/strong&amp;gt;] Every package of some intrinsically dull commodity includes a small and exciting plastic object. There are [math]c[/math] different types of object, and each package is equally likely to contain any given type. You buy one package each day.&amp;lt;/p&amp;gt;&lt;br /&gt;
    &amp;lt;ol&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse between the acquisitions of the [math]j[/math]-th new type of object and the [math](j + 1)[/math]-th new type.&amp;lt;/li&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse before you have a full set of objects.&amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (II)&amp;lt;/strong&amp;gt;] Prove that geometric distribution is the only discrete memoryless distribution with range values &amp;lt;math&amp;gt;\mathbb{N}_+&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;n_1,n_2 \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;0 \le p \le 1&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Bin}(n_1,p),Y \sim \mathrm{Bin}(n_2,p)&amp;lt;/math&amp;gt; be independent random variables. Prove that &amp;lt;math&amp;gt;X+Y \sim \mathrm{Bin}(n_1+n_2,p)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Negative binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; follows the negative binomial distribution with parameter &amp;lt;math&amp;gt;r \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt;. Calculate &amp;lt;math&amp;gt;\mathbf{Var}[X] = \mathbf{E}[X^2] - \left(\mathbf{E}[X]\right)^2&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Hypergeometric distribution&amp;lt;/strong&amp;gt;] An urn contains [math]N[/math] balls, [math]b[/math] of which are blue and [math]r = N -b[/math] of which are red. A random sample of [math]n[/math] balls is drawn &amp;lt;strong&amp;gt;without replacement&amp;lt;/strong&amp;gt; (无放回) from the urn. Let [math]B[/math] the number of blue balls in this sample. Show that if [math]N, b[/math], and [math]r[/math] approach [math]+\infty[/math] in such a way that [math]b/N \rightarrow  p[/math] and [math]r/N \rightarrow 1 - p[/math], then [math]\mathbf{Pr}(B = k) \rightarrow {n\choose k}p^k(1-p)^{n-k}[/math] for [math]0\leq k \leq n[/math].&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Poisson distribution&amp;lt;/strong&amp;gt;] In your pocket is a random number &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; of coins, where &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; has the Poisson distribution with parameter &amp;lt;math&amp;gt;\lambda&amp;lt;/math&amp;gt;. You toss each coin once, with heads showing with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; each time. Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the (random) number of heads outcomes and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the (also random) number of tails.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the joint mass function of &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find PDF of the marginal distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; in &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;. Are &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; independent?&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Conditional distribution&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;\lambda,\mu &amp;gt; 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;n \in \mathbb{N}&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Pois}(\lambda), Y \sim \mathrm{Pois}(\mu)&amp;lt;/math&amp;gt; be independent random variables. Find out the conditional distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, given &amp;lt;math&amp;gt;X+Y = n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Linearity of Expectation, 12 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Streak&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Suppose we flip a fair coin &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; times independently to obtain a sequence of flips &amp;lt;math&amp;gt;X_1, X_2, \ldots , X_n&amp;lt;/math&amp;gt;. A streak of flips is a consecutive subsequence of flips that are all the same. For example, if &amp;lt;math&amp;gt;X_3&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;X_4&amp;lt;/math&amp;gt;, and &amp;lt;math&amp;gt;X_5&amp;lt;/math&amp;gt; are all heads, there is a streak of length &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; starting at the third lip. (If &amp;lt;math&amp;gt;X_6&amp;lt;/math&amp;gt; is also heads, then there is also a streak of length &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; starting at the third lip.) Find the expected number of streaks of length &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; for some integer &amp;lt;math&amp;gt;k \ge 1&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Number of cycles&amp;lt;/strong&amp;gt;] &lt;br /&gt;
At a banquet, there are &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people who shake hands according to the following process: In each round, two idle hands are randomly selected and shaken (&amp;lt;strong&amp;gt;these two hands are no longer idle&amp;lt;/strong&amp;gt;). After &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds, there will be no idle hands left, and the &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people will form several cycles. For example, when &amp;lt;math&amp;gt;n=3&amp;lt;/math&amp;gt;, the following situation may occur: the left and right hands of the first person are held together, the left hand of the second person and the right hand of the third person are held together, and the right hand of the second person and the left hand of the third person are held together. In this case, three people form two cycles. How many cycles are expected to be formed after &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds?&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Number of operations&amp;lt;/strong&amp;gt;]&lt;br /&gt;
We have a directed graph &amp;lt;math&amp;gt;G=(V,E)&amp;lt;/math&amp;gt; without self-loops and multi-edges.&lt;br /&gt;
&lt;br /&gt;
Until the graph becomes empty, repeat the following operation:&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
Choose one (unerased) vertex uniformly at random (independently from the previous choices). Then, erase that vertex and all vertices that are reachable from the chosen vertex by traversing some edges. Erasing a vertex will also erase the edges incident to it.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;R_v&amp;lt;/math&amp;gt; denotes the number of vertices from which vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt; is reachable. Prove that the expected value of the number of operations equals to &amp;lt;math&amp;gt;\sum_{v \in V} \frac{1}{R_v}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Expected Mex&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X_1,X_2,\ldots,X_{100} \sim \mathrm{Geo}(1/2)&amp;lt;/math&amp;gt; be independent random variables. Compute &amp;lt;math&amp;gt;\mathbf{E}[\mathrm{mex}(X_1,X_2,\ldots,X_{100})]&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\mathrm{mex}(a_1,a_2,\ldots,a_n)&amp;lt;/math&amp;gt; is the smallest positive integer that does not appear in &amp;lt;math&amp;gt;a_1,a_2,\ldots,a_n&amp;lt;/math&amp;gt;. Your answer is considered correct if the absolute error does not exceed &amp;lt;math&amp;gt;10^{-6}&amp;lt;/math&amp;gt;. (Hint: Although mathematics will help you arrive at elegant and efficient methods, the use of a computer and programming skills will be required.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 4 (Probability meets graph theory) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random social networks&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G = (V, E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph without isolating vertex. &lt;br /&gt;
Let &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; be the degree of vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Let &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be a uniformly chosen vertex, and &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; a uniformly chosen neighbor of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Show that &amp;lt;math&amp;gt;\mathbf{E}[d_Z] \geq \mathbf{E}[d_Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Interpret this inequality in the context of social networks, in which the vertices represent people, and the edges represent friendship.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Turán&#039;s Theorem&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G=(V,E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph, and write &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; for the degree of the vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Use probablistic method to prove that &amp;lt;math&amp;gt;\alpha(G) \ge \sum_{v \in V} \frac{1}{d_v + 1}&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\alpha(G)&amp;lt;/math&amp;gt; is the size of a maximum independent set. (Hint: Consider the following random procedure for generating an independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; from a graph with vertex set &amp;lt;math&amp;gt;V&amp;lt;/math&amp;gt;: First, generate a random permutation of the vertices, denoted as &amp;lt;math&amp;gt;v_1,v_2,\ldots,v_n&amp;lt;/math&amp;gt;. Then, construct the independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; as follows: For each vertex &amp;lt;math&amp;gt;v_i \in V&amp;lt;/math&amp;gt;, add &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; if and only if none of its predecessors in the permutation, i.e., &amp;lt;math&amp;gt;v_1,\ldots,v_{i-1}&amp;lt;/math&amp;gt;, are neighbors of &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Tournament&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Prove that there is a [https://en.wikipedia.org/wiki/Tournament_(graph_theory) tournament] &amp;lt;math&amp;gt;T&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; players and at least &amp;lt;math&amp;gt;n!/2^{n-1}&amp;lt;/math&amp;gt;Hamiltonian paths.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 5 (1D random walk, 8 points) ==&lt;br /&gt;
Let &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt; be a constant, and &amp;lt;math&amp;gt;\{X_n\}_{n \ge 1}&amp;lt;/math&amp;gt; be independent Bernoulli trials with successful probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;.&lt;br /&gt;
Define &amp;lt;math&amp;gt;S_n = 2\sum_{i=1}^n X_i - n&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;S_0 = 0&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Range of random walk&amp;lt;/strong&amp;gt;] The range &amp;lt;math&amp;gt;R_n&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;S_0, S_1, \ldots, S_n&amp;lt;/math&amp;gt; is defined as the number of distinct values taken by the sequence. Show that &amp;lt;math&amp;gt;\mathbf{Pr}\left(R_n = R_{n-1}+1\right) = \mathbf{Pr}\left(\forall 1 \le i \le n, S_i \neq 0\right)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, and deduce that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n]\to \mathbf{Pr}(\forall i \ge 1, S_i \neq 0)&amp;lt;/math&amp;gt;. Hence show that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n] \to |2p-1|&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Symmetric 1D random walk (III)&amp;lt;/strong&amp;gt;] Suppose &amp;lt;math&amp;gt;p = \frac{1}{2}&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{E}[|S_n|] = \Theta(\sqrt{n})&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12964</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 2</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12964"/>
		<updated>2025-03-13T15:43:23Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: /* Problem 1 (Warm-up problems, 12 points) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 2==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Warm-up problems, 16 points) ==&lt;br /&gt;
* [&#039;&#039;&#039;Cumulative distribution function (CDF) (I)&#039;&#039;&#039;] Express the distribution functions of [math]X^+ = \max\{0,X\}[/math], [math]X^- = -\min\{0,X\}[/math], [math]|X|=X^+ + X^-[/math], [math]-X[/math], in terms of the distribution function [math]F[/math] of the random variable [math]X[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Cumulative distribution function (CDF) (II)&#039;&#039;&#039;] Let [math]X[/math] be a random variable with distribution function [math]\max(0,\min(1,x))[/math]. Let [math]F[/math] be a distribution function which is continuous and strictly increasing. Show that [math]Y=F^{-1}(X)[/math] be a random variable with distribution function [math]F[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Probability density function (PDF)&#039;&#039;&#039;] We toss &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; coins, and each one shows heads with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;, independently of each of the others. Each coin which shows head is tossed again. (If the coin shows tail, it won&#039;t be tossed again.) Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of heads resulting from the &amp;lt;strong&amp;gt;second&amp;lt;/strong&amp;gt; round of tosses, and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the number of heads resulting from &amp;lt;strong&amp;gt;all&amp;lt;/strong&amp;gt; tosses, which includes the first and (possible) second round of each toss.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the PDF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{E}[Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Let &amp;lt;math&amp;gt;p_X&amp;lt;/math&amp;gt; be the PMF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, show that [math]p_X(k-1)p_X(k+1)\leq [p_X(k)]^2[/math] for &amp;lt;math&amp;gt;1\leq k \leq n-1&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
* [&#039;&#039;&#039;Independence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;1\leq r\leq n&amp;lt;/math&amp;gt; be independent random variables which are symmetric about &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;; that is, &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-X_r&amp;lt;/math&amp;gt; have the same distributions. Show that, for all &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}[S_n \geq x] = \mathbf{Pr}[S_n \leq -x]&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;S_n = \sum_{r=1}^n X_r&amp;lt;/math&amp;gt;. Is the conclusion true without the assumtion of independence?&lt;br /&gt;
* [&#039;&#039;&#039;Dependence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be discrete random variables with joint mass function &amp;lt;math&amp;gt;f(x,y) = \frac{C}{(x+y-1)(x+y)(x+y+1)}&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;x,y \in \mathbb{N}_+&amp;lt;/math&amp;gt; (in other words, &amp;lt;math&amp;gt;x,y = 1,2,3,\cdots&amp;lt;/math&amp;gt;). Find (1) the value of &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;, (2) marginal mass function of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and (3) &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt;.&lt;br /&gt;
* [&#039;&#039;&#039;Expectation&#039;&#039;&#039;] It is required to place in order &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; books &amp;lt;math&amp;gt;B_1, B_2, \cdots, B_n&amp;lt;/math&amp;gt; on a library shelf in such way that readers searching from left to right waste as little time as possible on average. Assuming that a random reader requires book &amp;lt;math&amp;gt;B_i&amp;lt;/math&amp;gt; with probability &amp;lt;math&amp;gt;p_i&amp;lt;/math&amp;gt;, find the ordering of the books which minimizes the number of titles examined by a random reader before discovery of the required book.&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Law of total expectation&amp;lt;/strong&amp;gt;] Let [math]X \sim \mathrm{Geom}(p)[/math] for some parameter [math]p \in (0,1)[/math]. Calculate [math]\mathbf{E}[X][/math] using the law of total expectation.&lt;br /&gt;
&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Random number of random variables&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\{X_n\}_{n \ge 1}&amp;lt;/math&amp;gt; be identically distributed random variable and &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; be a random variable taking values in the non-negative integers and independent of the &amp;lt;math&amp;gt;X_n&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;n \ge 1&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{E}\left[\sum_{i=1}^N X_i\right] = \mathbf{E}[N] \mathbf{E}[X_1]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Discrete random variable) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (I)&amp;lt;/strong&amp;gt;] Every package of some intrinsically dull commodity includes a small and exciting plastic object. There are [math]c[/math] different types of object, and each package is equally likely to contain any given type. You buy one package each day.&amp;lt;/p&amp;gt;&lt;br /&gt;
    &amp;lt;ol&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse between the acquisitions of the [math]j[/math]-th new type of object and the [math](j + 1)[/math]-th new type.&amp;lt;/li&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse before you have a full set of objects.&amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (II)&amp;lt;/strong&amp;gt;] Prove that geometric distribution is the only discrete memoryless distribution with range values &amp;lt;math&amp;gt;\mathbb{N}_+&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;n_1,n_2 \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;0 \le p \le 1&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Bin}(n_1,p),Y \sim \mathrm{Bin}(n_2,p)&amp;lt;/math&amp;gt; be independent random variables. Prove that &amp;lt;math&amp;gt;X+Y \sim \mathrm{Bin}(n_1+n_2,p)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Negative binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; follows the negative binomial distribution with parameter &amp;lt;math&amp;gt;r \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt;. Calculate &amp;lt;math&amp;gt;\mathbf{Var}[X] = \mathbf{E}[X^2] - \left(\mathbf{E}[X]\right)^2&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Hypergeometric distribution&amp;lt;/strong&amp;gt;] An urn contains [math]N[/math] balls, [math]b[/math] of which are blue and [math]r = N -b[/math] of which are red. A random sample of [math]n[/math] balls is drawn &amp;lt;strong&amp;gt;without replacement&amp;lt;/strong&amp;gt; (无放回) from the urn. Let [math]B[/math] the number of blue balls in this sample. Show that if [math]N, b[/math], and [math]r[/math] approach [math]+\infty[/math] in such a way that [math]b/N \rightarrow  p[/math] and [math]r/N \rightarrow 1 - p[/math], then [math]\mathbf{Pr}(B = k) \rightarrow {n\choose k}p^k(1-p)^{n-k}[/math] for [math]0\leq k \leq n[/math].&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Poisson distribution&amp;lt;/strong&amp;gt;] In your pocket is a random number &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; of coins, where &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; has the Poisson distribution with parameter &amp;lt;math&amp;gt;\lambda&amp;lt;/math&amp;gt;. You toss each coin once, with heads showing with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; each time. Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the (random) number of heads outcomes and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the (also random) number of tails.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the joint mass function of &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find PDF of the marginal distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; in &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;. Are &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; independent?&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Conditional distribution&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;\lambda,\mu &amp;gt; 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;n \in \mathbb{N}&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Pois}(\lambda), Y \sim \mathrm{Pois}(\mu)&amp;lt;/math&amp;gt; be independent random variables. Find out the conditional distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, given &amp;lt;math&amp;gt;X+Y = n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Linearity of Expectation, 12 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Streak&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Suppose we flip a fair coin &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; times independently to obtain a sequence of flips &amp;lt;math&amp;gt;X_1, X_2, \ldots , X_n&amp;lt;/math&amp;gt;. A streak of flips is a consecutive subsequence of flips that are all the same. For example, if &amp;lt;math&amp;gt;X_3&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;X_4&amp;lt;/math&amp;gt;, and &amp;lt;math&amp;gt;X_5&amp;lt;/math&amp;gt; are all heads, there is a streak of length &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; starting at the third lip. (If &amp;lt;math&amp;gt;X_6&amp;lt;/math&amp;gt; is also heads, then there is also a streak of length &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; starting at the third lip.) Find the expected number of streaks of length &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; for some integer &amp;lt;math&amp;gt;k \ge 1&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Number of cycles&amp;lt;/strong&amp;gt;] &lt;br /&gt;
At a banquet, there are &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people who shake hands according to the following process: In each round, two idle hands are randomly selected and shaken (&amp;lt;strong&amp;gt;these two hands are no longer idle&amp;lt;/strong&amp;gt;). After &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds, there will be no idle hands left, and the &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people will form several cycles. For example, when &amp;lt;math&amp;gt;n=3&amp;lt;/math&amp;gt;, the following situation may occur: the left and right hands of the first person are held together, the left hand of the second person and the right hand of the third person are held together, and the right hand of the second person and the left hand of the third person are held together. In this case, three people form two cycles. How many cycles are expected to be formed after &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds?&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Number of operations&amp;lt;/strong&amp;gt;]&lt;br /&gt;
We have a directed graph &amp;lt;math&amp;gt;G=(V,E)&amp;lt;/math&amp;gt; without self-loops and multi-edges.&lt;br /&gt;
&lt;br /&gt;
Until the graph becomes empty, repeat the following operation:&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
Choose one (unerased) vertex uniformly at random (independently from the previous choices). Then, erase that vertex and all vertices that are reachable from the chosen vertex by traversing some edges. Erasing a vertex will also erase the edges incident to it.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;R_v&amp;lt;/math&amp;gt; denotes the number of vertices from which vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt; is reachable. Prove that the expected value of the number of operations equals to &amp;lt;math&amp;gt;\sum_{v \in V} \frac{1}{R_v}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Expected Mex&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X_1,X_2,\ldots,X_{100} \sim \mathrm{Geo}(1/2)&amp;lt;/math&amp;gt; be independent random variables. Compute &amp;lt;math&amp;gt;\mathbf{E}[\mathrm{mex}(X_1,X_2,\ldots,X_{100})]&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\mathrm{mex}(a_1,a_2,\ldots,a_n)&amp;lt;/math&amp;gt; is the smallest positive integer that does not appear in &amp;lt;math&amp;gt;a_1,a_2,\ldots,a_n&amp;lt;/math&amp;gt;. Your answer is considered correct if the absolute error does not exceed &amp;lt;math&amp;gt;10^{-6}&amp;lt;/math&amp;gt;. (Hint: Although mathematics will help you arrive at elegant and efficient methods, the use of a computer and programming skills will be required.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 4 (Probability meets graph theory) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random social networks&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G = (V, E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph without isolating vertex. &lt;br /&gt;
Let &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; be the degree of vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Let &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be a uniformly chosen vertex, and &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; a uniformly chosen neighbor of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Show that &amp;lt;math&amp;gt;\mathbf{E}[d_Z] \geq \mathbf{E}[d_Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Interpret this inequality in the context of social networks, in which the vertices represent people, and the edges represent friendship.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Turán&#039;s Theorem&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G=(V,E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph, and write &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; for the degree of the vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Use probablistic method to prove that &amp;lt;math&amp;gt;\alpha(G) \ge \sum_{v \in V} \frac{1}{d_v + 1}&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\alpha(G)&amp;lt;/math&amp;gt; is the size of a maximum independent set. (Hint: Consider the following random procedure for generating an independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; from a graph with vertex set &amp;lt;math&amp;gt;V&amp;lt;/math&amp;gt;: First, generate a random permutation of the vertices, denoted as &amp;lt;math&amp;gt;v_1,v_2,\ldots,v_n&amp;lt;/math&amp;gt;. Then, construct the independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; as follows: For each vertex &amp;lt;math&amp;gt;v_i \in V&amp;lt;/math&amp;gt;, add &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; if and only if none of its predecessors in the permutation, i.e., &amp;lt;math&amp;gt;v_1,\ldots,v_{i-1}&amp;lt;/math&amp;gt;, are neighbors of &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Tournament&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Prove that there is a [https://en.wikipedia.org/wiki/Tournament_(graph_theory) tournament] &amp;lt;math&amp;gt;T&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; players and at least &amp;lt;math&amp;gt;n!/2^{n-1}&amp;lt;/math&amp;gt;Hamiltonian paths.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 5 (1D random walk, 8 points) ==&lt;br /&gt;
Let &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt; be a constant, and &amp;lt;math&amp;gt;\{X_n\}_{n \ge 1}&amp;lt;/math&amp;gt; be independent Bernoulli trials with successful probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;.&lt;br /&gt;
Define &amp;lt;math&amp;gt;S_n = 2\sum_{i=1}^n X_i - n&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;S_0 = 0&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Range of random walk&amp;lt;/strong&amp;gt;] The range &amp;lt;math&amp;gt;R_n&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;S_0, S_1, \ldots, S_n&amp;lt;/math&amp;gt; is defined as the number of distinct values taken by the sequence. Show that &amp;lt;math&amp;gt;\mathbf{Pr}\left(R_n = R_{n-1}+1\right) = \mathbf{Pr}\left(\forall 1 \le i \le n, S_i \neq 0\right)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, and deduce that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n]\to \mathbf{Pr}(\forall i \ge 1, S_i \neq 0)&amp;lt;/math&amp;gt;. Hence show that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n] \to |2p-1|&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Symmetric 1D random walk (III)&amp;lt;/strong&amp;gt;] Suppose &amp;lt;math&amp;gt;p = \frac{1}{2}&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{E}[|S_n|] = \Theta(\sqrt{n})&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12963</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 2</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12963"/>
		<updated>2025-03-13T15:43:03Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: /* Problem 1 (Warm-up problems, 12 points) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 2==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Warm-up problems, 12 points) ==&lt;br /&gt;
* [&#039;&#039;&#039;Cumulative distribution function (CDF) (I)&#039;&#039;&#039;] Express the distribution functions of [math]X^+ = \max\{0,X\}[/math], [math]X^- = -\min\{0,X\}[/math], [math]|X|=X^+ + X^-[/math], [math]-X[/math], in terms of the distribution function [math]F[/math] of the random variable [math]X[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Cumulative distribution function (CDF) (II)&#039;&#039;&#039;] Let [math]X[/math] be a random variable with distribution function [math]\max(0,\min(1,x))[/math]. Let [math]F[/math] be a distribution function which is continuous and strictly increasing. Show that [math]Y=F^{-1}(X)[/math] be a random variable with distribution function [math]F[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Probability density function (PDF)&#039;&#039;&#039;] We toss &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; coins, and each one shows heads with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;, independently of each of the others. Each coin which shows head is tossed again. (If the coin shows tail, it won&#039;t be tossed again.) Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of heads resulting from the &amp;lt;strong&amp;gt;second&amp;lt;/strong&amp;gt; round of tosses, and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the number of heads resulting from &amp;lt;strong&amp;gt;all&amp;lt;/strong&amp;gt; tosses, which includes the first and (possible) second round of each toss.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the PDF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{E}[Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Let &amp;lt;math&amp;gt;p_X&amp;lt;/math&amp;gt; be the PMF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, show that [math]p_X(k-1)p_X(k+1)\leq [p_X(k)]^2[/math] for &amp;lt;math&amp;gt;1\leq k \leq n-1&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
* [&#039;&#039;&#039;Independence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;1\leq r\leq n&amp;lt;/math&amp;gt; be independent random variables which are symmetric about &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;; that is, &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-X_r&amp;lt;/math&amp;gt; have the same distributions. Show that, for all &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}[S_n \geq x] = \mathbf{Pr}[S_n \leq -x]&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;S_n = \sum_{r=1}^n X_r&amp;lt;/math&amp;gt;. Is the conclusion true without the assumtion of independence?&lt;br /&gt;
* [&#039;&#039;&#039;Dependence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be discrete random variables with joint mass function &amp;lt;math&amp;gt;f(x,y) = \frac{C}{(x+y-1)(x+y)(x+y+1)}&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;x,y \in \mathbb{N}_+&amp;lt;/math&amp;gt; (in other words, &amp;lt;math&amp;gt;x,y = 1,2,3,\cdots&amp;lt;/math&amp;gt;). Find (1) the value of &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;, (2) marginal mass function of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and (3) &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt;.&lt;br /&gt;
* [&#039;&#039;&#039;Expectation&#039;&#039;&#039;] It is required to place in order &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; books &amp;lt;math&amp;gt;B_1, B_2, \cdots, B_n&amp;lt;/math&amp;gt; on a library shelf in such way that readers searching from left to right waste as little time as possible on average. Assuming that a random reader requires book &amp;lt;math&amp;gt;B_i&amp;lt;/math&amp;gt; with probability &amp;lt;math&amp;gt;p_i&amp;lt;/math&amp;gt;, find the ordering of the books which minimizes the number of titles examined by a random reader before discovery of the required book.&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Law of total expectation&amp;lt;/strong&amp;gt;] Let [math]X \sim \mathrm{Geom}(p)[/math] for some parameter [math]p \in (0,1)[/math]. Calculate [math]\mathbf{E}[X][/math] using the law of total expectation.&lt;br /&gt;
&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Random number of random variables&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\{X_n\}_{n \ge 1}&amp;lt;/math&amp;gt; be identically distributed random variable and &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; be a random variable taking values in the non-negative integers and independent of the &amp;lt;math&amp;gt;X_n&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;n \ge 1&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{E}\left[\sum_{i=1}^N X_i\right] = \mathbf{E}[N] \mathbf{E}[X_1]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Discrete random variable) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (I)&amp;lt;/strong&amp;gt;] Every package of some intrinsically dull commodity includes a small and exciting plastic object. There are [math]c[/math] different types of object, and each package is equally likely to contain any given type. You buy one package each day.&amp;lt;/p&amp;gt;&lt;br /&gt;
    &amp;lt;ol&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse between the acquisitions of the [math]j[/math]-th new type of object and the [math](j + 1)[/math]-th new type.&amp;lt;/li&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse before you have a full set of objects.&amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (II)&amp;lt;/strong&amp;gt;] Prove that geometric distribution is the only discrete memoryless distribution with range values &amp;lt;math&amp;gt;\mathbb{N}_+&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;n_1,n_2 \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;0 \le p \le 1&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Bin}(n_1,p),Y \sim \mathrm{Bin}(n_2,p)&amp;lt;/math&amp;gt; be independent random variables. Prove that &amp;lt;math&amp;gt;X+Y \sim \mathrm{Bin}(n_1+n_2,p)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Negative binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; follows the negative binomial distribution with parameter &amp;lt;math&amp;gt;r \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt;. Calculate &amp;lt;math&amp;gt;\mathbf{Var}[X] = \mathbf{E}[X^2] - \left(\mathbf{E}[X]\right)^2&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Hypergeometric distribution&amp;lt;/strong&amp;gt;] An urn contains [math]N[/math] balls, [math]b[/math] of which are blue and [math]r = N -b[/math] of which are red. A random sample of [math]n[/math] balls is drawn &amp;lt;strong&amp;gt;without replacement&amp;lt;/strong&amp;gt; (无放回) from the urn. Let [math]B[/math] the number of blue balls in this sample. Show that if [math]N, b[/math], and [math]r[/math] approach [math]+\infty[/math] in such a way that [math]b/N \rightarrow  p[/math] and [math]r/N \rightarrow 1 - p[/math], then [math]\mathbf{Pr}(B = k) \rightarrow {n\choose k}p^k(1-p)^{n-k}[/math] for [math]0\leq k \leq n[/math].&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Poisson distribution&amp;lt;/strong&amp;gt;] In your pocket is a random number &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; of coins, where &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; has the Poisson distribution with parameter &amp;lt;math&amp;gt;\lambda&amp;lt;/math&amp;gt;. You toss each coin once, with heads showing with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; each time. Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the (random) number of heads outcomes and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the (also random) number of tails.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the joint mass function of &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find PDF of the marginal distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; in &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;. Are &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; independent?&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Conditional distribution&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;\lambda,\mu &amp;gt; 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;n \in \mathbb{N}&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Pois}(\lambda), Y \sim \mathrm{Pois}(\mu)&amp;lt;/math&amp;gt; be independent random variables. Find out the conditional distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, given &amp;lt;math&amp;gt;X+Y = n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Linearity of Expectation, 12 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Streak&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Suppose we flip a fair coin &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; times independently to obtain a sequence of flips &amp;lt;math&amp;gt;X_1, X_2, \ldots , X_n&amp;lt;/math&amp;gt;. A streak of flips is a consecutive subsequence of flips that are all the same. For example, if &amp;lt;math&amp;gt;X_3&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;X_4&amp;lt;/math&amp;gt;, and &amp;lt;math&amp;gt;X_5&amp;lt;/math&amp;gt; are all heads, there is a streak of length &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; starting at the third lip. (If &amp;lt;math&amp;gt;X_6&amp;lt;/math&amp;gt; is also heads, then there is also a streak of length &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; starting at the third lip.) Find the expected number of streaks of length &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; for some integer &amp;lt;math&amp;gt;k \ge 1&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Number of cycles&amp;lt;/strong&amp;gt;] &lt;br /&gt;
At a banquet, there are &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people who shake hands according to the following process: In each round, two idle hands are randomly selected and shaken (&amp;lt;strong&amp;gt;these two hands are no longer idle&amp;lt;/strong&amp;gt;). After &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds, there will be no idle hands left, and the &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people will form several cycles. For example, when &amp;lt;math&amp;gt;n=3&amp;lt;/math&amp;gt;, the following situation may occur: the left and right hands of the first person are held together, the left hand of the second person and the right hand of the third person are held together, and the right hand of the second person and the left hand of the third person are held together. In this case, three people form two cycles. How many cycles are expected to be formed after &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds?&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Number of operations&amp;lt;/strong&amp;gt;]&lt;br /&gt;
We have a directed graph &amp;lt;math&amp;gt;G=(V,E)&amp;lt;/math&amp;gt; without self-loops and multi-edges.&lt;br /&gt;
&lt;br /&gt;
Until the graph becomes empty, repeat the following operation:&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
Choose one (unerased) vertex uniformly at random (independently from the previous choices). Then, erase that vertex and all vertices that are reachable from the chosen vertex by traversing some edges. Erasing a vertex will also erase the edges incident to it.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;R_v&amp;lt;/math&amp;gt; denotes the number of vertices from which vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt; is reachable. Prove that the expected value of the number of operations equals to &amp;lt;math&amp;gt;\sum_{v \in V} \frac{1}{R_v}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Expected Mex&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X_1,X_2,\ldots,X_{100} \sim \mathrm{Geo}(1/2)&amp;lt;/math&amp;gt; be independent random variables. Compute &amp;lt;math&amp;gt;\mathbf{E}[\mathrm{mex}(X_1,X_2,\ldots,X_{100})]&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\mathrm{mex}(a_1,a_2,\ldots,a_n)&amp;lt;/math&amp;gt; is the smallest positive integer that does not appear in &amp;lt;math&amp;gt;a_1,a_2,\ldots,a_n&amp;lt;/math&amp;gt;. Your answer is considered correct if the absolute error does not exceed &amp;lt;math&amp;gt;10^{-6}&amp;lt;/math&amp;gt;. (Hint: Although mathematics will help you arrive at elegant and efficient methods, the use of a computer and programming skills will be required.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 4 (Probability meets graph theory) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random social networks&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G = (V, E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph without isolating vertex. &lt;br /&gt;
Let &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; be the degree of vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Let &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be a uniformly chosen vertex, and &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; a uniformly chosen neighbor of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Show that &amp;lt;math&amp;gt;\mathbf{E}[d_Z] \geq \mathbf{E}[d_Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Interpret this inequality in the context of social networks, in which the vertices represent people, and the edges represent friendship.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Turán&#039;s Theorem&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G=(V,E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph, and write &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; for the degree of the vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Use probablistic method to prove that &amp;lt;math&amp;gt;\alpha(G) \ge \sum_{v \in V} \frac{1}{d_v + 1}&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\alpha(G)&amp;lt;/math&amp;gt; is the size of a maximum independent set. (Hint: Consider the following random procedure for generating an independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; from a graph with vertex set &amp;lt;math&amp;gt;V&amp;lt;/math&amp;gt;: First, generate a random permutation of the vertices, denoted as &amp;lt;math&amp;gt;v_1,v_2,\ldots,v_n&amp;lt;/math&amp;gt;. Then, construct the independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; as follows: For each vertex &amp;lt;math&amp;gt;v_i \in V&amp;lt;/math&amp;gt;, add &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; if and only if none of its predecessors in the permutation, i.e., &amp;lt;math&amp;gt;v_1,\ldots,v_{i-1}&amp;lt;/math&amp;gt;, are neighbors of &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Tournament&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Prove that there is a [https://en.wikipedia.org/wiki/Tournament_(graph_theory) tournament] &amp;lt;math&amp;gt;T&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; players and at least &amp;lt;math&amp;gt;n!/2^{n-1}&amp;lt;/math&amp;gt;Hamiltonian paths.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 5 (1D random walk, 8 points) ==&lt;br /&gt;
Let &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt; be a constant, and &amp;lt;math&amp;gt;\{X_n\}_{n \ge 1}&amp;lt;/math&amp;gt; be independent Bernoulli trials with successful probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;.&lt;br /&gt;
Define &amp;lt;math&amp;gt;S_n = 2\sum_{i=1}^n X_i - n&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;S_0 = 0&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Range of random walk&amp;lt;/strong&amp;gt;] The range &amp;lt;math&amp;gt;R_n&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;S_0, S_1, \ldots, S_n&amp;lt;/math&amp;gt; is defined as the number of distinct values taken by the sequence. Show that &amp;lt;math&amp;gt;\mathbf{Pr}\left(R_n = R_{n-1}+1\right) = \mathbf{Pr}\left(\forall 1 \le i \le n, S_i \neq 0\right)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, and deduce that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n]\to \mathbf{Pr}(\forall i \ge 1, S_i \neq 0)&amp;lt;/math&amp;gt;. Hence show that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n] \to |2p-1|&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Symmetric 1D random walk (III)&amp;lt;/strong&amp;gt;] Suppose &amp;lt;math&amp;gt;p = \frac{1}{2}&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{E}[|S_n|] = \Theta(\sqrt{n})&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12962</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 2</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12962"/>
		<updated>2025-03-13T15:42:43Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: /* Problem 2 (Discrete random variable) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 2==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Warm-up problems, 12 points) ==&lt;br /&gt;
* [&#039;&#039;&#039;Cumulative distribution function (CDF) (I)&#039;&#039;&#039;] Express the distribution functions of [math]X^+ = \max\{0,X\}[/math], [math]X^- = -\min\{0,X\}[/math], [math]|X|=X^+ + X^-[/math], [math]-X[/math], in terms of the distribution function [math]F[/math] of the random variable [math]X[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Cumulative distribution function (CDF) (II)&#039;&#039;&#039;] Let [math]X[/math] be a random variable with distribution function [math]\max(0,\min(1,x))[/math]. Let [math]F[/math] be a distribution function which is continuous and strictly increasing. Show that [math]Y=F^{-1}(X)[/math] be a random variable with distribution function [math]F[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Probability density function (PDF)&#039;&#039;&#039;] We toss &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; coins, and each one shows heads with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;, independently of each of the others. Each coin which shows head is tossed again. (If the coin shows tail, it won&#039;t be tossed again.) Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of heads resulting from the &amp;lt;strong&amp;gt;second&amp;lt;/strong&amp;gt; round of tosses, and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the number of heads resulting from &amp;lt;strong&amp;gt;all&amp;lt;/strong&amp;gt; tosses, which includes the first and (possible) second round of each toss.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the PDF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{E}[Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Let &amp;lt;math&amp;gt;p_X&amp;lt;/math&amp;gt; be the PMF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, show that [math]p_X(k-1)p_X(k+1)\leq [p_X(k)]^2[/math] for &amp;lt;math&amp;gt;1\leq k \leq n-1&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
* [&#039;&#039;&#039;Independence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;1\leq r\leq n&amp;lt;/math&amp;gt; be independent random variables which are symmetric about &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;; that is, &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-X_r&amp;lt;/math&amp;gt; have the same distributions. Show that, for all &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}[S_n \geq x] = \mathbf{Pr}[S_n \leq -x]&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;S_n = \sum_{r=1}^n X_r&amp;lt;/math&amp;gt;. Is the conclusion true without the assumtion of independence?&lt;br /&gt;
* [&#039;&#039;&#039;Dependence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be discrete random variables with joint mass function &amp;lt;math&amp;gt;f(x,y) = \frac{C}{(x+y-1)(x+y)(x+y+1)}&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;x,y \in \mathbb{N}_+&amp;lt;/math&amp;gt; (in other words, &amp;lt;math&amp;gt;x,y = 1,2,3,\cdots&amp;lt;/math&amp;gt;). Find (1) the value of &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;, (2) marginal mass function of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and (3) &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt;.&lt;br /&gt;
* [&#039;&#039;&#039;Expectation&#039;&#039;&#039;] It is required to place in order &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; books &amp;lt;math&amp;gt;B_1, B_2, \cdots, B_n&amp;lt;/math&amp;gt; on a library shelf in such way that readers searching from left to right waste as little time as possible on average. Assuming that a random reader requires book &amp;lt;math&amp;gt;B_i&amp;lt;/math&amp;gt; with probability &amp;lt;math&amp;gt;p_i&amp;lt;/math&amp;gt;, find the ordering of the books which minimizes the number of titles examined by a random reader before discovery of the required book.&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Entropy of discrete random variable&amp;lt;/strong&amp;gt;] Let [math]X[/math] be a discrete random variable with range of values [math][N] = \{1,2,\ldots,N\}[/math] and probability mass function [math]p[/math]. Define [math]H(X) = -\sum_{n \ge 1} p(n) \log p(n)[/math] with convention [math]0\log 0 = 0[/math]. Prove that [math]H(X) \le \log N[/math] using Jensen&#039;s inequality.&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Law of total expectation&amp;lt;/strong&amp;gt;] Let [math]X \sim \mathrm{Geom}(p)[/math] for some parameter [math]p \in (0,1)[/math]. Calculate [math]\mathbf{E}[X][/math] using the law of total expectation.&lt;br /&gt;
&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Random number of random variables&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\{X_n\}_{n \ge 1}&amp;lt;/math&amp;gt; be identically distributed random variable and &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; be a random variable taking values in the non-negative integers and independent of the &amp;lt;math&amp;gt;X_n&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;n \ge 1&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{E}\left[\sum_{i=1}^N X_i\right] = \mathbf{E}[N] \mathbf{E}[X_1]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Discrete random variable) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (I)&amp;lt;/strong&amp;gt;] Every package of some intrinsically dull commodity includes a small and exciting plastic object. There are [math]c[/math] different types of object, and each package is equally likely to contain any given type. You buy one package each day.&amp;lt;/p&amp;gt;&lt;br /&gt;
    &amp;lt;ol&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse between the acquisitions of the [math]j[/math]-th new type of object and the [math](j + 1)[/math]-th new type.&amp;lt;/li&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse before you have a full set of objects.&amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (II)&amp;lt;/strong&amp;gt;] Prove that geometric distribution is the only discrete memoryless distribution with range values &amp;lt;math&amp;gt;\mathbb{N}_+&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;n_1,n_2 \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;0 \le p \le 1&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Bin}(n_1,p),Y \sim \mathrm{Bin}(n_2,p)&amp;lt;/math&amp;gt; be independent random variables. Prove that &amp;lt;math&amp;gt;X+Y \sim \mathrm{Bin}(n_1+n_2,p)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Negative binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; follows the negative binomial distribution with parameter &amp;lt;math&amp;gt;r \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt;. Calculate &amp;lt;math&amp;gt;\mathbf{Var}[X] = \mathbf{E}[X^2] - \left(\mathbf{E}[X]\right)^2&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Hypergeometric distribution&amp;lt;/strong&amp;gt;] An urn contains [math]N[/math] balls, [math]b[/math] of which are blue and [math]r = N -b[/math] of which are red. A random sample of [math]n[/math] balls is drawn &amp;lt;strong&amp;gt;without replacement&amp;lt;/strong&amp;gt; (无放回) from the urn. Let [math]B[/math] the number of blue balls in this sample. Show that if [math]N, b[/math], and [math]r[/math] approach [math]+\infty[/math] in such a way that [math]b/N \rightarrow  p[/math] and [math]r/N \rightarrow 1 - p[/math], then [math]\mathbf{Pr}(B = k) \rightarrow {n\choose k}p^k(1-p)^{n-k}[/math] for [math]0\leq k \leq n[/math].&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Poisson distribution&amp;lt;/strong&amp;gt;] In your pocket is a random number &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; of coins, where &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; has the Poisson distribution with parameter &amp;lt;math&amp;gt;\lambda&amp;lt;/math&amp;gt;. You toss each coin once, with heads showing with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; each time. Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the (random) number of heads outcomes and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the (also random) number of tails.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the joint mass function of &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find PDF of the marginal distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; in &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;. Are &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; independent?&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Conditional distribution&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;\lambda,\mu &amp;gt; 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;n \in \mathbb{N}&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Pois}(\lambda), Y \sim \mathrm{Pois}(\mu)&amp;lt;/math&amp;gt; be independent random variables. Find out the conditional distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, given &amp;lt;math&amp;gt;X+Y = n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Linearity of Expectation, 12 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Streak&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Suppose we flip a fair coin &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; times independently to obtain a sequence of flips &amp;lt;math&amp;gt;X_1, X_2, \ldots , X_n&amp;lt;/math&amp;gt;. A streak of flips is a consecutive subsequence of flips that are all the same. For example, if &amp;lt;math&amp;gt;X_3&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;X_4&amp;lt;/math&amp;gt;, and &amp;lt;math&amp;gt;X_5&amp;lt;/math&amp;gt; are all heads, there is a streak of length &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; starting at the third lip. (If &amp;lt;math&amp;gt;X_6&amp;lt;/math&amp;gt; is also heads, then there is also a streak of length &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; starting at the third lip.) Find the expected number of streaks of length &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; for some integer &amp;lt;math&amp;gt;k \ge 1&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Number of cycles&amp;lt;/strong&amp;gt;] &lt;br /&gt;
At a banquet, there are &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people who shake hands according to the following process: In each round, two idle hands are randomly selected and shaken (&amp;lt;strong&amp;gt;these two hands are no longer idle&amp;lt;/strong&amp;gt;). After &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds, there will be no idle hands left, and the &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people will form several cycles. For example, when &amp;lt;math&amp;gt;n=3&amp;lt;/math&amp;gt;, the following situation may occur: the left and right hands of the first person are held together, the left hand of the second person and the right hand of the third person are held together, and the right hand of the second person and the left hand of the third person are held together. In this case, three people form two cycles. How many cycles are expected to be formed after &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds?&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Number of operations&amp;lt;/strong&amp;gt;]&lt;br /&gt;
We have a directed graph &amp;lt;math&amp;gt;G=(V,E)&amp;lt;/math&amp;gt; without self-loops and multi-edges.&lt;br /&gt;
&lt;br /&gt;
Until the graph becomes empty, repeat the following operation:&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
Choose one (unerased) vertex uniformly at random (independently from the previous choices). Then, erase that vertex and all vertices that are reachable from the chosen vertex by traversing some edges. Erasing a vertex will also erase the edges incident to it.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;R_v&amp;lt;/math&amp;gt; denotes the number of vertices from which vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt; is reachable. Prove that the expected value of the number of operations equals to &amp;lt;math&amp;gt;\sum_{v \in V} \frac{1}{R_v}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Expected Mex&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X_1,X_2,\ldots,X_{100} \sim \mathrm{Geo}(1/2)&amp;lt;/math&amp;gt; be independent random variables. Compute &amp;lt;math&amp;gt;\mathbf{E}[\mathrm{mex}(X_1,X_2,\ldots,X_{100})]&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\mathrm{mex}(a_1,a_2,\ldots,a_n)&amp;lt;/math&amp;gt; is the smallest positive integer that does not appear in &amp;lt;math&amp;gt;a_1,a_2,\ldots,a_n&amp;lt;/math&amp;gt;. Your answer is considered correct if the absolute error does not exceed &amp;lt;math&amp;gt;10^{-6}&amp;lt;/math&amp;gt;. (Hint: Although mathematics will help you arrive at elegant and efficient methods, the use of a computer and programming skills will be required.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 4 (Probability meets graph theory) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random social networks&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G = (V, E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph without isolating vertex. &lt;br /&gt;
Let &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; be the degree of vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Let &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be a uniformly chosen vertex, and &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; a uniformly chosen neighbor of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Show that &amp;lt;math&amp;gt;\mathbf{E}[d_Z] \geq \mathbf{E}[d_Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Interpret this inequality in the context of social networks, in which the vertices represent people, and the edges represent friendship.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Turán&#039;s Theorem&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G=(V,E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph, and write &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; for the degree of the vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Use probablistic method to prove that &amp;lt;math&amp;gt;\alpha(G) \ge \sum_{v \in V} \frac{1}{d_v + 1}&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\alpha(G)&amp;lt;/math&amp;gt; is the size of a maximum independent set. (Hint: Consider the following random procedure for generating an independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; from a graph with vertex set &amp;lt;math&amp;gt;V&amp;lt;/math&amp;gt;: First, generate a random permutation of the vertices, denoted as &amp;lt;math&amp;gt;v_1,v_2,\ldots,v_n&amp;lt;/math&amp;gt;. Then, construct the independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; as follows: For each vertex &amp;lt;math&amp;gt;v_i \in V&amp;lt;/math&amp;gt;, add &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; if and only if none of its predecessors in the permutation, i.e., &amp;lt;math&amp;gt;v_1,\ldots,v_{i-1}&amp;lt;/math&amp;gt;, are neighbors of &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Tournament&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Prove that there is a [https://en.wikipedia.org/wiki/Tournament_(graph_theory) tournament] &amp;lt;math&amp;gt;T&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; players and at least &amp;lt;math&amp;gt;n!/2^{n-1}&amp;lt;/math&amp;gt;Hamiltonian paths.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 5 (1D random walk, 8 points) ==&lt;br /&gt;
Let &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt; be a constant, and &amp;lt;math&amp;gt;\{X_n\}_{n \ge 1}&amp;lt;/math&amp;gt; be independent Bernoulli trials with successful probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;.&lt;br /&gt;
Define &amp;lt;math&amp;gt;S_n = 2\sum_{i=1}^n X_i - n&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;S_0 = 0&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Range of random walk&amp;lt;/strong&amp;gt;] The range &amp;lt;math&amp;gt;R_n&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;S_0, S_1, \ldots, S_n&amp;lt;/math&amp;gt; is defined as the number of distinct values taken by the sequence. Show that &amp;lt;math&amp;gt;\mathbf{Pr}\left(R_n = R_{n-1}+1\right) = \mathbf{Pr}\left(\forall 1 \le i \le n, S_i \neq 0\right)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, and deduce that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n]\to \mathbf{Pr}(\forall i \ge 1, S_i \neq 0)&amp;lt;/math&amp;gt;. Hence show that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n] \to |2p-1|&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Symmetric 1D random walk (III)&amp;lt;/strong&amp;gt;] Suppose &amp;lt;math&amp;gt;p = \frac{1}{2}&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{E}[|S_n|] = \Theta(\sqrt{n})&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12961</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 2</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12961"/>
		<updated>2025-03-13T15:39:14Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 2==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Warm-up problems, 12 points) ==&lt;br /&gt;
* [&#039;&#039;&#039;Cumulative distribution function (CDF) (I)&#039;&#039;&#039;] Express the distribution functions of [math]X^+ = \max\{0,X\}[/math], [math]X^- = -\min\{0,X\}[/math], [math]|X|=X^+ + X^-[/math], [math]-X[/math], in terms of the distribution function [math]F[/math] of the random variable [math]X[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Cumulative distribution function (CDF) (II)&#039;&#039;&#039;] Let [math]X[/math] be a random variable with distribution function [math]\max(0,\min(1,x))[/math]. Let [math]F[/math] be a distribution function which is continuous and strictly increasing. Show that [math]Y=F^{-1}(X)[/math] be a random variable with distribution function [math]F[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Probability density function (PDF)&#039;&#039;&#039;] We toss &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; coins, and each one shows heads with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;, independently of each of the others. Each coin which shows head is tossed again. (If the coin shows tail, it won&#039;t be tossed again.) Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of heads resulting from the &amp;lt;strong&amp;gt;second&amp;lt;/strong&amp;gt; round of tosses, and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the number of heads resulting from &amp;lt;strong&amp;gt;all&amp;lt;/strong&amp;gt; tosses, which includes the first and (possible) second round of each toss.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the PDF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{E}[Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Let &amp;lt;math&amp;gt;p_X&amp;lt;/math&amp;gt; be the PMF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, show that [math]p_X(k-1)p_X(k+1)\leq [p_X(k)]^2[/math] for &amp;lt;math&amp;gt;1\leq k \leq n-1&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
* [&#039;&#039;&#039;Independence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;1\leq r\leq n&amp;lt;/math&amp;gt; be independent random variables which are symmetric about &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;; that is, &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-X_r&amp;lt;/math&amp;gt; have the same distributions. Show that, for all &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}[S_n \geq x] = \mathbf{Pr}[S_n \leq -x]&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;S_n = \sum_{r=1}^n X_r&amp;lt;/math&amp;gt;. Is the conclusion true without the assumtion of independence?&lt;br /&gt;
* [&#039;&#039;&#039;Dependence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be discrete random variables with joint mass function &amp;lt;math&amp;gt;f(x,y) = \frac{C}{(x+y-1)(x+y)(x+y+1)}&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;x,y \in \mathbb{N}_+&amp;lt;/math&amp;gt; (in other words, &amp;lt;math&amp;gt;x,y = 1,2,3,\cdots&amp;lt;/math&amp;gt;). Find (1) the value of &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;, (2) marginal mass function of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and (3) &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt;.&lt;br /&gt;
* [&#039;&#039;&#039;Expectation&#039;&#039;&#039;] It is required to place in order &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; books &amp;lt;math&amp;gt;B_1, B_2, \cdots, B_n&amp;lt;/math&amp;gt; on a library shelf in such way that readers searching from left to right waste as little time as possible on average. Assuming that a random reader requires book &amp;lt;math&amp;gt;B_i&amp;lt;/math&amp;gt; with probability &amp;lt;math&amp;gt;p_i&amp;lt;/math&amp;gt;, find the ordering of the books which minimizes the number of titles examined by a random reader before discovery of the required book.&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Entropy of discrete random variable&amp;lt;/strong&amp;gt;] Let [math]X[/math] be a discrete random variable with range of values [math][N] = \{1,2,\ldots,N\}[/math] and probability mass function [math]p[/math]. Define [math]H(X) = -\sum_{n \ge 1} p(n) \log p(n)[/math] with convention [math]0\log 0 = 0[/math]. Prove that [math]H(X) \le \log N[/math] using Jensen&#039;s inequality.&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Law of total expectation&amp;lt;/strong&amp;gt;] Let [math]X \sim \mathrm{Geom}(p)[/math] for some parameter [math]p \in (0,1)[/math]. Calculate [math]\mathbf{E}[X][/math] using the law of total expectation.&lt;br /&gt;
&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Random number of random variables&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\{X_n\}_{n \ge 1}&amp;lt;/math&amp;gt; be identically distributed random variable and &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; be a random variable taking values in the non-negative integers and independent of the &amp;lt;math&amp;gt;X_n&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;n \ge 1&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{E}\left[\sum_{i=1}^N X_i\right] = \mathbf{E}[N] \mathbf{E}[X_1]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Discrete random variable) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (I)&amp;lt;/strong&amp;gt;] Every package of some intrinsically dull commodity includes a small and exciting plastic object. There are [math]c[/math] different types of object, and each package is equally likely to contain any given type. You buy one package each day.&amp;lt;/p&amp;gt;&lt;br /&gt;
    &amp;lt;ol&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse between the acquisitions of the [math]j[/math]-th new type of object and the [math](j + 1)[/math]-th new type.&amp;lt;/li&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse before you have a full set of objects.&amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (II)&amp;lt;/strong&amp;gt;] Prove that geometric distribution is the only discrete memoryless distribution with range values &amp;lt;math&amp;gt;\mathbb{N}_+&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;n_1,n_2 \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;0 \le p \le 1&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Bin}(n_1,p),Y \sim \mathrm{Bin}(n_2,p)&amp;lt;/math&amp;gt; be independent random variables. Prove that &amp;lt;math&amp;gt;X+Y \sim \mathrm{Bin}(n_1+n_2,p)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Negative binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; follows the negative binomial distribution with parameter &amp;lt;math&amp;gt;r \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt;. Calculate &amp;lt;math&amp;gt;\mathbf{Var}[X] = \mathbf{E}[X^2] - \left(\mathbf{E}[X]\right)^2&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Hypergeometric distribution&amp;lt;/strong&amp;gt;] An urn contains [math]N[/math] balls, [math]b[/math] of which are blue and [math]r = N -b[/math] of which are red. A random sample of [math]n[/math] balls is drawn &amp;lt;strong&amp;gt;without replacement&amp;lt;/strong&amp;gt; (无放回) from the urn. Let [math]B[/math] the number of blue balls in this sample. Show that if [math]N, b[/math], and [math]r[/math] approach [math]+\infty[/math] in such a way that [math]b/N \rightarrow  p[/math] and [math]r/N \rightarrow 1 - p[/math], then [math]\mathbf{Pr}(B = k) \rightarrow {n\choose k}p^k(1-p)^{n-k}[/math] for [math]0\leq k \leq n[/math].&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Poisson distribution&amp;lt;/strong&amp;gt;] In your pocket is a random number &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; of coins, where &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; has the Poisson distribution with parameter &amp;lt;math&amp;gt;\lambda&amp;lt;/math&amp;gt;. You toss each coin once, with heads showing with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; each time. Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the (random) number of heads outcomes and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the (also random) number of tails.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the joint mass function of &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find PDF of the marginal distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; in &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;. Are &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; independent?&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Conditional distribution (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be independent &amp;lt;math&amp;gt;\text{Bin}(n, p)&amp;lt;/math&amp;gt; random variables, and let &amp;lt;math&amp;gt;Z = X + Y&amp;lt;/math&amp;gt;. Show that the conditional distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; given &amp;lt;math&amp;gt;Z = N&amp;lt;/math&amp;gt; is the hypergeometric distribution.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Conditional distribution (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;\lambda,\mu &amp;gt; 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;n \in \mathbb{N}&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Pois}(\lambda), Y \sim \mathrm{Pois}(\mu)&amp;lt;/math&amp;gt; be independent random variables. Find out the conditional distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, given &amp;lt;math&amp;gt;X+Y = n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Linearity of Expectation, 12 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Streak&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Suppose we flip a fair coin &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; times independently to obtain a sequence of flips &amp;lt;math&amp;gt;X_1, X_2, \ldots , X_n&amp;lt;/math&amp;gt;. A streak of flips is a consecutive subsequence of flips that are all the same. For example, if &amp;lt;math&amp;gt;X_3&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;X_4&amp;lt;/math&amp;gt;, and &amp;lt;math&amp;gt;X_5&amp;lt;/math&amp;gt; are all heads, there is a streak of length &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; starting at the third lip. (If &amp;lt;math&amp;gt;X_6&amp;lt;/math&amp;gt; is also heads, then there is also a streak of length &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; starting at the third lip.) Find the expected number of streaks of length &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; for some integer &amp;lt;math&amp;gt;k \ge 1&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Number of cycles&amp;lt;/strong&amp;gt;] &lt;br /&gt;
At a banquet, there are &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people who shake hands according to the following process: In each round, two idle hands are randomly selected and shaken (&amp;lt;strong&amp;gt;these two hands are no longer idle&amp;lt;/strong&amp;gt;). After &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds, there will be no idle hands left, and the &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people will form several cycles. For example, when &amp;lt;math&amp;gt;n=3&amp;lt;/math&amp;gt;, the following situation may occur: the left and right hands of the first person are held together, the left hand of the second person and the right hand of the third person are held together, and the right hand of the second person and the left hand of the third person are held together. In this case, three people form two cycles. How many cycles are expected to be formed after &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds?&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Number of operations&amp;lt;/strong&amp;gt;]&lt;br /&gt;
We have a directed graph &amp;lt;math&amp;gt;G=(V,E)&amp;lt;/math&amp;gt; without self-loops and multi-edges.&lt;br /&gt;
&lt;br /&gt;
Until the graph becomes empty, repeat the following operation:&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
Choose one (unerased) vertex uniformly at random (independently from the previous choices). Then, erase that vertex and all vertices that are reachable from the chosen vertex by traversing some edges. Erasing a vertex will also erase the edges incident to it.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;R_v&amp;lt;/math&amp;gt; denotes the number of vertices from which vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt; is reachable. Prove that the expected value of the number of operations equals to &amp;lt;math&amp;gt;\sum_{v \in V} \frac{1}{R_v}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Expected Mex&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X_1,X_2,\ldots,X_{100} \sim \mathrm{Geo}(1/2)&amp;lt;/math&amp;gt; be independent random variables. Compute &amp;lt;math&amp;gt;\mathbf{E}[\mathrm{mex}(X_1,X_2,\ldots,X_{100})]&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\mathrm{mex}(a_1,a_2,\ldots,a_n)&amp;lt;/math&amp;gt; is the smallest positive integer that does not appear in &amp;lt;math&amp;gt;a_1,a_2,\ldots,a_n&amp;lt;/math&amp;gt;. Your answer is considered correct if the absolute error does not exceed &amp;lt;math&amp;gt;10^{-6}&amp;lt;/math&amp;gt;. (Hint: Although mathematics will help you arrive at elegant and efficient methods, the use of a computer and programming skills will be required.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 4 (Probability meets graph theory) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random social networks&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G = (V, E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph without isolating vertex. &lt;br /&gt;
Let &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; be the degree of vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Let &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be a uniformly chosen vertex, and &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; a uniformly chosen neighbor of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Show that &amp;lt;math&amp;gt;\mathbf{E}[d_Z] \geq \mathbf{E}[d_Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Interpret this inequality in the context of social networks, in which the vertices represent people, and the edges represent friendship.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Turán&#039;s Theorem&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G=(V,E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph, and write &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; for the degree of the vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Use probablistic method to prove that &amp;lt;math&amp;gt;\alpha(G) \ge \sum_{v \in V} \frac{1}{d_v + 1}&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\alpha(G)&amp;lt;/math&amp;gt; is the size of a maximum independent set. (Hint: Consider the following random procedure for generating an independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; from a graph with vertex set &amp;lt;math&amp;gt;V&amp;lt;/math&amp;gt;: First, generate a random permutation of the vertices, denoted as &amp;lt;math&amp;gt;v_1,v_2,\ldots,v_n&amp;lt;/math&amp;gt;. Then, construct the independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; as follows: For each vertex &amp;lt;math&amp;gt;v_i \in V&amp;lt;/math&amp;gt;, add &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; if and only if none of its predecessors in the permutation, i.e., &amp;lt;math&amp;gt;v_1,\ldots,v_{i-1}&amp;lt;/math&amp;gt;, are neighbors of &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Tournament&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Prove that there is a [https://en.wikipedia.org/wiki/Tournament_(graph_theory) tournament] &amp;lt;math&amp;gt;T&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; players and at least &amp;lt;math&amp;gt;n!/2^{n-1}&amp;lt;/math&amp;gt;Hamiltonian paths.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 5 (1D random walk, 8 points) ==&lt;br /&gt;
Let &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt; be a constant, and &amp;lt;math&amp;gt;\{X_n\}_{n \ge 1}&amp;lt;/math&amp;gt; be independent Bernoulli trials with successful probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;.&lt;br /&gt;
Define &amp;lt;math&amp;gt;S_n = 2\sum_{i=1}^n X_i - n&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;S_0 = 0&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Range of random walk&amp;lt;/strong&amp;gt;] The range &amp;lt;math&amp;gt;R_n&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;S_0, S_1, \ldots, S_n&amp;lt;/math&amp;gt; is defined as the number of distinct values taken by the sequence. Show that &amp;lt;math&amp;gt;\mathbf{Pr}\left(R_n = R_{n-1}+1\right) = \mathbf{Pr}\left(\forall 1 \le i \le n, S_i \neq 0\right)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, and deduce that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n]\to \mathbf{Pr}(\forall i \ge 1, S_i \neq 0)&amp;lt;/math&amp;gt;. Hence show that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n] \to |2p-1|&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Symmetric 1D random walk (III)&amp;lt;/strong&amp;gt;] Suppose &amp;lt;math&amp;gt;p = \frac{1}{2}&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{E}[|S_n|] = \Theta(\sqrt{n})&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12960</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 2</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12960"/>
		<updated>2025-03-13T15:38:21Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: /* Problem 1 (Warm-up problems, 12 points) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 2==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Warm-up problems, 12 points) ==&lt;br /&gt;
* [&#039;&#039;&#039;Cumulative distribution function (CDF) (I)&#039;&#039;&#039;] Express the distribution functions of [math]X^+ = \max\{0,X\}[/math], [math]X^- = -\min\{0,X\}[/math], [math]|X|=X^+ + X^-[/math], [math]-X[/math], in terms of the distribution function [math]F[/math] of the random variable [math]X[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Cumulative distribution function (CDF) (II)&#039;&#039;&#039;] Let [math]X[/math] be a random variable with distribution function [math]\max(0,\min(1,x))[/math]. Let [math]F[/math] be a distribution function which is continuous and strictly increasing. Show that [math]Y=F^{-1}(X)[/math] be a random variable with distribution function [math]F[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Probability density function (PDF)&#039;&#039;&#039;] We toss &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; coins, and each one shows heads with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;, independently of each of the others. Each coin which shows head is tossed again. (If the coin shows tail, it won&#039;t be tossed again.) Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of heads resulting from the &amp;lt;strong&amp;gt;second&amp;lt;/strong&amp;gt; round of tosses, and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the number of heads resulting from &amp;lt;strong&amp;gt;all&amp;lt;/strong&amp;gt; tosses, which includes the first and (possible) second round of each toss.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the PDF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{E}[Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Let &amp;lt;math&amp;gt;p_X&amp;lt;/math&amp;gt; be the PMF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, show that [math]p_X(k-1)p_X(k+1)\leq [p_X(k)]^2[/math] for &amp;lt;math&amp;gt;1\leq k \leq n-1&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
* [&#039;&#039;&#039;Independence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;1\leq r\leq n&amp;lt;/math&amp;gt; be independent random variables which are symmetric about &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;; that is, &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-X_r&amp;lt;/math&amp;gt; have the same distributions. Show that, for all &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}[S_n \geq x] = \mathbf{Pr}[S_n \leq -x]&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;S_n = \sum_{r=1}^n X_r&amp;lt;/math&amp;gt;. Is the conclusion true without the assumtion of independence?&lt;br /&gt;
* [&#039;&#039;&#039;Dependence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be discrete random variables with joint mass function &amp;lt;math&amp;gt;f(x,y) = \frac{C}{(x+y-1)(x+y)(x+y+1)}&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;x,y \in \mathbb{N}_+&amp;lt;/math&amp;gt; (in other words, &amp;lt;math&amp;gt;x,y = 1,2,3,\cdots&amp;lt;/math&amp;gt;). Find (1) the value of &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;, (2) marginal mass function of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and (3) &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt;.&lt;br /&gt;
* [&#039;&#039;&#039;Expectation&#039;&#039;&#039;] It is required to place in order &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; books &amp;lt;math&amp;gt;B_1, B_2, \cdots, B_n&amp;lt;/math&amp;gt; on a library shelf in such way that readers searching from left to right waste as little time as possible on average. Assuming that a random reader requires book &amp;lt;math&amp;gt;B_i&amp;lt;/math&amp;gt; with probability &amp;lt;math&amp;gt;p_i&amp;lt;/math&amp;gt;, find the ordering of the books which minimizes the number of titles examined by a random reader before discovery of the required book.&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Entropy of discrete random variable&amp;lt;/strong&amp;gt;] Let [math]X[/math] be a discrete random variable with range of values [math][N] = \{1,2,\ldots,N\}[/math] and probability mass function [math]p[/math]. Define [math]H(X) = -\sum_{n \ge 1} p(n) \log p(n)[/math] with convention [math]0\log 0 = 0[/math]. Prove that [math]H(X) \le \log N[/math] using Jensen&#039;s inequality.&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Law of total expectation&amp;lt;/strong&amp;gt;] Let [math]X \sim \mathrm{Geom}(p)[/math] for some parameter [math]p \in (0,1)[/math]. Calculate [math]\mathbf{E}[X][/math] using the law of total expectation.&lt;br /&gt;
&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Random number of random variables&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\{X_n\}_{n \ge 1}&amp;lt;/math&amp;gt; be identically distributed random variable and &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; be a random variable taking values in the non-negative integers and independent of the &amp;lt;math&amp;gt;X_n&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;n \ge 1&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{E}\left[\sum_{i=1}^N X_i\right] = \mathbf{E}[N] \mathbf{E}[X_1]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Distribution of random variable, 8 points) ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Cumulative distribution function (CDF)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a random variable with cumulative distribution function &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Show that &amp;lt;math&amp;gt;Y = aX+b&amp;lt;/math&amp;gt; is a random variable where &amp;lt;math&amp;gt;a&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;b&amp;lt;/math&amp;gt; are real constants, and express the CDF of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; by &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;. (Hint: Try expressing the event [math]Y=aX+b\le y[/math] by countably many set operations on the events defined on [math]X[/math].)&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Let &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; be the CDF of random variable &amp;lt;math&amp;gt;Z:\Omega\rightarrow \mathbb{R}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;0\leq \lambda \leq 1&amp;lt;/math&amp;gt;, show that &lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;math&amp;gt;\lambda F + (1-\lambda)G&amp;lt;/math&amp;gt; is a CDF function. &amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;The product &amp;lt;math&amp;gt;FG&amp;lt;/math&amp;gt; is a CDF function, and if &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; are independent, then &amp;lt;math&amp;gt;FG&amp;lt;/math&amp;gt; is the CDF of &amp;lt;math&amp;gt;\max\{X,Z\}&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Probability density function (PDF)&amp;lt;/strong&amp;gt;] We toss &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; coins, and each one shows heads with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;, independently of each of the others. Each coin which shows head is tossed again. (If the coin shows tail, it won&#039;t be tossed again.) Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of heads resulting from the &amp;lt;strong&amp;gt;second&amp;lt;/strong&amp;gt; round of tosses, and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the number of heads resulting from &amp;lt;strong&amp;gt;all&amp;lt;/strong&amp;gt; tosses, which includes the first and (possible) second round of each toss.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the PDF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{E}[Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Let &amp;lt;math&amp;gt;p_X&amp;lt;/math&amp;gt; be the PMF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, show that [math]p_X(k-1)p_X(k+1)\leq [p_X(k)]^2[/math] for &amp;lt;math&amp;gt;1\leq k \leq n-1&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Discrete random variable) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (I)&amp;lt;/strong&amp;gt;] Every package of some intrinsically dull commodity includes a small and exciting plastic object. There are [math]c[/math] different types of object, and each package is equally likely to contain any given type. You buy one package each day.&amp;lt;/p&amp;gt;&lt;br /&gt;
    &amp;lt;ol&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse between the acquisitions of the [math]j[/math]-th new type of object and the [math](j + 1)[/math]-th new type.&amp;lt;/li&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse before you have a full set of objects.&amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (II)&amp;lt;/strong&amp;gt;] Prove that geometric distribution is the only discrete memoryless distribution with range values &amp;lt;math&amp;gt;\mathbb{N}_+&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;n_1,n_2 \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;0 \le p \le 1&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Bin}(n_1,p),Y \sim \mathrm{Bin}(n_2,p)&amp;lt;/math&amp;gt; be independent random variables. Prove that &amp;lt;math&amp;gt;X+Y \sim \mathrm{Bin}(n_1+n_2,p)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Negative binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; follows the negative binomial distribution with parameter &amp;lt;math&amp;gt;r \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt;. Calculate &amp;lt;math&amp;gt;\mathbf{Var}[X] = \mathbf{E}[X^2] - \left(\mathbf{E}[X]\right)^2&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Hypergeometric distribution&amp;lt;/strong&amp;gt;] An urn contains [math]N[/math] balls, [math]b[/math] of which are blue and [math]r = N -b[/math] of which are red. A random sample of [math]n[/math] balls is drawn &amp;lt;strong&amp;gt;without replacement&amp;lt;/strong&amp;gt; (无放回) from the urn. Let [math]B[/math] the number of blue balls in this sample. Show that if [math]N, b[/math], and [math]r[/math] approach [math]+\infty[/math] in such a way that [math]b/N \rightarrow  p[/math] and [math]r/N \rightarrow 1 - p[/math], then [math]\mathbf{Pr}(B = k) \rightarrow {n\choose k}p^k(1-p)^{n-k}[/math] for [math]0\leq k \leq n[/math].&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Poisson distribution&amp;lt;/strong&amp;gt;] In your pocket is a random number &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; of coins, where &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; has the Poisson distribution with parameter &amp;lt;math&amp;gt;\lambda&amp;lt;/math&amp;gt;. You toss each coin once, with heads showing with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; each time. Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the (random) number of heads outcomes and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the (also random) number of tails.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the joint mass function of &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find PDF of the marginal distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; in &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;. Are &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; independent?&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Conditional distribution (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be independent &amp;lt;math&amp;gt;\text{Bin}(n, p)&amp;lt;/math&amp;gt; random variables, and let &amp;lt;math&amp;gt;Z = X + Y&amp;lt;/math&amp;gt;. Show that the conditional distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; given &amp;lt;math&amp;gt;Z = N&amp;lt;/math&amp;gt; is the hypergeometric distribution.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Conditional distribution (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;\lambda,\mu &amp;gt; 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;n \in \mathbb{N}&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Pois}(\lambda), Y \sim \mathrm{Pois}(\mu)&amp;lt;/math&amp;gt; be independent random variables. Find out the conditional distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, given &amp;lt;math&amp;gt;X+Y = n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 4 (Linearity of Expectation, 12 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Streak&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Suppose we flip a fair coin &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; times independently to obtain a sequence of flips &amp;lt;math&amp;gt;X_1, X_2, \ldots , X_n&amp;lt;/math&amp;gt;. A streak of flips is a consecutive subsequence of flips that are all the same. For example, if &amp;lt;math&amp;gt;X_3&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;X_4&amp;lt;/math&amp;gt;, and &amp;lt;math&amp;gt;X_5&amp;lt;/math&amp;gt; are all heads, there is a streak of length &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; starting at the third lip. (If &amp;lt;math&amp;gt;X_6&amp;lt;/math&amp;gt; is also heads, then there is also a streak of length &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; starting at the third lip.) Find the expected number of streaks of length &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; for some integer &amp;lt;math&amp;gt;k \ge 1&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Number of cycles&amp;lt;/strong&amp;gt;] &lt;br /&gt;
At a banquet, there are &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people who shake hands according to the following process: In each round, two idle hands are randomly selected and shaken (&amp;lt;strong&amp;gt;these two hands are no longer idle&amp;lt;/strong&amp;gt;). After &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds, there will be no idle hands left, and the &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people will form several cycles. For example, when &amp;lt;math&amp;gt;n=3&amp;lt;/math&amp;gt;, the following situation may occur: the left and right hands of the first person are held together, the left hand of the second person and the right hand of the third person are held together, and the right hand of the second person and the left hand of the third person are held together. In this case, three people form two cycles. How many cycles are expected to be formed after &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds?&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Number of operations&amp;lt;/strong&amp;gt;]&lt;br /&gt;
We have a directed graph &amp;lt;math&amp;gt;G=(V,E)&amp;lt;/math&amp;gt; without self-loops and multi-edges.&lt;br /&gt;
&lt;br /&gt;
Until the graph becomes empty, repeat the following operation:&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
Choose one (unerased) vertex uniformly at random (independently from the previous choices). Then, erase that vertex and all vertices that are reachable from the chosen vertex by traversing some edges. Erasing a vertex will also erase the edges incident to it.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;R_v&amp;lt;/math&amp;gt; denotes the number of vertices from which vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt; is reachable. Prove that the expected value of the number of operations equals to &amp;lt;math&amp;gt;\sum_{v \in V} \frac{1}{R_v}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Expected Mex&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X_1,X_2,\ldots,X_{100} \sim \mathrm{Geo}(1/2)&amp;lt;/math&amp;gt; be independent random variables. Compute &amp;lt;math&amp;gt;\mathbf{E}[\mathrm{mex}(X_1,X_2,\ldots,X_{100})]&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\mathrm{mex}(a_1,a_2,\ldots,a_n)&amp;lt;/math&amp;gt; is the smallest positive integer that does not appear in &amp;lt;math&amp;gt;a_1,a_2,\ldots,a_n&amp;lt;/math&amp;gt;. Your answer is considered correct if the absolute error does not exceed &amp;lt;math&amp;gt;10^{-6}&amp;lt;/math&amp;gt;. (Hint: Although mathematics will help you arrive at elegant and efficient methods, the use of a computer and programming skills will be required.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 5 (Probability meets graph theory) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random social networks&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G = (V, E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph without isolating vertex. &lt;br /&gt;
Let &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; be the degree of vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Let &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be a uniformly chosen vertex, and &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; a uniformly chosen neighbor of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Show that &amp;lt;math&amp;gt;\mathbf{E}[d_Z] \geq \mathbf{E}[d_Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Interpret this inequality in the context of social networks, in which the vertices represent people, and the edges represent friendship.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Turán&#039;s Theorem&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G=(V,E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph, and write &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; for the degree of the vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Use probablistic method to prove that &amp;lt;math&amp;gt;\alpha(G) \ge \sum_{v \in V} \frac{1}{d_v + 1}&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\alpha(G)&amp;lt;/math&amp;gt; is the size of a maximum independent set. (Hint: Consider the following random procedure for generating an independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; from a graph with vertex set &amp;lt;math&amp;gt;V&amp;lt;/math&amp;gt;: First, generate a random permutation of the vertices, denoted as &amp;lt;math&amp;gt;v_1,v_2,\ldots,v_n&amp;lt;/math&amp;gt;. Then, construct the independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; as follows: For each vertex &amp;lt;math&amp;gt;v_i \in V&amp;lt;/math&amp;gt;, add &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; if and only if none of its predecessors in the permutation, i.e., &amp;lt;math&amp;gt;v_1,\ldots,v_{i-1}&amp;lt;/math&amp;gt;, are neighbors of &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Tournament&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Prove that there is a [https://en.wikipedia.org/wiki/Tournament_(graph_theory) tournament] &amp;lt;math&amp;gt;T&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; players and at least &amp;lt;math&amp;gt;n!/2^{n-1}&amp;lt;/math&amp;gt;Hamiltonian paths.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 6 (1D random walk, 8 points) ==&lt;br /&gt;
Let &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt; be a constant, and &amp;lt;math&amp;gt;\{X_n\}_{n \ge 1}&amp;lt;/math&amp;gt; be independent Bernoulli trials with successful probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;.&lt;br /&gt;
Define &amp;lt;math&amp;gt;S_n = 2\sum_{i=1}^n X_i - n&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;S_0 = 0&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Range of random walk&amp;lt;/strong&amp;gt;] The range &amp;lt;math&amp;gt;R_n&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;S_0, S_1, \ldots, S_n&amp;lt;/math&amp;gt; is defined as the number of distinct values taken by the sequence. Show that &amp;lt;math&amp;gt;\mathbf{Pr}\left(R_n = R_{n-1}+1\right) = \mathbf{Pr}\left(\forall 1 \le i \le n, S_i \neq 0\right)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, and deduce that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n]\to \mathbf{Pr}(\forall i \ge 1, S_i \neq 0)&amp;lt;/math&amp;gt;. Hence show that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n] \to |2p-1|&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Symmetric 1D random walk (III)&amp;lt;/strong&amp;gt;] Suppose &amp;lt;math&amp;gt;p = \frac{1}{2}&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{E}[|S_n|] = \Theta(\sqrt{n})&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12959</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 2</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12959"/>
		<updated>2025-03-13T15:37:55Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: /* Problem 1 (Warm-up problems, 12 points) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 2==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Warm-up problems, 12 points) ==&lt;br /&gt;
* [&#039;&#039;&#039;Function of random variable (I)&#039;&#039;&#039;] Express the distribution functions of [math]X^+ = \max\{0,X\}[/math], [math]X^- = -\min\{0,X\}[/math], [math]|X|=X^+ + X^-[/math], [math]-X[/math], in terms of the distribution function [math]F[/math] of the random variable [math]X[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Function of random variable (II)&#039;&#039;&#039;] Let [math]X[/math] be a random variable with distribution function [math]\max(0,\min(1,x))[/math]. Let [math]F[/math] be a distribution function which is continuous and strictly increasing. Show that [math]Y=F^{-1}(X)[/math] be a random variable with distribution function [math]F[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Probability density function (PDF)&#039;&#039;&#039;] We toss &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; coins, and each one shows heads with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;, independently of each of the others. Each coin which shows head is tossed again. (If the coin shows tail, it won&#039;t be tossed again.) Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of heads resulting from the &amp;lt;strong&amp;gt;second&amp;lt;/strong&amp;gt; round of tosses, and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the number of heads resulting from &amp;lt;strong&amp;gt;all&amp;lt;/strong&amp;gt; tosses, which includes the first and (possible) second round of each toss.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the PDF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{E}[Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Let &amp;lt;math&amp;gt;p_X&amp;lt;/math&amp;gt; be the PMF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, show that [math]p_X(k-1)p_X(k+1)\leq [p_X(k)]^2[/math] for &amp;lt;math&amp;gt;1\leq k \leq n-1&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
* [&#039;&#039;&#039;Independence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;1\leq r\leq n&amp;lt;/math&amp;gt; be independent random variables which are symmetric about &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;; that is, &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-X_r&amp;lt;/math&amp;gt; have the same distributions. Show that, for all &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}[S_n \geq x] = \mathbf{Pr}[S_n \leq -x]&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;S_n = \sum_{r=1}^n X_r&amp;lt;/math&amp;gt;. Is the conclusion true without the assumtion of independence?&lt;br /&gt;
* [&#039;&#039;&#039;Dependence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be discrete random variables with joint mass function &amp;lt;math&amp;gt;f(x,y) = \frac{C}{(x+y-1)(x+y)(x+y+1)}&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;x,y \in \mathbb{N}_+&amp;lt;/math&amp;gt; (in other words, &amp;lt;math&amp;gt;x,y = 1,2,3,\cdots&amp;lt;/math&amp;gt;). Find (1) the value of &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;, (2) marginal mass function of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and (3) &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt;.&lt;br /&gt;
* [&#039;&#039;&#039;Expectation&#039;&#039;&#039;] It is required to place in order &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; books &amp;lt;math&amp;gt;B_1, B_2, \cdots, B_n&amp;lt;/math&amp;gt; on a library shelf in such way that readers searching from left to right waste as little time as possible on average. Assuming that a random reader requires book &amp;lt;math&amp;gt;B_i&amp;lt;/math&amp;gt; with probability &amp;lt;math&amp;gt;p_i&amp;lt;/math&amp;gt;, find the ordering of the books which minimizes the number of titles examined by a random reader before discovery of the required book.&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Entropy of discrete random variable&amp;lt;/strong&amp;gt;] Let [math]X[/math] be a discrete random variable with range of values [math][N] = \{1,2,\ldots,N\}[/math] and probability mass function [math]p[/math]. Define [math]H(X) = -\sum_{n \ge 1} p(n) \log p(n)[/math] with convention [math]0\log 0 = 0[/math]. Prove that [math]H(X) \le \log N[/math] using Jensen&#039;s inequality.&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Law of total expectation&amp;lt;/strong&amp;gt;] Let [math]X \sim \mathrm{Geom}(p)[/math] for some parameter [math]p \in (0,1)[/math]. Calculate [math]\mathbf{E}[X][/math] using the law of total expectation.&lt;br /&gt;
&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Random number of random variables&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\{X_n\}_{n \ge 1}&amp;lt;/math&amp;gt; be identically distributed random variable and &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; be a random variable taking values in the non-negative integers and independent of the &amp;lt;math&amp;gt;X_n&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;n \ge 1&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{E}\left[\sum_{i=1}^N X_i\right] = \mathbf{E}[N] \mathbf{E}[X_1]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Distribution of random variable, 8 points) ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Cumulative distribution function (CDF)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a random variable with cumulative distribution function &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Show that &amp;lt;math&amp;gt;Y = aX+b&amp;lt;/math&amp;gt; is a random variable where &amp;lt;math&amp;gt;a&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;b&amp;lt;/math&amp;gt; are real constants, and express the CDF of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; by &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;. (Hint: Try expressing the event [math]Y=aX+b\le y[/math] by countably many set operations on the events defined on [math]X[/math].)&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Let &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; be the CDF of random variable &amp;lt;math&amp;gt;Z:\Omega\rightarrow \mathbb{R}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;0\leq \lambda \leq 1&amp;lt;/math&amp;gt;, show that &lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;math&amp;gt;\lambda F + (1-\lambda)G&amp;lt;/math&amp;gt; is a CDF function. &amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;The product &amp;lt;math&amp;gt;FG&amp;lt;/math&amp;gt; is a CDF function, and if &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; are independent, then &amp;lt;math&amp;gt;FG&amp;lt;/math&amp;gt; is the CDF of &amp;lt;math&amp;gt;\max\{X,Z\}&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Probability density function (PDF)&amp;lt;/strong&amp;gt;] We toss &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; coins, and each one shows heads with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;, independently of each of the others. Each coin which shows head is tossed again. (If the coin shows tail, it won&#039;t be tossed again.) Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of heads resulting from the &amp;lt;strong&amp;gt;second&amp;lt;/strong&amp;gt; round of tosses, and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the number of heads resulting from &amp;lt;strong&amp;gt;all&amp;lt;/strong&amp;gt; tosses, which includes the first and (possible) second round of each toss.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the PDF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{E}[Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Let &amp;lt;math&amp;gt;p_X&amp;lt;/math&amp;gt; be the PMF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, show that [math]p_X(k-1)p_X(k+1)\leq [p_X(k)]^2[/math] for &amp;lt;math&amp;gt;1\leq k \leq n-1&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Discrete random variable) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (I)&amp;lt;/strong&amp;gt;] Every package of some intrinsically dull commodity includes a small and exciting plastic object. There are [math]c[/math] different types of object, and each package is equally likely to contain any given type. You buy one package each day.&amp;lt;/p&amp;gt;&lt;br /&gt;
    &amp;lt;ol&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse between the acquisitions of the [math]j[/math]-th new type of object and the [math](j + 1)[/math]-th new type.&amp;lt;/li&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse before you have a full set of objects.&amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (II)&amp;lt;/strong&amp;gt;] Prove that geometric distribution is the only discrete memoryless distribution with range values &amp;lt;math&amp;gt;\mathbb{N}_+&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;n_1,n_2 \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;0 \le p \le 1&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Bin}(n_1,p),Y \sim \mathrm{Bin}(n_2,p)&amp;lt;/math&amp;gt; be independent random variables. Prove that &amp;lt;math&amp;gt;X+Y \sim \mathrm{Bin}(n_1+n_2,p)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Negative binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; follows the negative binomial distribution with parameter &amp;lt;math&amp;gt;r \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt;. Calculate &amp;lt;math&amp;gt;\mathbf{Var}[X] = \mathbf{E}[X^2] - \left(\mathbf{E}[X]\right)^2&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Hypergeometric distribution&amp;lt;/strong&amp;gt;] An urn contains [math]N[/math] balls, [math]b[/math] of which are blue and [math]r = N -b[/math] of which are red. A random sample of [math]n[/math] balls is drawn &amp;lt;strong&amp;gt;without replacement&amp;lt;/strong&amp;gt; (无放回) from the urn. Let [math]B[/math] the number of blue balls in this sample. Show that if [math]N, b[/math], and [math]r[/math] approach [math]+\infty[/math] in such a way that [math]b/N \rightarrow  p[/math] and [math]r/N \rightarrow 1 - p[/math], then [math]\mathbf{Pr}(B = k) \rightarrow {n\choose k}p^k(1-p)^{n-k}[/math] for [math]0\leq k \leq n[/math].&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Poisson distribution&amp;lt;/strong&amp;gt;] In your pocket is a random number &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; of coins, where &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; has the Poisson distribution with parameter &amp;lt;math&amp;gt;\lambda&amp;lt;/math&amp;gt;. You toss each coin once, with heads showing with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; each time. Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the (random) number of heads outcomes and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the (also random) number of tails.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the joint mass function of &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find PDF of the marginal distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; in &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;. Are &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; independent?&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Conditional distribution (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be independent &amp;lt;math&amp;gt;\text{Bin}(n, p)&amp;lt;/math&amp;gt; random variables, and let &amp;lt;math&amp;gt;Z = X + Y&amp;lt;/math&amp;gt;. Show that the conditional distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; given &amp;lt;math&amp;gt;Z = N&amp;lt;/math&amp;gt; is the hypergeometric distribution.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Conditional distribution (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;\lambda,\mu &amp;gt; 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;n \in \mathbb{N}&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Pois}(\lambda), Y \sim \mathrm{Pois}(\mu)&amp;lt;/math&amp;gt; be independent random variables. Find out the conditional distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, given &amp;lt;math&amp;gt;X+Y = n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 4 (Linearity of Expectation, 12 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Streak&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Suppose we flip a fair coin &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; times independently to obtain a sequence of flips &amp;lt;math&amp;gt;X_1, X_2, \ldots , X_n&amp;lt;/math&amp;gt;. A streak of flips is a consecutive subsequence of flips that are all the same. For example, if &amp;lt;math&amp;gt;X_3&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;X_4&amp;lt;/math&amp;gt;, and &amp;lt;math&amp;gt;X_5&amp;lt;/math&amp;gt; are all heads, there is a streak of length &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; starting at the third lip. (If &amp;lt;math&amp;gt;X_6&amp;lt;/math&amp;gt; is also heads, then there is also a streak of length &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; starting at the third lip.) Find the expected number of streaks of length &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; for some integer &amp;lt;math&amp;gt;k \ge 1&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Number of cycles&amp;lt;/strong&amp;gt;] &lt;br /&gt;
At a banquet, there are &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people who shake hands according to the following process: In each round, two idle hands are randomly selected and shaken (&amp;lt;strong&amp;gt;these two hands are no longer idle&amp;lt;/strong&amp;gt;). After &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds, there will be no idle hands left, and the &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people will form several cycles. For example, when &amp;lt;math&amp;gt;n=3&amp;lt;/math&amp;gt;, the following situation may occur: the left and right hands of the first person are held together, the left hand of the second person and the right hand of the third person are held together, and the right hand of the second person and the left hand of the third person are held together. In this case, three people form two cycles. How many cycles are expected to be formed after &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds?&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Number of operations&amp;lt;/strong&amp;gt;]&lt;br /&gt;
We have a directed graph &amp;lt;math&amp;gt;G=(V,E)&amp;lt;/math&amp;gt; without self-loops and multi-edges.&lt;br /&gt;
&lt;br /&gt;
Until the graph becomes empty, repeat the following operation:&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
Choose one (unerased) vertex uniformly at random (independently from the previous choices). Then, erase that vertex and all vertices that are reachable from the chosen vertex by traversing some edges. Erasing a vertex will also erase the edges incident to it.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;R_v&amp;lt;/math&amp;gt; denotes the number of vertices from which vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt; is reachable. Prove that the expected value of the number of operations equals to &amp;lt;math&amp;gt;\sum_{v \in V} \frac{1}{R_v}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Expected Mex&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X_1,X_2,\ldots,X_{100} \sim \mathrm{Geo}(1/2)&amp;lt;/math&amp;gt; be independent random variables. Compute &amp;lt;math&amp;gt;\mathbf{E}[\mathrm{mex}(X_1,X_2,\ldots,X_{100})]&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\mathrm{mex}(a_1,a_2,\ldots,a_n)&amp;lt;/math&amp;gt; is the smallest positive integer that does not appear in &amp;lt;math&amp;gt;a_1,a_2,\ldots,a_n&amp;lt;/math&amp;gt;. Your answer is considered correct if the absolute error does not exceed &amp;lt;math&amp;gt;10^{-6}&amp;lt;/math&amp;gt;. (Hint: Although mathematics will help you arrive at elegant and efficient methods, the use of a computer and programming skills will be required.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 5 (Probability meets graph theory) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random social networks&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G = (V, E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph without isolating vertex. &lt;br /&gt;
Let &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; be the degree of vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Let &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be a uniformly chosen vertex, and &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; a uniformly chosen neighbor of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Show that &amp;lt;math&amp;gt;\mathbf{E}[d_Z] \geq \mathbf{E}[d_Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Interpret this inequality in the context of social networks, in which the vertices represent people, and the edges represent friendship.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Turán&#039;s Theorem&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G=(V,E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph, and write &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; for the degree of the vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Use probablistic method to prove that &amp;lt;math&amp;gt;\alpha(G) \ge \sum_{v \in V} \frac{1}{d_v + 1}&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\alpha(G)&amp;lt;/math&amp;gt; is the size of a maximum independent set. (Hint: Consider the following random procedure for generating an independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; from a graph with vertex set &amp;lt;math&amp;gt;V&amp;lt;/math&amp;gt;: First, generate a random permutation of the vertices, denoted as &amp;lt;math&amp;gt;v_1,v_2,\ldots,v_n&amp;lt;/math&amp;gt;. Then, construct the independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; as follows: For each vertex &amp;lt;math&amp;gt;v_i \in V&amp;lt;/math&amp;gt;, add &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; if and only if none of its predecessors in the permutation, i.e., &amp;lt;math&amp;gt;v_1,\ldots,v_{i-1}&amp;lt;/math&amp;gt;, are neighbors of &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Tournament&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Prove that there is a [https://en.wikipedia.org/wiki/Tournament_(graph_theory) tournament] &amp;lt;math&amp;gt;T&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; players and at least &amp;lt;math&amp;gt;n!/2^{n-1}&amp;lt;/math&amp;gt;Hamiltonian paths.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 6 (1D random walk, 8 points) ==&lt;br /&gt;
Let &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt; be a constant, and &amp;lt;math&amp;gt;\{X_n\}_{n \ge 1}&amp;lt;/math&amp;gt; be independent Bernoulli trials with successful probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;.&lt;br /&gt;
Define &amp;lt;math&amp;gt;S_n = 2\sum_{i=1}^n X_i - n&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;S_0 = 0&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Range of random walk&amp;lt;/strong&amp;gt;] The range &amp;lt;math&amp;gt;R_n&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;S_0, S_1, \ldots, S_n&amp;lt;/math&amp;gt; is defined as the number of distinct values taken by the sequence. Show that &amp;lt;math&amp;gt;\mathbf{Pr}\left(R_n = R_{n-1}+1\right) = \mathbf{Pr}\left(\forall 1 \le i \le n, S_i \neq 0\right)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, and deduce that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n]\to \mathbf{Pr}(\forall i \ge 1, S_i \neq 0)&amp;lt;/math&amp;gt;. Hence show that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n] \to |2p-1|&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Symmetric 1D random walk (III)&amp;lt;/strong&amp;gt;] Suppose &amp;lt;math&amp;gt;p = \frac{1}{2}&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{E}[|S_n|] = \Theta(\sqrt{n})&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12958</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 2</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12958"/>
		<updated>2025-03-13T15:32:01Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: /* Problem 3 (Discrete random variable) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 2==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Warm-up problems, 12 points) ==&lt;br /&gt;
* [&#039;&#039;&#039;Function of random variable (I)&#039;&#039;&#039;] Express the distribution functions of [math]X^+ = \max\{0,X\}[/math], [math]X^- = -\min\{0,X\}[/math], [math]|X|=X^+ + X^-[/math], [math]-X[/math], in terms of the distribution function [math]F[/math] of the random variable [math]X[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Function of random variable (II)&#039;&#039;&#039;] Let [math]X[/math] be a random variable with distribution function [math]\max(0,\min(1,x))[/math]. Let [math]F[/math] be a distribution function which is continuous and strictly increasing. Show that [math]Y=F^{-1}(X)[/math] be a random variable with distribution function [math]F[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Independence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;1\leq r\leq n&amp;lt;/math&amp;gt; be independent random variables which are symmetric about &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;; that is, &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-X_r&amp;lt;/math&amp;gt; have the same distributions. Show that, for all &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}[S_n \geq x] = \mathbf{Pr}[S_n \leq -x]&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;S_n = \sum_{r=1}^n X_r&amp;lt;/math&amp;gt;. Is the conclusion true without the assumtion of independence?&lt;br /&gt;
* [&#039;&#039;&#039;Dependence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be discrete random variables with joint mass function &amp;lt;math&amp;gt;f(x,y) = \frac{C}{(x+y-1)(x+y)(x+y+1)}&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;x,y \in \mathbb{N}_+&amp;lt;/math&amp;gt; (in other words, &amp;lt;math&amp;gt;x,y = 1,2,3,\cdots&amp;lt;/math&amp;gt;). Find (1) the value of &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;, (2) marginal mass function of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and (3) &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt;.&lt;br /&gt;
* [&#039;&#039;&#039;Expectation&#039;&#039;&#039;] It is required to place in order &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; books &amp;lt;math&amp;gt;B_1, B_2, \cdots, B_n&amp;lt;/math&amp;gt; on a library shelf in such way that readers searching from left to right waste as little time as possible on average. Assuming that a random reader requires book &amp;lt;math&amp;gt;B_i&amp;lt;/math&amp;gt; with probability &amp;lt;math&amp;gt;p_i&amp;lt;/math&amp;gt;, find the ordering of the books which minimizes the number of titles examined by a random reader before discovery of the required book.&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Entropy of discrete random variable&amp;lt;/strong&amp;gt;] Let [math]X[/math] be a discrete random variable with range of values [math][N] = \{1,2,\ldots,N\}[/math] and probability mass function [math]p[/math]. Define [math]H(X) = -\sum_{n \ge 1} p(n) \log p(n)[/math] with convention [math]0\log 0 = 0[/math]. Prove that [math]H(X) \le \log N[/math] using Jensen&#039;s inequality.&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Law of total expectation&amp;lt;/strong&amp;gt;] Let [math]X \sim \mathrm{Geom}(p)[/math] for some parameter [math]p \in (0,1)[/math]. Calculate [math]\mathbf{E}[X][/math] using the law of total expectation.&lt;br /&gt;
&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Random number of random variables&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\{X_n\}_{n \ge 1}&amp;lt;/math&amp;gt; be identically distributed random variable and &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; be a random variable taking values in the non-negative integers and independent of the &amp;lt;math&amp;gt;X_n&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;n \ge 1&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{E}\left[\sum_{i=1}^N X_i\right] = \mathbf{E}[N] \mathbf{E}[X_1]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Distribution of random variable, 8 points) ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Cumulative distribution function (CDF)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a random variable with cumulative distribution function &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Show that &amp;lt;math&amp;gt;Y = aX+b&amp;lt;/math&amp;gt; is a random variable where &amp;lt;math&amp;gt;a&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;b&amp;lt;/math&amp;gt; are real constants, and express the CDF of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; by &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;. (Hint: Try expressing the event [math]Y=aX+b\le y[/math] by countably many set operations on the events defined on [math]X[/math].)&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Let &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; be the CDF of random variable &amp;lt;math&amp;gt;Z:\Omega\rightarrow \mathbb{R}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;0\leq \lambda \leq 1&amp;lt;/math&amp;gt;, show that &lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;math&amp;gt;\lambda F + (1-\lambda)G&amp;lt;/math&amp;gt; is a CDF function. &amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;The product &amp;lt;math&amp;gt;FG&amp;lt;/math&amp;gt; is a CDF function, and if &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; are independent, then &amp;lt;math&amp;gt;FG&amp;lt;/math&amp;gt; is the CDF of &amp;lt;math&amp;gt;\max\{X,Z\}&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Probability density function (PDF)&amp;lt;/strong&amp;gt;] We toss &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; coins, and each one shows heads with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;, independently of each of the others. Each coin which shows head is tossed again. (If the coin shows tail, it won&#039;t be tossed again.) Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of heads resulting from the &amp;lt;strong&amp;gt;second&amp;lt;/strong&amp;gt; round of tosses, and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the number of heads resulting from &amp;lt;strong&amp;gt;all&amp;lt;/strong&amp;gt; tosses, which includes the first and (possible) second round of each toss.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the PDF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{E}[Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Let &amp;lt;math&amp;gt;p_X&amp;lt;/math&amp;gt; be the PMF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, show that [math]p_X(k-1)p_X(k+1)\leq [p_X(k)]^2[/math] for &amp;lt;math&amp;gt;1\leq k \leq n-1&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Discrete random variable) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (I)&amp;lt;/strong&amp;gt;] Every package of some intrinsically dull commodity includes a small and exciting plastic object. There are [math]c[/math] different types of object, and each package is equally likely to contain any given type. You buy one package each day.&amp;lt;/p&amp;gt;&lt;br /&gt;
    &amp;lt;ol&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse between the acquisitions of the [math]j[/math]-th new type of object and the [math](j + 1)[/math]-th new type.&amp;lt;/li&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse before you have a full set of objects.&amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (II)&amp;lt;/strong&amp;gt;] Prove that geometric distribution is the only discrete memoryless distribution with range values &amp;lt;math&amp;gt;\mathbb{N}_+&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;n_1,n_2 \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;0 \le p \le 1&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Bin}(n_1,p),Y \sim \mathrm{Bin}(n_2,p)&amp;lt;/math&amp;gt; be independent random variables. Prove that &amp;lt;math&amp;gt;X+Y \sim \mathrm{Bin}(n_1+n_2,p)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Negative binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; follows the negative binomial distribution with parameter &amp;lt;math&amp;gt;r \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt;. Calculate &amp;lt;math&amp;gt;\mathbf{Var}[X] = \mathbf{E}[X^2] - \left(\mathbf{E}[X]\right)^2&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Hypergeometric distribution&amp;lt;/strong&amp;gt;] An urn contains [math]N[/math] balls, [math]b[/math] of which are blue and [math]r = N -b[/math] of which are red. A random sample of [math]n[/math] balls is drawn &amp;lt;strong&amp;gt;without replacement&amp;lt;/strong&amp;gt; (无放回) from the urn. Let [math]B[/math] the number of blue balls in this sample. Show that if [math]N, b[/math], and [math]r[/math] approach [math]+\infty[/math] in such a way that [math]b/N \rightarrow  p[/math] and [math]r/N \rightarrow 1 - p[/math], then [math]\mathbf{Pr}(B = k) \rightarrow {n\choose k}p^k(1-p)^{n-k}[/math] for [math]0\leq k \leq n[/math].&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Poisson distribution&amp;lt;/strong&amp;gt;] In your pocket is a random number &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; of coins, where &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; has the Poisson distribution with parameter &amp;lt;math&amp;gt;\lambda&amp;lt;/math&amp;gt;. You toss each coin once, with heads showing with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; each time. Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the (random) number of heads outcomes and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the (also random) number of tails.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the joint mass function of &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find PDF of the marginal distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; in &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;. Are &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; independent?&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Conditional distribution (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be independent &amp;lt;math&amp;gt;\text{Bin}(n, p)&amp;lt;/math&amp;gt; random variables, and let &amp;lt;math&amp;gt;Z = X + Y&amp;lt;/math&amp;gt;. Show that the conditional distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; given &amp;lt;math&amp;gt;Z = N&amp;lt;/math&amp;gt; is the hypergeometric distribution.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Conditional distribution (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;\lambda,\mu &amp;gt; 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;n \in \mathbb{N}&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Pois}(\lambda), Y \sim \mathrm{Pois}(\mu)&amp;lt;/math&amp;gt; be independent random variables. Find out the conditional distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, given &amp;lt;math&amp;gt;X+Y = n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 4 (Linearity of Expectation, 12 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Streak&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Suppose we flip a fair coin &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; times independently to obtain a sequence of flips &amp;lt;math&amp;gt;X_1, X_2, \ldots , X_n&amp;lt;/math&amp;gt;. A streak of flips is a consecutive subsequence of flips that are all the same. For example, if &amp;lt;math&amp;gt;X_3&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;X_4&amp;lt;/math&amp;gt;, and &amp;lt;math&amp;gt;X_5&amp;lt;/math&amp;gt; are all heads, there is a streak of length &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; starting at the third lip. (If &amp;lt;math&amp;gt;X_6&amp;lt;/math&amp;gt; is also heads, then there is also a streak of length &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; starting at the third lip.) Find the expected number of streaks of length &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; for some integer &amp;lt;math&amp;gt;k \ge 1&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Number of cycles&amp;lt;/strong&amp;gt;] &lt;br /&gt;
At a banquet, there are &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people who shake hands according to the following process: In each round, two idle hands are randomly selected and shaken (&amp;lt;strong&amp;gt;these two hands are no longer idle&amp;lt;/strong&amp;gt;). After &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds, there will be no idle hands left, and the &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people will form several cycles. For example, when &amp;lt;math&amp;gt;n=3&amp;lt;/math&amp;gt;, the following situation may occur: the left and right hands of the first person are held together, the left hand of the second person and the right hand of the third person are held together, and the right hand of the second person and the left hand of the third person are held together. In this case, three people form two cycles. How many cycles are expected to be formed after &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds?&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Number of operations&amp;lt;/strong&amp;gt;]&lt;br /&gt;
We have a directed graph &amp;lt;math&amp;gt;G=(V,E)&amp;lt;/math&amp;gt; without self-loops and multi-edges.&lt;br /&gt;
&lt;br /&gt;
Until the graph becomes empty, repeat the following operation:&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
Choose one (unerased) vertex uniformly at random (independently from the previous choices). Then, erase that vertex and all vertices that are reachable from the chosen vertex by traversing some edges. Erasing a vertex will also erase the edges incident to it.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;R_v&amp;lt;/math&amp;gt; denotes the number of vertices from which vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt; is reachable. Prove that the expected value of the number of operations equals to &amp;lt;math&amp;gt;\sum_{v \in V} \frac{1}{R_v}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Expected Mex&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X_1,X_2,\ldots,X_{100} \sim \mathrm{Geo}(1/2)&amp;lt;/math&amp;gt; be independent random variables. Compute &amp;lt;math&amp;gt;\mathbf{E}[\mathrm{mex}(X_1,X_2,\ldots,X_{100})]&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\mathrm{mex}(a_1,a_2,\ldots,a_n)&amp;lt;/math&amp;gt; is the smallest positive integer that does not appear in &amp;lt;math&amp;gt;a_1,a_2,\ldots,a_n&amp;lt;/math&amp;gt;. Your answer is considered correct if the absolute error does not exceed &amp;lt;math&amp;gt;10^{-6}&amp;lt;/math&amp;gt;. (Hint: Although mathematics will help you arrive at elegant and efficient methods, the use of a computer and programming skills will be required.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 5 (Probability meets graph theory) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random social networks&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G = (V, E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph without isolating vertex. &lt;br /&gt;
Let &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; be the degree of vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Let &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be a uniformly chosen vertex, and &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; a uniformly chosen neighbor of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Show that &amp;lt;math&amp;gt;\mathbf{E}[d_Z] \geq \mathbf{E}[d_Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Interpret this inequality in the context of social networks, in which the vertices represent people, and the edges represent friendship.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Turán&#039;s Theorem&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G=(V,E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph, and write &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; for the degree of the vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Use probablistic method to prove that &amp;lt;math&amp;gt;\alpha(G) \ge \sum_{v \in V} \frac{1}{d_v + 1}&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\alpha(G)&amp;lt;/math&amp;gt; is the size of a maximum independent set. (Hint: Consider the following random procedure for generating an independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; from a graph with vertex set &amp;lt;math&amp;gt;V&amp;lt;/math&amp;gt;: First, generate a random permutation of the vertices, denoted as &amp;lt;math&amp;gt;v_1,v_2,\ldots,v_n&amp;lt;/math&amp;gt;. Then, construct the independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; as follows: For each vertex &amp;lt;math&amp;gt;v_i \in V&amp;lt;/math&amp;gt;, add &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; if and only if none of its predecessors in the permutation, i.e., &amp;lt;math&amp;gt;v_1,\ldots,v_{i-1}&amp;lt;/math&amp;gt;, are neighbors of &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Tournament&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Prove that there is a [https://en.wikipedia.org/wiki/Tournament_(graph_theory) tournament] &amp;lt;math&amp;gt;T&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; players and at least &amp;lt;math&amp;gt;n!/2^{n-1}&amp;lt;/math&amp;gt;Hamiltonian paths.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 6 (1D random walk, 8 points) ==&lt;br /&gt;
Let &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt; be a constant, and &amp;lt;math&amp;gt;\{X_n\}_{n \ge 1}&amp;lt;/math&amp;gt; be independent Bernoulli trials with successful probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;.&lt;br /&gt;
Define &amp;lt;math&amp;gt;S_n = 2\sum_{i=1}^n X_i - n&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;S_0 = 0&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Range of random walk&amp;lt;/strong&amp;gt;] The range &amp;lt;math&amp;gt;R_n&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;S_0, S_1, \ldots, S_n&amp;lt;/math&amp;gt; is defined as the number of distinct values taken by the sequence. Show that &amp;lt;math&amp;gt;\mathbf{Pr}\left(R_n = R_{n-1}+1\right) = \mathbf{Pr}\left(\forall 1 \le i \le n, S_i \neq 0\right)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, and deduce that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n]\to \mathbf{Pr}(\forall i \ge 1, S_i \neq 0)&amp;lt;/math&amp;gt;. Hence show that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n] \to |2p-1|&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Symmetric 1D random walk (III)&amp;lt;/strong&amp;gt;] Suppose &amp;lt;math&amp;gt;p = \frac{1}{2}&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{E}[|S_n|] = \Theta(\sqrt{n})&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12957</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 2</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12957"/>
		<updated>2025-03-13T15:31:45Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: /* Problem 2 (Distribution of random variable, 8 points) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 2==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Warm-up problems, 12 points) ==&lt;br /&gt;
* [&#039;&#039;&#039;Function of random variable (I)&#039;&#039;&#039;] Express the distribution functions of [math]X^+ = \max\{0,X\}[/math], [math]X^- = -\min\{0,X\}[/math], [math]|X|=X^+ + X^-[/math], [math]-X[/math], in terms of the distribution function [math]F[/math] of the random variable [math]X[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Function of random variable (II)&#039;&#039;&#039;] Let [math]X[/math] be a random variable with distribution function [math]\max(0,\min(1,x))[/math]. Let [math]F[/math] be a distribution function which is continuous and strictly increasing. Show that [math]Y=F^{-1}(X)[/math] be a random variable with distribution function [math]F[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Independence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;1\leq r\leq n&amp;lt;/math&amp;gt; be independent random variables which are symmetric about &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;; that is, &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-X_r&amp;lt;/math&amp;gt; have the same distributions. Show that, for all &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}[S_n \geq x] = \mathbf{Pr}[S_n \leq -x]&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;S_n = \sum_{r=1}^n X_r&amp;lt;/math&amp;gt;. Is the conclusion true without the assumtion of independence?&lt;br /&gt;
* [&#039;&#039;&#039;Dependence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be discrete random variables with joint mass function &amp;lt;math&amp;gt;f(x,y) = \frac{C}{(x+y-1)(x+y)(x+y+1)}&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;x,y \in \mathbb{N}_+&amp;lt;/math&amp;gt; (in other words, &amp;lt;math&amp;gt;x,y = 1,2,3,\cdots&amp;lt;/math&amp;gt;). Find (1) the value of &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;, (2) marginal mass function of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and (3) &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt;.&lt;br /&gt;
* [&#039;&#039;&#039;Expectation&#039;&#039;&#039;] It is required to place in order &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; books &amp;lt;math&amp;gt;B_1, B_2, \cdots, B_n&amp;lt;/math&amp;gt; on a library shelf in such way that readers searching from left to right waste as little time as possible on average. Assuming that a random reader requires book &amp;lt;math&amp;gt;B_i&amp;lt;/math&amp;gt; with probability &amp;lt;math&amp;gt;p_i&amp;lt;/math&amp;gt;, find the ordering of the books which minimizes the number of titles examined by a random reader before discovery of the required book.&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Entropy of discrete random variable&amp;lt;/strong&amp;gt;] Let [math]X[/math] be a discrete random variable with range of values [math][N] = \{1,2,\ldots,N\}[/math] and probability mass function [math]p[/math]. Define [math]H(X) = -\sum_{n \ge 1} p(n) \log p(n)[/math] with convention [math]0\log 0 = 0[/math]. Prove that [math]H(X) \le \log N[/math] using Jensen&#039;s inequality.&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Law of total expectation&amp;lt;/strong&amp;gt;] Let [math]X \sim \mathrm{Geom}(p)[/math] for some parameter [math]p \in (0,1)[/math]. Calculate [math]\mathbf{E}[X][/math] using the law of total expectation.&lt;br /&gt;
&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Random number of random variables&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\{X_n\}_{n \ge 1}&amp;lt;/math&amp;gt; be identically distributed random variable and &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; be a random variable taking values in the non-negative integers and independent of the &amp;lt;math&amp;gt;X_n&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;n \ge 1&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{E}\left[\sum_{i=1}^N X_i\right] = \mathbf{E}[N] \mathbf{E}[X_1]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Distribution of random variable, 8 points) ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Cumulative distribution function (CDF)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a random variable with cumulative distribution function &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Show that &amp;lt;math&amp;gt;Y = aX+b&amp;lt;/math&amp;gt; is a random variable where &amp;lt;math&amp;gt;a&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;b&amp;lt;/math&amp;gt; are real constants, and express the CDF of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; by &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;. (Hint: Try expressing the event [math]Y=aX+b\le y[/math] by countably many set operations on the events defined on [math]X[/math].)&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Let &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; be the CDF of random variable &amp;lt;math&amp;gt;Z:\Omega\rightarrow \mathbb{R}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;0\leq \lambda \leq 1&amp;lt;/math&amp;gt;, show that &lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;math&amp;gt;\lambda F + (1-\lambda)G&amp;lt;/math&amp;gt; is a CDF function. &amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;The product &amp;lt;math&amp;gt;FG&amp;lt;/math&amp;gt; is a CDF function, and if &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; are independent, then &amp;lt;math&amp;gt;FG&amp;lt;/math&amp;gt; is the CDF of &amp;lt;math&amp;gt;\max\{X,Z\}&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Probability density function (PDF)&amp;lt;/strong&amp;gt;] We toss &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; coins, and each one shows heads with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;, independently of each of the others. Each coin which shows head is tossed again. (If the coin shows tail, it won&#039;t be tossed again.) Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of heads resulting from the &amp;lt;strong&amp;gt;second&amp;lt;/strong&amp;gt; round of tosses, and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the number of heads resulting from &amp;lt;strong&amp;gt;all&amp;lt;/strong&amp;gt; tosses, which includes the first and (possible) second round of each toss.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the PDF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{E}[Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Let &amp;lt;math&amp;gt;p_X&amp;lt;/math&amp;gt; be the PMF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, show that [math]p_X(k-1)p_X(k+1)\leq [p_X(k)]^2[/math] for &amp;lt;math&amp;gt;1\leq k \leq n-1&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Discrete random variable) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (I)&amp;lt;/strong&amp;gt;] Every package of some intrinsically dull commodity includes a small and exciting plastic object. There are [math]c[/math] different types of object, and each package is equally likely to contain any given type. You buy one package each day.&amp;lt;/p&amp;gt;&lt;br /&gt;
    &amp;lt;ol&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse between the acquisitions of the [math]j[/math]-th new type of object and the [math](j + 1)[/math]-th new type.&amp;lt;/li&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse before you have a full set of objects.&amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (II)&amp;lt;/strong&amp;gt;] Prove that geometric distribution is the only discrete memoryless distribution with range values &amp;lt;math&amp;gt;\mathbb{N}_+&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;n_1,n_2 \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;0 \le p \le 1&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Bin}(n_1,p),Y \sim \mathrm{Bin}(n_2,p)&amp;lt;/math&amp;gt; be independent random variables. Prove that &amp;lt;math&amp;gt;X+Y \sim \mathrm{Bin}(n_1+n_2,p)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Negative binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; follows the negative binomial distribution with parameter &amp;lt;math&amp;gt;r \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt;. Calculate &amp;lt;math&amp;gt;\mathbf{Var}[X] = \mathbf{E}[X^2] - \left(\mathbf{E}[X]\right)^2&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Hypergeometric distribution&amp;lt;/strong&amp;gt;] An urn contains [math]N[/math] balls, [math]b[/math] of which are blue and [math]r = N -b[/math] of which are red. A random sample of [math]n[/math] balls is drawn &amp;lt;strong&amp;gt;without replacement&amp;lt;/strong&amp;gt; (无放回) from the urn. Let [math]B[/math] the number of blue balls in this sample. Show that if [math]N, b[/math], and [math]r[/math] approach [math]+\infty[/math] in such a way that [math]b/N \rightarrow  p[/math] and [math]r/N \rightarrow 1 - p[/math], then [math]\mathbf{Pr}(B = k) \rightarrow {n\choose k}p^k(1-p)^{n-k}[/math] for [math]0\leq k \leq n[/math].&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Poisson distribution&amp;lt;/strong&amp;gt;] In your pocket is a random number &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; of coins, where &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; has the Poisson distribution with parameter &amp;lt;math&amp;gt;\lambda&amp;lt;/math&amp;gt;. You toss each coin once, with heads showing with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; each time. Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the (random) number of heads outcomes and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the (also random) number of tails.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the joint mass function of &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find PMF of the marginal distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; in &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;. Are &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; independent?&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Conditional distribution (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be independent &amp;lt;math&amp;gt;\text{Bin}(n, p)&amp;lt;/math&amp;gt; random variables, and let &amp;lt;math&amp;gt;Z = X + Y&amp;lt;/math&amp;gt;. Show that the conditional distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; given &amp;lt;math&amp;gt;Z = N&amp;lt;/math&amp;gt; is the hypergeometric distribution.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Conditional distribution (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;\lambda,\mu &amp;gt; 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;n \in \mathbb{N}&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Pois}(\lambda), Y \sim \mathrm{Pois}(\mu)&amp;lt;/math&amp;gt; be independent random variables. Find out the conditional distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, given &amp;lt;math&amp;gt;X+Y = n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 4 (Linearity of Expectation, 12 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Streak&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Suppose we flip a fair coin &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; times independently to obtain a sequence of flips &amp;lt;math&amp;gt;X_1, X_2, \ldots , X_n&amp;lt;/math&amp;gt;. A streak of flips is a consecutive subsequence of flips that are all the same. For example, if &amp;lt;math&amp;gt;X_3&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;X_4&amp;lt;/math&amp;gt;, and &amp;lt;math&amp;gt;X_5&amp;lt;/math&amp;gt; are all heads, there is a streak of length &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; starting at the third lip. (If &amp;lt;math&amp;gt;X_6&amp;lt;/math&amp;gt; is also heads, then there is also a streak of length &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; starting at the third lip.) Find the expected number of streaks of length &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; for some integer &amp;lt;math&amp;gt;k \ge 1&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Number of cycles&amp;lt;/strong&amp;gt;] &lt;br /&gt;
At a banquet, there are &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people who shake hands according to the following process: In each round, two idle hands are randomly selected and shaken (&amp;lt;strong&amp;gt;these two hands are no longer idle&amp;lt;/strong&amp;gt;). After &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds, there will be no idle hands left, and the &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people will form several cycles. For example, when &amp;lt;math&amp;gt;n=3&amp;lt;/math&amp;gt;, the following situation may occur: the left and right hands of the first person are held together, the left hand of the second person and the right hand of the third person are held together, and the right hand of the second person and the left hand of the third person are held together. In this case, three people form two cycles. How many cycles are expected to be formed after &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds?&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Number of operations&amp;lt;/strong&amp;gt;]&lt;br /&gt;
We have a directed graph &amp;lt;math&amp;gt;G=(V,E)&amp;lt;/math&amp;gt; without self-loops and multi-edges.&lt;br /&gt;
&lt;br /&gt;
Until the graph becomes empty, repeat the following operation:&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
Choose one (unerased) vertex uniformly at random (independently from the previous choices). Then, erase that vertex and all vertices that are reachable from the chosen vertex by traversing some edges. Erasing a vertex will also erase the edges incident to it.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;R_v&amp;lt;/math&amp;gt; denotes the number of vertices from which vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt; is reachable. Prove that the expected value of the number of operations equals to &amp;lt;math&amp;gt;\sum_{v \in V} \frac{1}{R_v}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Expected Mex&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X_1,X_2,\ldots,X_{100} \sim \mathrm{Geo}(1/2)&amp;lt;/math&amp;gt; be independent random variables. Compute &amp;lt;math&amp;gt;\mathbf{E}[\mathrm{mex}(X_1,X_2,\ldots,X_{100})]&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\mathrm{mex}(a_1,a_2,\ldots,a_n)&amp;lt;/math&amp;gt; is the smallest positive integer that does not appear in &amp;lt;math&amp;gt;a_1,a_2,\ldots,a_n&amp;lt;/math&amp;gt;. Your answer is considered correct if the absolute error does not exceed &amp;lt;math&amp;gt;10^{-6}&amp;lt;/math&amp;gt;. (Hint: Although mathematics will help you arrive at elegant and efficient methods, the use of a computer and programming skills will be required.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 5 (Probability meets graph theory) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random social networks&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G = (V, E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph without isolating vertex. &lt;br /&gt;
Let &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; be the degree of vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Let &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be a uniformly chosen vertex, and &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; a uniformly chosen neighbor of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Show that &amp;lt;math&amp;gt;\mathbf{E}[d_Z] \geq \mathbf{E}[d_Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Interpret this inequality in the context of social networks, in which the vertices represent people, and the edges represent friendship.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Turán&#039;s Theorem&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G=(V,E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph, and write &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; for the degree of the vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Use probablistic method to prove that &amp;lt;math&amp;gt;\alpha(G) \ge \sum_{v \in V} \frac{1}{d_v + 1}&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\alpha(G)&amp;lt;/math&amp;gt; is the size of a maximum independent set. (Hint: Consider the following random procedure for generating an independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; from a graph with vertex set &amp;lt;math&amp;gt;V&amp;lt;/math&amp;gt;: First, generate a random permutation of the vertices, denoted as &amp;lt;math&amp;gt;v_1,v_2,\ldots,v_n&amp;lt;/math&amp;gt;. Then, construct the independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; as follows: For each vertex &amp;lt;math&amp;gt;v_i \in V&amp;lt;/math&amp;gt;, add &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; if and only if none of its predecessors in the permutation, i.e., &amp;lt;math&amp;gt;v_1,\ldots,v_{i-1}&amp;lt;/math&amp;gt;, are neighbors of &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Tournament&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Prove that there is a [https://en.wikipedia.org/wiki/Tournament_(graph_theory) tournament] &amp;lt;math&amp;gt;T&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; players and at least &amp;lt;math&amp;gt;n!/2^{n-1}&amp;lt;/math&amp;gt;Hamiltonian paths.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 6 (1D random walk, 8 points) ==&lt;br /&gt;
Let &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt; be a constant, and &amp;lt;math&amp;gt;\{X_n\}_{n \ge 1}&amp;lt;/math&amp;gt; be independent Bernoulli trials with successful probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;.&lt;br /&gt;
Define &amp;lt;math&amp;gt;S_n = 2\sum_{i=1}^n X_i - n&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;S_0 = 0&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Range of random walk&amp;lt;/strong&amp;gt;] The range &amp;lt;math&amp;gt;R_n&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;S_0, S_1, \ldots, S_n&amp;lt;/math&amp;gt; is defined as the number of distinct values taken by the sequence. Show that &amp;lt;math&amp;gt;\mathbf{Pr}\left(R_n = R_{n-1}+1\right) = \mathbf{Pr}\left(\forall 1 \le i \le n, S_i \neq 0\right)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, and deduce that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n]\to \mathbf{Pr}(\forall i \ge 1, S_i \neq 0)&amp;lt;/math&amp;gt;. Hence show that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n] \to |2p-1|&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Symmetric 1D random walk (III)&amp;lt;/strong&amp;gt;] Suppose &amp;lt;math&amp;gt;p = \frac{1}{2}&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{E}[|S_n|] = \Theta(\sqrt{n})&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12955</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 2</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12955"/>
		<updated>2025-03-13T15:28:48Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: /* Problem 1 (Warm-up problems, 12 points) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 2==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Warm-up problems, 12 points) ==&lt;br /&gt;
* [&#039;&#039;&#039;Function of random variable (I)&#039;&#039;&#039;] Express the distribution functions of [math]X^+ = \max\{0,X\}[/math], [math]X^- = -\min\{0,X\}[/math], [math]|X|=X^+ + X^-[/math], [math]-X[/math], in terms of the distribution function [math]F[/math] of the random variable [math]X[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Function of random variable (II)&#039;&#039;&#039;] Let [math]X[/math] be a random variable with distribution function [math]\max(0,\min(1,x))[/math]. Let [math]F[/math] be a distribution function which is continuous and strictly increasing. Show that [math]Y=F^{-1}(X)[/math] be a random variable with distribution function [math]F[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Independence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;1\leq r\leq n&amp;lt;/math&amp;gt; be independent random variables which are symmetric about &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;; that is, &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-X_r&amp;lt;/math&amp;gt; have the same distributions. Show that, for all &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}[S_n \geq x] = \mathbf{Pr}[S_n \leq -x]&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;S_n = \sum_{r=1}^n X_r&amp;lt;/math&amp;gt;. Is the conclusion true without the assumtion of independence?&lt;br /&gt;
* [&#039;&#039;&#039;Dependence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be discrete random variables with joint mass function &amp;lt;math&amp;gt;f(x,y) = \frac{C}{(x+y-1)(x+y)(x+y+1)}&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;x,y \in \mathbb{N}_+&amp;lt;/math&amp;gt; (in other words, &amp;lt;math&amp;gt;x,y = 1,2,3,\cdots&amp;lt;/math&amp;gt;). Find (1) the value of &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;, (2) marginal mass function of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and (3) &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt;.&lt;br /&gt;
* [&#039;&#039;&#039;Expectation&#039;&#039;&#039;] It is required to place in order &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; books &amp;lt;math&amp;gt;B_1, B_2, \cdots, B_n&amp;lt;/math&amp;gt; on a library shelf in such way that readers searching from left to right waste as little time as possible on average. Assuming that a random reader requires book &amp;lt;math&amp;gt;B_i&amp;lt;/math&amp;gt; with probability &amp;lt;math&amp;gt;p_i&amp;lt;/math&amp;gt;, find the ordering of the books which minimizes the number of titles examined by a random reader before discovery of the required book.&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Entropy of discrete random variable&amp;lt;/strong&amp;gt;] Let [math]X[/math] be a discrete random variable with range of values [math][N] = \{1,2,\ldots,N\}[/math] and probability mass function [math]p[/math]. Define [math]H(X) = -\sum_{n \ge 1} p(n) \log p(n)[/math] with convention [math]0\log 0 = 0[/math]. Prove that [math]H(X) \le \log N[/math] using Jensen&#039;s inequality.&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Law of total expectation&amp;lt;/strong&amp;gt;] Let [math]X \sim \mathrm{Geom}(p)[/math] for some parameter [math]p \in (0,1)[/math]. Calculate [math]\mathbf{E}[X][/math] using the law of total expectation.&lt;br /&gt;
&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Random number of random variables&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\{X_n\}_{n \ge 1}&amp;lt;/math&amp;gt; be identically distributed random variable and &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; be a random variable taking values in the non-negative integers and independent of the &amp;lt;math&amp;gt;X_n&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;n \ge 1&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{E}\left[\sum_{i=1}^N X_i\right] = \mathbf{E}[N] \mathbf{E}[X_1]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Distribution of random variable, 8 points) ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Cumulative distribution function (CDF)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a random variable with cumulative distribution function &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Show that &amp;lt;math&amp;gt;Y = aX+b&amp;lt;/math&amp;gt; is a random variable where &amp;lt;math&amp;gt;a&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;b&amp;lt;/math&amp;gt; are real constants, and express the CDF of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; by &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;. (Hint: Try expressing the event [math]Y=aX+b\le y[/math] by countably many set operations on the events defined on [math]X[/math].)&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Let &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; be the CDF of random variable &amp;lt;math&amp;gt;Z:\Omega\rightarrow \mathbb{R}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;0\leq \lambda \leq 1&amp;lt;/math&amp;gt;, show that &lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;math&amp;gt;\lambda F + (1-\lambda)G&amp;lt;/math&amp;gt; is a CDF function. &amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;The product &amp;lt;math&amp;gt;FG&amp;lt;/math&amp;gt; is a CDF function, and if &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; are independent, then &amp;lt;math&amp;gt;FG&amp;lt;/math&amp;gt; is the CDF of &amp;lt;math&amp;gt;\max\{X,Z\}&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Probability mass function (PMF)&amp;lt;/strong&amp;gt;] We toss &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; coins, and each one shows heads with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;, independently of each of the others. Each coin which shows head is tossed again. (If the coin shows tail, it won&#039;t be tossed again.) Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of heads resulting from the &amp;lt;strong&amp;gt;second&amp;lt;/strong&amp;gt; round of tosses, and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the number of heads resulting from &amp;lt;strong&amp;gt;all&amp;lt;/strong&amp;gt; tosses, which includes the first and (possible) second round of each toss.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the PMF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{E}[Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Let &amp;lt;math&amp;gt;p_X&amp;lt;/math&amp;gt; be the PMF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, show that [math]p_X(k-1)p_X(k+1)\leq [p_X(k)]^2[/math] for &amp;lt;math&amp;gt;1\leq k \leq n-1&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Discrete random variable) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (I)&amp;lt;/strong&amp;gt;] Every package of some intrinsically dull commodity includes a small and exciting plastic object. There are [math]c[/math] different types of object, and each package is equally likely to contain any given type. You buy one package each day.&amp;lt;/p&amp;gt;&lt;br /&gt;
    &amp;lt;ol&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse between the acquisitions of the [math]j[/math]-th new type of object and the [math](j + 1)[/math]-th new type.&amp;lt;/li&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse before you have a full set of objects.&amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (II)&amp;lt;/strong&amp;gt;] Prove that geometric distribution is the only discrete memoryless distribution with range values &amp;lt;math&amp;gt;\mathbb{N}_+&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;n_1,n_2 \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;0 \le p \le 1&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Bin}(n_1,p),Y \sim \mathrm{Bin}(n_2,p)&amp;lt;/math&amp;gt; be independent random variables. Prove that &amp;lt;math&amp;gt;X+Y \sim \mathrm{Bin}(n_1+n_2,p)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Negative binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; follows the negative binomial distribution with parameter &amp;lt;math&amp;gt;r \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt;. Calculate &amp;lt;math&amp;gt;\mathbf{Var}[X] = \mathbf{E}[X^2] - \left(\mathbf{E}[X]\right)^2&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Hypergeometric distribution&amp;lt;/strong&amp;gt;] An urn contains [math]N[/math] balls, [math]b[/math] of which are blue and [math]r = N -b[/math] of which are red. A random sample of [math]n[/math] balls is drawn &amp;lt;strong&amp;gt;without replacement&amp;lt;/strong&amp;gt; (无放回) from the urn. Let [math]B[/math] the number of blue balls in this sample. Show that if [math]N, b[/math], and [math]r[/math] approach [math]+\infty[/math] in such a way that [math]b/N \rightarrow  p[/math] and [math]r/N \rightarrow 1 - p[/math], then [math]\mathbf{Pr}(B = k) \rightarrow {n\choose k}p^k(1-p)^{n-k}[/math] for [math]0\leq k \leq n[/math].&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Poisson distribution&amp;lt;/strong&amp;gt;] In your pocket is a random number &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; of coins, where &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; has the Poisson distribution with parameter &amp;lt;math&amp;gt;\lambda&amp;lt;/math&amp;gt;. You toss each coin once, with heads showing with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; each time. Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the (random) number of heads outcomes and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the (also random) number of tails.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the joint mass function of &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find PMF of the marginal distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; in &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;. Are &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; independent?&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Conditional distribution (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be independent &amp;lt;math&amp;gt;\text{Bin}(n, p)&amp;lt;/math&amp;gt; random variables, and let &amp;lt;math&amp;gt;Z = X + Y&amp;lt;/math&amp;gt;. Show that the conditional distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; given &amp;lt;math&amp;gt;Z = N&amp;lt;/math&amp;gt; is the hypergeometric distribution.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Conditional distribution (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;\lambda,\mu &amp;gt; 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;n \in \mathbb{N}&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Pois}(\lambda), Y \sim \mathrm{Pois}(\mu)&amp;lt;/math&amp;gt; be independent random variables. Find out the conditional distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, given &amp;lt;math&amp;gt;X+Y = n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 4 (Linearity of Expectation, 12 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Streak&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Suppose we flip a fair coin &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; times independently to obtain a sequence of flips &amp;lt;math&amp;gt;X_1, X_2, \ldots , X_n&amp;lt;/math&amp;gt;. A streak of flips is a consecutive subsequence of flips that are all the same. For example, if &amp;lt;math&amp;gt;X_3&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;X_4&amp;lt;/math&amp;gt;, and &amp;lt;math&amp;gt;X_5&amp;lt;/math&amp;gt; are all heads, there is a streak of length &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; starting at the third lip. (If &amp;lt;math&amp;gt;X_6&amp;lt;/math&amp;gt; is also heads, then there is also a streak of length &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; starting at the third lip.) Find the expected number of streaks of length &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; for some integer &amp;lt;math&amp;gt;k \ge 1&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Number of cycles&amp;lt;/strong&amp;gt;] &lt;br /&gt;
At a banquet, there are &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people who shake hands according to the following process: In each round, two idle hands are randomly selected and shaken (&amp;lt;strong&amp;gt;these two hands are no longer idle&amp;lt;/strong&amp;gt;). After &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds, there will be no idle hands left, and the &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people will form several cycles. For example, when &amp;lt;math&amp;gt;n=3&amp;lt;/math&amp;gt;, the following situation may occur: the left and right hands of the first person are held together, the left hand of the second person and the right hand of the third person are held together, and the right hand of the second person and the left hand of the third person are held together. In this case, three people form two cycles. How many cycles are expected to be formed after &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds?&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Number of operations&amp;lt;/strong&amp;gt;]&lt;br /&gt;
We have a directed graph &amp;lt;math&amp;gt;G=(V,E)&amp;lt;/math&amp;gt; without self-loops and multi-edges.&lt;br /&gt;
&lt;br /&gt;
Until the graph becomes empty, repeat the following operation:&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
Choose one (unerased) vertex uniformly at random (independently from the previous choices). Then, erase that vertex and all vertices that are reachable from the chosen vertex by traversing some edges. Erasing a vertex will also erase the edges incident to it.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;R_v&amp;lt;/math&amp;gt; denotes the number of vertices from which vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt; is reachable. Prove that the expected value of the number of operations equals to &amp;lt;math&amp;gt;\sum_{v \in V} \frac{1}{R_v}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Expected Mex&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X_1,X_2,\ldots,X_{100} \sim \mathrm{Geo}(1/2)&amp;lt;/math&amp;gt; be independent random variables. Compute &amp;lt;math&amp;gt;\mathbf{E}[\mathrm{mex}(X_1,X_2,\ldots,X_{100})]&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\mathrm{mex}(a_1,a_2,\ldots,a_n)&amp;lt;/math&amp;gt; is the smallest positive integer that does not appear in &amp;lt;math&amp;gt;a_1,a_2,\ldots,a_n&amp;lt;/math&amp;gt;. Your answer is considered correct if the absolute error does not exceed &amp;lt;math&amp;gt;10^{-6}&amp;lt;/math&amp;gt;. (Hint: Although mathematics will help you arrive at elegant and efficient methods, the use of a computer and programming skills will be required.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 5 (Probability meets graph theory) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random social networks&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G = (V, E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph without isolating vertex. &lt;br /&gt;
Let &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; be the degree of vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Let &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be a uniformly chosen vertex, and &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; a uniformly chosen neighbor of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Show that &amp;lt;math&amp;gt;\mathbf{E}[d_Z] \geq \mathbf{E}[d_Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Interpret this inequality in the context of social networks, in which the vertices represent people, and the edges represent friendship.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Turán&#039;s Theorem&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G=(V,E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph, and write &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; for the degree of the vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Use probablistic method to prove that &amp;lt;math&amp;gt;\alpha(G) \ge \sum_{v \in V} \frac{1}{d_v + 1}&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\alpha(G)&amp;lt;/math&amp;gt; is the size of a maximum independent set. (Hint: Consider the following random procedure for generating an independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; from a graph with vertex set &amp;lt;math&amp;gt;V&amp;lt;/math&amp;gt;: First, generate a random permutation of the vertices, denoted as &amp;lt;math&amp;gt;v_1,v_2,\ldots,v_n&amp;lt;/math&amp;gt;. Then, construct the independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; as follows: For each vertex &amp;lt;math&amp;gt;v_i \in V&amp;lt;/math&amp;gt;, add &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; if and only if none of its predecessors in the permutation, i.e., &amp;lt;math&amp;gt;v_1,\ldots,v_{i-1}&amp;lt;/math&amp;gt;, are neighbors of &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Dominating set&amp;lt;/strong&amp;gt;]&lt;br /&gt;
A &#039;&#039;dominating set&#039;&#039; of vertices in an undirected graph &amp;lt;math&amp;gt;G = (V, E)&amp;lt;/math&amp;gt; is a set &amp;lt;math&amp;gt;S \subseteq V&amp;lt;/math&amp;gt; such that every vertex of&lt;br /&gt;
&amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; belongs to &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; or has a neighbor in &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;. &lt;br /&gt;
&lt;br /&gt;
Let &amp;lt;math&amp;gt;G = (V, E)&amp;lt;/math&amp;gt; be an &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;-vertex graph with minimum degree &amp;lt;math&amp;gt;d &amp;gt; 1&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; has a dominating set with at most &amp;lt;math&amp;gt;\frac{n\left(1+\log(d+1)\right)}{d+1}&amp;lt;/math&amp;gt; vertices. (Hint: Consider a random vertex subset &amp;lt;math&amp;gt;S \subseteq V&amp;lt;/math&amp;gt; by including each vertex independently with&lt;br /&gt;
probability &amp;lt;math&amp;gt;p := \log(d + 1)/(d + 1)&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 6 (1D random walk, 8 points) ==&lt;br /&gt;
Let &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt; be a constant, and &amp;lt;math&amp;gt;\{X_n\}_{n \ge 1}&amp;lt;/math&amp;gt; be independent Bernoulli trials with successful probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;.&lt;br /&gt;
Define &amp;lt;math&amp;gt;S_n = 2\sum_{i=1}^n X_i - n&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;S_0 = 0&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Range of random walk&amp;lt;/strong&amp;gt;] The range &amp;lt;math&amp;gt;R_n&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;S_0, S_1, \ldots, S_n&amp;lt;/math&amp;gt; is defined as the number of distinct values taken by the sequence. Show that &amp;lt;math&amp;gt;\mathbf{Pr}\left(R_n = R_{n-1}+1\right) = \mathbf{Pr}\left(\forall 1 \le i \le n, S_i \neq 0\right)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, and deduce that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n]\to \mathbf{Pr}(\forall i \ge 1, S_i \neq 0)&amp;lt;/math&amp;gt;. Hence show that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n] \to |2p-1|&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Symmetric 1D random walk (III)&amp;lt;/strong&amp;gt;] Suppose &amp;lt;math&amp;gt;p = \frac{1}{2}&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{E}[|S_n|] = \Theta(\sqrt{n})&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12954</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 2</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12954"/>
		<updated>2025-03-13T15:21:47Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: /* Problem 1 (Warm-up problems, 12 points) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 2==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Warm-up problems, 12 points) ==&lt;br /&gt;
* [&#039;&#039;&#039;Function of random variable (I)&#039;&#039;&#039;] Express the distribution functions of [math]X^+ = \max\{0,X\}[/math], [math]X^- = -\min\{0,X\}[/math], [math]|X|=X^+ + X^-[/math], [math]-X[/math], in terms of the distribution function [math]F[/math] of the random variable [math]X[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Function of random variable (II)&#039;&#039;&#039;] Let [math]X[/math] be a random variable with distribution function [math]\max(0,\min(1,x))[/math]. Let [math]F[/math] be a distribution function which is continuous and strictly increasing. Show that [math]Y=F^{-1}(X)[/math] be a random variable with distribution function [math]F[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Independence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;1\leq r\leq n&amp;lt;/math&amp;gt; be independent random variables which are symmetric about &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;; that is, &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-X_r&amp;lt;/math&amp;gt; have the same distributions. Show that, for all &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}[S_n \geq x] = \mathbf{Pr}[S_n \leq -x]&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;S_n = \sum_{r=1}^n X_r&amp;lt;/math&amp;gt;. Is the conclusion true without the assumtion of independence?&lt;br /&gt;
* [&#039;&#039;&#039;Dependence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be discrete random variables with joint mass function &amp;lt;math&amp;gt;f(x,y) = \frac{C}{(x+y-1)(x+y)(x+y+1)}&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;x,y \in \mathbb{N}_+&amp;lt;/math&amp;gt; (in other words, &amp;lt;math&amp;gt;x,y = 1,2,3,\cdots&amp;lt;/math&amp;gt;). Find (1) the value of &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;, (2) marginal mass function of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and (3) &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt;.&lt;br /&gt;
* [&#039;&#039;&#039;Sorted books&#039;&#039;&#039;]&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Entropy of discrete random variable&amp;lt;/strong&amp;gt;] Let [math]X[/math] be a discrete random variable with range of values [math][N] = \{1,2,\ldots,N\}[/math] and probability mass function [math]p[/math]. Define [math]H(X) = -\sum_{n \ge 1} p(n) \log p(n)[/math] with convention [math]0\log 0 = 0[/math]. Prove that [math]H(X) \le \log N[/math] using Jensen&#039;s inequality.&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Law of total expectation&amp;lt;/strong&amp;gt;] Let [math]X \sim \mathrm{Geom}(p)[/math] for some parameter [math]p \in (0,1)[/math]. Calculate [math]\mathbf{E}[X][/math] using the law of total expectation.&lt;br /&gt;
&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Random number of random variables&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\{X_n\}_{n \ge 1}&amp;lt;/math&amp;gt; be identically distributed random variable and &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; be a random variable taking values in the non-negative integers and independent of the &amp;lt;math&amp;gt;X_n&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;n \ge 1&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{E}\left[\sum_{i=1}^N X_i\right] = \mathbf{E}[N] \mathbf{E}[X_1]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Distribution of random variable, 8 points) ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Cumulative distribution function (CDF)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a random variable with cumulative distribution function &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Show that &amp;lt;math&amp;gt;Y = aX+b&amp;lt;/math&amp;gt; is a random variable where &amp;lt;math&amp;gt;a&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;b&amp;lt;/math&amp;gt; are real constants, and express the CDF of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; by &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;. (Hint: Try expressing the event [math]Y=aX+b\le y[/math] by countably many set operations on the events defined on [math]X[/math].)&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Let &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; be the CDF of random variable &amp;lt;math&amp;gt;Z:\Omega\rightarrow \mathbb{R}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;0\leq \lambda \leq 1&amp;lt;/math&amp;gt;, show that &lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;math&amp;gt;\lambda F + (1-\lambda)G&amp;lt;/math&amp;gt; is a CDF function. &amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;The product &amp;lt;math&amp;gt;FG&amp;lt;/math&amp;gt; is a CDF function, and if &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; are independent, then &amp;lt;math&amp;gt;FG&amp;lt;/math&amp;gt; is the CDF of &amp;lt;math&amp;gt;\max\{X,Z\}&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Probability mass function (PMF)&amp;lt;/strong&amp;gt;] We toss &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; coins, and each one shows heads with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;, independently of each of the others. Each coin which shows head is tossed again. (If the coin shows tail, it won&#039;t be tossed again.) Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of heads resulting from the &amp;lt;strong&amp;gt;second&amp;lt;/strong&amp;gt; round of tosses, and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the number of heads resulting from &amp;lt;strong&amp;gt;all&amp;lt;/strong&amp;gt; tosses, which includes the first and (possible) second round of each toss.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the PMF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{E}[Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Let &amp;lt;math&amp;gt;p_X&amp;lt;/math&amp;gt; be the PMF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, show that [math]p_X(k-1)p_X(k+1)\leq [p_X(k)]^2[/math] for &amp;lt;math&amp;gt;1\leq k \leq n-1&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Discrete random variable) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (I)&amp;lt;/strong&amp;gt;] Every package of some intrinsically dull commodity includes a small and exciting plastic object. There are [math]c[/math] different types of object, and each package is equally likely to contain any given type. You buy one package each day.&amp;lt;/p&amp;gt;&lt;br /&gt;
    &amp;lt;ol&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse between the acquisitions of the [math]j[/math]-th new type of object and the [math](j + 1)[/math]-th new type.&amp;lt;/li&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse before you have a full set of objects.&amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (II)&amp;lt;/strong&amp;gt;] Prove that geometric distribution is the only discrete memoryless distribution with range values &amp;lt;math&amp;gt;\mathbb{N}_+&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;n_1,n_2 \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;0 \le p \le 1&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Bin}(n_1,p),Y \sim \mathrm{Bin}(n_2,p)&amp;lt;/math&amp;gt; be independent random variables. Prove that &amp;lt;math&amp;gt;X+Y \sim \mathrm{Bin}(n_1+n_2,p)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Negative binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; follows the negative binomial distribution with parameter &amp;lt;math&amp;gt;r \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt;. Calculate &amp;lt;math&amp;gt;\mathbf{Var}[X] = \mathbf{E}[X^2] - \left(\mathbf{E}[X]\right)^2&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Hypergeometric distribution&amp;lt;/strong&amp;gt;] An urn contains [math]N[/math] balls, [math]b[/math] of which are blue and [math]r = N -b[/math] of which are red. A random sample of [math]n[/math] balls is drawn &amp;lt;strong&amp;gt;without replacement&amp;lt;/strong&amp;gt; (无放回) from the urn. Let [math]B[/math] the number of blue balls in this sample. Show that if [math]N, b[/math], and [math]r[/math] approach [math]+\infty[/math] in such a way that [math]b/N \rightarrow  p[/math] and [math]r/N \rightarrow 1 - p[/math], then [math]\mathbf{Pr}(B = k) \rightarrow {n\choose k}p^k(1-p)^{n-k}[/math] for [math]0\leq k \leq n[/math].&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Poisson distribution&amp;lt;/strong&amp;gt;] In your pocket is a random number &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; of coins, where &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; has the Poisson distribution with parameter &amp;lt;math&amp;gt;\lambda&amp;lt;/math&amp;gt;. You toss each coin once, with heads showing with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; each time. Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the (random) number of heads outcomes and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the (also random) number of tails.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the joint mass function of &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find PMF of the marginal distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; in &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;. Are &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; independent?&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Conditional distribution (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be independent &amp;lt;math&amp;gt;\text{Bin}(n, p)&amp;lt;/math&amp;gt; random variables, and let &amp;lt;math&amp;gt;Z = X + Y&amp;lt;/math&amp;gt;. Show that the conditional distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; given &amp;lt;math&amp;gt;Z = N&amp;lt;/math&amp;gt; is the hypergeometric distribution.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Conditional distribution (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;\lambda,\mu &amp;gt; 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;n \in \mathbb{N}&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Pois}(\lambda), Y \sim \mathrm{Pois}(\mu)&amp;lt;/math&amp;gt; be independent random variables. Find out the conditional distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, given &amp;lt;math&amp;gt;X+Y = n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 4 (Linearity of Expectation, 12 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Streak&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Suppose we flip a fair coin &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; times independently to obtain a sequence of flips &amp;lt;math&amp;gt;X_1, X_2, \ldots , X_n&amp;lt;/math&amp;gt;. A streak of flips is a consecutive subsequence of flips that are all the same. For example, if &amp;lt;math&amp;gt;X_3&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;X_4&amp;lt;/math&amp;gt;, and &amp;lt;math&amp;gt;X_5&amp;lt;/math&amp;gt; are all heads, there is a streak of length &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; starting at the third lip. (If &amp;lt;math&amp;gt;X_6&amp;lt;/math&amp;gt; is also heads, then there is also a streak of length &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; starting at the third lip.) Find the expected number of streaks of length &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; for some integer &amp;lt;math&amp;gt;k \ge 1&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Number of cycles&amp;lt;/strong&amp;gt;] &lt;br /&gt;
At a banquet, there are &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people who shake hands according to the following process: In each round, two idle hands are randomly selected and shaken (&amp;lt;strong&amp;gt;these two hands are no longer idle&amp;lt;/strong&amp;gt;). After &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds, there will be no idle hands left, and the &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people will form several cycles. For example, when &amp;lt;math&amp;gt;n=3&amp;lt;/math&amp;gt;, the following situation may occur: the left and right hands of the first person are held together, the left hand of the second person and the right hand of the third person are held together, and the right hand of the second person and the left hand of the third person are held together. In this case, three people form two cycles. How many cycles are expected to be formed after &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds?&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Number of operations&amp;lt;/strong&amp;gt;]&lt;br /&gt;
We have a directed graph &amp;lt;math&amp;gt;G=(V,E)&amp;lt;/math&amp;gt; without self-loops and multi-edges.&lt;br /&gt;
&lt;br /&gt;
Until the graph becomes empty, repeat the following operation:&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
Choose one (unerased) vertex uniformly at random (independently from the previous choices). Then, erase that vertex and all vertices that are reachable from the chosen vertex by traversing some edges. Erasing a vertex will also erase the edges incident to it.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;R_v&amp;lt;/math&amp;gt; denotes the number of vertices from which vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt; is reachable. Prove that the expected value of the number of operations equals to &amp;lt;math&amp;gt;\sum_{v \in V} \frac{1}{R_v}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Expected Mex&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X_1,X_2,\ldots,X_{100} \sim \mathrm{Geo}(1/2)&amp;lt;/math&amp;gt; be independent random variables. Compute &amp;lt;math&amp;gt;\mathbf{E}[\mathrm{mex}(X_1,X_2,\ldots,X_{100})]&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\mathrm{mex}(a_1,a_2,\ldots,a_n)&amp;lt;/math&amp;gt; is the smallest positive integer that does not appear in &amp;lt;math&amp;gt;a_1,a_2,\ldots,a_n&amp;lt;/math&amp;gt;. Your answer is considered correct if the absolute error does not exceed &amp;lt;math&amp;gt;10^{-6}&amp;lt;/math&amp;gt;. (Hint: Although mathematics will help you arrive at elegant and efficient methods, the use of a computer and programming skills will be required.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 5 (Probability meets graph theory) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random social networks&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G = (V, E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph without isolating vertex. &lt;br /&gt;
Let &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; be the degree of vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Let &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be a uniformly chosen vertex, and &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; a uniformly chosen neighbor of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Show that &amp;lt;math&amp;gt;\mathbf{E}[d_Z] \geq \mathbf{E}[d_Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Interpret this inequality in the context of social networks, in which the vertices represent people, and the edges represent friendship.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Turán&#039;s Theorem&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G=(V,E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph, and write &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; for the degree of the vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Use probablistic method to prove that &amp;lt;math&amp;gt;\alpha(G) \ge \sum_{v \in V} \frac{1}{d_v + 1}&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\alpha(G)&amp;lt;/math&amp;gt; is the size of a maximum independent set. (Hint: Consider the following random procedure for generating an independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; from a graph with vertex set &amp;lt;math&amp;gt;V&amp;lt;/math&amp;gt;: First, generate a random permutation of the vertices, denoted as &amp;lt;math&amp;gt;v_1,v_2,\ldots,v_n&amp;lt;/math&amp;gt;. Then, construct the independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; as follows: For each vertex &amp;lt;math&amp;gt;v_i \in V&amp;lt;/math&amp;gt;, add &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; if and only if none of its predecessors in the permutation, i.e., &amp;lt;math&amp;gt;v_1,\ldots,v_{i-1}&amp;lt;/math&amp;gt;, are neighbors of &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Dominating set&amp;lt;/strong&amp;gt;]&lt;br /&gt;
A &#039;&#039;dominating set&#039;&#039; of vertices in an undirected graph &amp;lt;math&amp;gt;G = (V, E)&amp;lt;/math&amp;gt; is a set &amp;lt;math&amp;gt;S \subseteq V&amp;lt;/math&amp;gt; such that every vertex of&lt;br /&gt;
&amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; belongs to &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; or has a neighbor in &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;. &lt;br /&gt;
&lt;br /&gt;
Let &amp;lt;math&amp;gt;G = (V, E)&amp;lt;/math&amp;gt; be an &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;-vertex graph with minimum degree &amp;lt;math&amp;gt;d &amp;gt; 1&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; has a dominating set with at most &amp;lt;math&amp;gt;\frac{n\left(1+\log(d+1)\right)}{d+1}&amp;lt;/math&amp;gt; vertices. (Hint: Consider a random vertex subset &amp;lt;math&amp;gt;S \subseteq V&amp;lt;/math&amp;gt; by including each vertex independently with&lt;br /&gt;
probability &amp;lt;math&amp;gt;p := \log(d + 1)/(d + 1)&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 6 (1D random walk, 8 points) ==&lt;br /&gt;
Let &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt; be a constant, and &amp;lt;math&amp;gt;\{X_n\}_{n \ge 1}&amp;lt;/math&amp;gt; be independent Bernoulli trials with successful probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;.&lt;br /&gt;
Define &amp;lt;math&amp;gt;S_n = 2\sum_{i=1}^n X_i - n&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;S_0 = 0&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Range of random walk&amp;lt;/strong&amp;gt;] The range &amp;lt;math&amp;gt;R_n&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;S_0, S_1, \ldots, S_n&amp;lt;/math&amp;gt; is defined as the number of distinct values taken by the sequence. Show that &amp;lt;math&amp;gt;\mathbf{Pr}\left(R_n = R_{n-1}+1\right) = \mathbf{Pr}\left(\forall 1 \le i \le n, S_i \neq 0\right)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, and deduce that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n]\to \mathbf{Pr}(\forall i \ge 1, S_i \neq 0)&amp;lt;/math&amp;gt;. Hence show that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n] \to |2p-1|&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Symmetric 1D random walk (III)&amp;lt;/strong&amp;gt;] Suppose &amp;lt;math&amp;gt;p = \frac{1}{2}&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{E}[|S_n|] = \Theta(\sqrt{n})&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12950</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 2</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12950"/>
		<updated>2025-03-13T15:09:45Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: /* Problem 1 (Warm-up problems, 12 points) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 2==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Warm-up problems, 12 points) ==&lt;br /&gt;
* [&#039;&#039;&#039;Function of random variable (I)&#039;&#039;&#039;] Express the distribution functions of [math]X^+ = \max\{0,X\}[/math], [math]X^- = -\min\{0,X\}[/math], [math]|X|=X^+ + X^-[/math], [math]-X[/math], in terms of the distribution function [math]F[/math] of the random variable [math]X[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Function of random variable (II)&#039;&#039;&#039;] Let [math]X[/math] be a random variable with distribution function [math]\max(0,\min(1,x))[/math]. Let [math]F[/math] be a distribution function which is continuous and strictly increasing. Show that [math]Y=F^{-1}(X)[/math] be a random variable with distribution function [math]F[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Independence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;1\leq r\leq n&amp;lt;/math&amp;gt; be independent random variables which are symmetric about &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;; that is, &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-X_r&amp;lt;/math&amp;gt; have the same distributions. Show that, for all &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}[S_n \geq x] = \mathbf{Pr}[S_n \leq -x]&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;S_n = \sum_{r=1}^n X_r&amp;lt;/math&amp;gt;. Is the conclusion true without the assumtion of independence?&lt;br /&gt;
* [&#039;&#039;&#039;Dependence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be discrete random variables with joint mass function &amp;lt;math&amp;gt;f(x,y) = \frac{C}{(x+y-1)(x+y)(x+y+1)}&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;x,y \in \mathbb{N}_+&amp;lt;/math&amp;gt; (in other words, &amp;lt;math&amp;gt;x,y = 1,2,3,\cdots&amp;lt;/math&amp;gt;). Find (1) the value of &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;, (2) marginal mass function of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and (3) &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt;.&lt;br /&gt;
* [&#039;&#039;&#039;Expectation&#039;&#039;&#039;] Is it generally true that [math]\mathbf{E}[1/X] = 1/\mathbf{E}[X][/math]? Is it ever true that [math]\mathbf{E}[1/X] = 1/\mathbf{E}[X][/math]?&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Entropy of discrete random variable&amp;lt;/strong&amp;gt;] Let [math]X[/math] be a discrete random variable with range of values [math][N] = \{1,2,\ldots,N\}[/math] and probability mass function [math]p[/math]. Define [math]H(X) = -\sum_{n \ge 1} p(n) \log p(n)[/math] with convention [math]0\log 0 = 0[/math]. Prove that [math]H(X) \le \log N[/math] using Jensen&#039;s inequality.&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Law of total expectation&amp;lt;/strong&amp;gt;] Let [math]X \sim \mathrm{Geom}(p)[/math] for some parameter [math]p \in (0,1)[/math]. Calculate [math]\mathbf{E}[X][/math] using the law of total expectation.&lt;br /&gt;
&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Random number of random variables&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\{X_n\}_{n \ge 1}&amp;lt;/math&amp;gt; be identically distributed random variable and &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; be a random variable taking values in the non-negative integers and independent of the &amp;lt;math&amp;gt;X_n&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;n \ge 1&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{E}\left[\sum_{i=1}^N X_i\right] = \mathbf{E}[N] \mathbf{E}[X_1]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Distribution of random variable, 8 points) ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Cumulative distribution function (CDF)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a random variable with cumulative distribution function &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Show that &amp;lt;math&amp;gt;Y = aX+b&amp;lt;/math&amp;gt; is a random variable where &amp;lt;math&amp;gt;a&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;b&amp;lt;/math&amp;gt; are real constants, and express the CDF of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; by &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;. (Hint: Try expressing the event [math]Y=aX+b\le y[/math] by countably many set operations on the events defined on [math]X[/math].)&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Let &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; be the CDF of random variable &amp;lt;math&amp;gt;Z:\Omega\rightarrow \mathbb{R}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;0\leq \lambda \leq 1&amp;lt;/math&amp;gt;, show that &lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;math&amp;gt;\lambda F + (1-\lambda)G&amp;lt;/math&amp;gt; is a CDF function. &amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;The product &amp;lt;math&amp;gt;FG&amp;lt;/math&amp;gt; is a CDF function, and if &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; are independent, then &amp;lt;math&amp;gt;FG&amp;lt;/math&amp;gt; is the CDF of &amp;lt;math&amp;gt;\max\{X,Z\}&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Probability mass function (PMF)&amp;lt;/strong&amp;gt;] We toss &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; coins, and each one shows heads with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;, independently of each of the others. Each coin which shows head is tossed again. (If the coin shows tail, it won&#039;t be tossed again.) Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of heads resulting from the &amp;lt;strong&amp;gt;second&amp;lt;/strong&amp;gt; round of tosses, and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the number of heads resulting from &amp;lt;strong&amp;gt;all&amp;lt;/strong&amp;gt; tosses, which includes the first and (possible) second round of each toss.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the PMF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{E}[Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Let &amp;lt;math&amp;gt;p_X&amp;lt;/math&amp;gt; be the PMF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, show that [math]p_X(k-1)p_X(k+1)\leq [p_X(k)]^2[/math] for &amp;lt;math&amp;gt;1\leq k \leq n-1&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Discrete random variable) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (I)&amp;lt;/strong&amp;gt;] Every package of some intrinsically dull commodity includes a small and exciting plastic object. There are [math]c[/math] different types of object, and each package is equally likely to contain any given type. You buy one package each day.&amp;lt;/p&amp;gt;&lt;br /&gt;
    &amp;lt;ol&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse between the acquisitions of the [math]j[/math]-th new type of object and the [math](j + 1)[/math]-th new type.&amp;lt;/li&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse before you have a full set of objects.&amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (II)&amp;lt;/strong&amp;gt;] Prove that geometric distribution is the only discrete memoryless distribution with range values &amp;lt;math&amp;gt;\mathbb{N}_+&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;n_1,n_2 \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;0 \le p \le 1&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Bin}(n_1,p),Y \sim \mathrm{Bin}(n_2,p)&amp;lt;/math&amp;gt; be independent random variables. Prove that &amp;lt;math&amp;gt;X+Y \sim \mathrm{Bin}(n_1+n_2,p)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Negative binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; follows the negative binomial distribution with parameter &amp;lt;math&amp;gt;r \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt;. Calculate &amp;lt;math&amp;gt;\mathbf{Var}[X] = \mathbf{E}[X^2] - \left(\mathbf{E}[X]\right)^2&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Hypergeometric distribution&amp;lt;/strong&amp;gt;] An urn contains [math]N[/math] balls, [math]b[/math] of which are blue and [math]r = N -b[/math] of which are red. A random sample of [math]n[/math] balls is drawn &amp;lt;strong&amp;gt;without replacement&amp;lt;/strong&amp;gt; (无放回) from the urn. Let [math]B[/math] the number of blue balls in this sample. Show that if [math]N, b[/math], and [math]r[/math] approach [math]+\infty[/math] in such a way that [math]b/N \rightarrow  p[/math] and [math]r/N \rightarrow 1 - p[/math], then [math]\mathbf{Pr}(B = k) \rightarrow {n\choose k}p^k(1-p)^{n-k}[/math] for [math]0\leq k \leq n[/math].&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Poisson distribution&amp;lt;/strong&amp;gt;] In your pocket is a random number &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; of coins, where &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; has the Poisson distribution with parameter &amp;lt;math&amp;gt;\lambda&amp;lt;/math&amp;gt;. You toss each coin once, with heads showing with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; each time. Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the (random) number of heads outcomes and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the (also random) number of tails.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the joint mass function of &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find PMF of the marginal distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; in &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;. Are &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; independent?&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Conditional distribution (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be independent &amp;lt;math&amp;gt;\text{Bin}(n, p)&amp;lt;/math&amp;gt; random variables, and let &amp;lt;math&amp;gt;Z = X + Y&amp;lt;/math&amp;gt;. Show that the conditional distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; given &amp;lt;math&amp;gt;Z = N&amp;lt;/math&amp;gt; is the hypergeometric distribution.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Conditional distribution (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;\lambda,\mu &amp;gt; 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;n \in \mathbb{N}&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Pois}(\lambda), Y \sim \mathrm{Pois}(\mu)&amp;lt;/math&amp;gt; be independent random variables. Find out the conditional distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, given &amp;lt;math&amp;gt;X+Y = n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 4 (Linearity of Expectation, 12 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Streak&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Suppose we flip a fair coin &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; times independently to obtain a sequence of flips &amp;lt;math&amp;gt;X_1, X_2, \ldots , X_n&amp;lt;/math&amp;gt;. A streak of flips is a consecutive subsequence of flips that are all the same. For example, if &amp;lt;math&amp;gt;X_3&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;X_4&amp;lt;/math&amp;gt;, and &amp;lt;math&amp;gt;X_5&amp;lt;/math&amp;gt; are all heads, there is a streak of length &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; starting at the third lip. (If &amp;lt;math&amp;gt;X_6&amp;lt;/math&amp;gt; is also heads, then there is also a streak of length &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; starting at the third lip.) Find the expected number of streaks of length &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; for some integer &amp;lt;math&amp;gt;k \ge 1&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Number of cycles&amp;lt;/strong&amp;gt;] &lt;br /&gt;
At a banquet, there are &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people who shake hands according to the following process: In each round, two idle hands are randomly selected and shaken (&amp;lt;strong&amp;gt;these two hands are no longer idle&amp;lt;/strong&amp;gt;). After &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds, there will be no idle hands left, and the &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people will form several cycles. For example, when &amp;lt;math&amp;gt;n=3&amp;lt;/math&amp;gt;, the following situation may occur: the left and right hands of the first person are held together, the left hand of the second person and the right hand of the third person are held together, and the right hand of the second person and the left hand of the third person are held together. In this case, three people form two cycles. How many cycles are expected to be formed after &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds?&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Expected Mex&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X_1,X_2,\ldots,X_{100} \sim \mathrm{Geo}(1/2)&amp;lt;/math&amp;gt; be independent random variables. Compute &amp;lt;math&amp;gt;\mathbf{E}[\mathrm{mex}(X_1,X_2,\ldots,X_{100})]&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\mathrm{mex}(a_1,a_2,\ldots,a_n)&amp;lt;/math&amp;gt; is the smallest positive integer that does not appear in &amp;lt;math&amp;gt;a_1,a_2,\ldots,a_n&amp;lt;/math&amp;gt;. Your answer is considered correct if the absolute error does not exceed &amp;lt;math&amp;gt;10^{-6}&amp;lt;/math&amp;gt;. (Hint: Although mathematics will help you arrive at elegant and efficient methods, the use of a computer and programming skills will be required.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 5 (Probability meets graph theory) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random social networks&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G = (V, E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph without isolating vertex. &lt;br /&gt;
Let &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; be the degree of vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Let &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be a uniformly chosen vertex, and &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; a uniformly chosen neighbor of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Show that &amp;lt;math&amp;gt;\mathbf{E}[d_Z] \geq \mathbf{E}[d_Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Interpret this inequality in the context of social networks, in which the vertices represent people, and the edges represent friendship.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Turán&#039;s Theorem&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G=(V,E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph, and write &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; for the degree of the vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Use probablistic method to prove that &amp;lt;math&amp;gt;\alpha(G) \ge \sum_{v \in V} \frac{1}{d_v + 1}&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\alpha(G)&amp;lt;/math&amp;gt; is the size of a maximum independent set. (Hint: Consider the following random procedure for generating an independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; from a graph with vertex set &amp;lt;math&amp;gt;V&amp;lt;/math&amp;gt;: First, generate a random permutation of the vertices, denoted as &amp;lt;math&amp;gt;v_1,v_2,\ldots,v_n&amp;lt;/math&amp;gt;. Then, construct the independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; as follows: For each vertex &amp;lt;math&amp;gt;v_i \in V&amp;lt;/math&amp;gt;, add &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; if and only if none of its predecessors in the permutation, i.e., &amp;lt;math&amp;gt;v_1,\ldots,v_{i-1}&amp;lt;/math&amp;gt;, are neighbors of &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Dominating set&amp;lt;/strong&amp;gt;]&lt;br /&gt;
A &#039;&#039;dominating set&#039;&#039; of vertices in an undirected graph &amp;lt;math&amp;gt;G = (V, E)&amp;lt;/math&amp;gt; is a set &amp;lt;math&amp;gt;S \subseteq V&amp;lt;/math&amp;gt; such that every vertex of&lt;br /&gt;
&amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; belongs to &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; or has a neighbor in &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;. &lt;br /&gt;
&lt;br /&gt;
Let &amp;lt;math&amp;gt;G = (V, E)&amp;lt;/math&amp;gt; be an &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;-vertex graph with minimum degree &amp;lt;math&amp;gt;d &amp;gt; 1&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; has a dominating set with at most &amp;lt;math&amp;gt;\frac{n\left(1+\log(d+1)\right)}{d+1}&amp;lt;/math&amp;gt; vertices. (Hint: Consider a random vertex subset &amp;lt;math&amp;gt;S \subseteq V&amp;lt;/math&amp;gt; by including each vertex independently with&lt;br /&gt;
probability &amp;lt;math&amp;gt;p := \log(d + 1)/(d + 1)&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 6 (1D random walk, 8 points) ==&lt;br /&gt;
Let &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt; be a constant, and &amp;lt;math&amp;gt;\{X_n\}_{n \ge 1}&amp;lt;/math&amp;gt; be independent Bernoulli trials with successful probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;.&lt;br /&gt;
Define &amp;lt;math&amp;gt;S_n = 2\sum_{i=1}^n X_i - n&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;S_0 = 0&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Range of random walk&amp;lt;/strong&amp;gt;] The range &amp;lt;math&amp;gt;R_n&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;S_0, S_1, \ldots, S_n&amp;lt;/math&amp;gt; is defined as the number of distinct values taken by the sequence. Show that &amp;lt;math&amp;gt;\mathbf{Pr}\left(R_n = R_{n-1}+1\right) = \mathbf{Pr}\left(\forall 1 \le i \le n, S_i \neq 0\right)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, and deduce that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n]\to \mathbf{Pr}(\forall i \ge 1, S_i \neq 0)&amp;lt;/math&amp;gt;. Hence show that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n] \to |2p-1|&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Symmetric 1D random walk (III)&amp;lt;/strong&amp;gt;] Suppose &amp;lt;math&amp;gt;p = \frac{1}{2}&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{E}[|S_n|] = \Theta(\sqrt{n})&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12949</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 2</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12949"/>
		<updated>2025-03-13T15:09:14Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: /* Problem 1 (Warm-up problems, 12 points) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 2==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Warm-up problems, 12 points) ==&lt;br /&gt;
* [&#039;&#039;&#039;Function of random variable (I)&#039;&#039;&#039;] Let [math]X[/math] be a random variable with distribution function [math]\max(0,\min(1,x))[/math]. Let [math]F[/math] be a distribution function which is continuous and strictly increasing. Show that [math]Y=F^{-1}(X)[/math] be a random variable with distribution function [math]F[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Function of random variable (II)&#039;&#039;&#039;] Express the distribution functions of [math]X^+ = \max\{0,X\}[/math], [math]X^- = -\min\{0,X\}[/math], [math]|X|=X^+ + X^-[/math], [math]-X[/math], in terms of the distribution function [math]F[/math] of the random variable [math]X[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Independence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;1\leq r\leq n&amp;lt;/math&amp;gt; be independent random variables which are symmetric about &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;; that is, &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-X_r&amp;lt;/math&amp;gt; have the same distributions. Show that, for all &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}[S_n \geq x] = \mathbf{Pr}[S_n \leq -x]&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;S_n = \sum_{r=1}^n X_r&amp;lt;/math&amp;gt;. Is the conclusion true without the assumtion of independence?&lt;br /&gt;
* [&#039;&#039;&#039;Dependence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be discrete random variables with joint mass function &amp;lt;math&amp;gt;f(x,y) = \frac{C}{(x+y-1)(x+y)(x+y+1)}&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;x,y \in \mathbb{N}_+&amp;lt;/math&amp;gt; (in other words, &amp;lt;math&amp;gt;x,y = 1,2,3,\cdots&amp;lt;/math&amp;gt;). Find (1) the value of &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;, (2) marginal mass function of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and (3) &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt;.&lt;br /&gt;
* [&#039;&#039;&#039;Expectation&#039;&#039;&#039;] Is it generally true that [math]\mathbf{E}[1/X] = 1/\mathbf{E}[X][/math]? Is it ever true that [math]\mathbf{E}[1/X] = 1/\mathbf{E}[X][/math]?&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Entropy of discrete random variable&amp;lt;/strong&amp;gt;] Let [math]X[/math] be a discrete random variable with range of values [math][N] = \{1,2,\ldots,N\}[/math] and probability mass function [math]p[/math]. Define [math]H(X) = -\sum_{n \ge 1} p(n) \log p(n)[/math] with convention [math]0\log 0 = 0[/math]. Prove that [math]H(X) \le \log N[/math] using Jensen&#039;s inequality.&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Law of total expectation&amp;lt;/strong&amp;gt;] Let [math]X \sim \mathrm{Geom}(p)[/math] for some parameter [math]p \in (0,1)[/math]. Calculate [math]\mathbf{E}[X][/math] using the law of total expectation.&lt;br /&gt;
&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Random number of random variables&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\{X_n\}_{n \ge 1}&amp;lt;/math&amp;gt; be identically distributed random variable and &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; be a random variable taking values in the non-negative integers and independent of the &amp;lt;math&amp;gt;X_n&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;n \ge 1&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{E}\left[\sum_{i=1}^N X_i\right] = \mathbf{E}[N] \mathbf{E}[X_1]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Distribution of random variable, 8 points) ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Cumulative distribution function (CDF)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a random variable with cumulative distribution function &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Show that &amp;lt;math&amp;gt;Y = aX+b&amp;lt;/math&amp;gt; is a random variable where &amp;lt;math&amp;gt;a&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;b&amp;lt;/math&amp;gt; are real constants, and express the CDF of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; by &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;. (Hint: Try expressing the event [math]Y=aX+b\le y[/math] by countably many set operations on the events defined on [math]X[/math].)&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Let &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; be the CDF of random variable &amp;lt;math&amp;gt;Z:\Omega\rightarrow \mathbb{R}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;0\leq \lambda \leq 1&amp;lt;/math&amp;gt;, show that &lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;math&amp;gt;\lambda F + (1-\lambda)G&amp;lt;/math&amp;gt; is a CDF function. &amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;The product &amp;lt;math&amp;gt;FG&amp;lt;/math&amp;gt; is a CDF function, and if &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; are independent, then &amp;lt;math&amp;gt;FG&amp;lt;/math&amp;gt; is the CDF of &amp;lt;math&amp;gt;\max\{X,Z\}&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Probability mass function (PMF)&amp;lt;/strong&amp;gt;] We toss &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; coins, and each one shows heads with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;, independently of each of the others. Each coin which shows head is tossed again. (If the coin shows tail, it won&#039;t be tossed again.) Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of heads resulting from the &amp;lt;strong&amp;gt;second&amp;lt;/strong&amp;gt; round of tosses, and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the number of heads resulting from &amp;lt;strong&amp;gt;all&amp;lt;/strong&amp;gt; tosses, which includes the first and (possible) second round of each toss.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the PMF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{E}[Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Let &amp;lt;math&amp;gt;p_X&amp;lt;/math&amp;gt; be the PMF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, show that [math]p_X(k-1)p_X(k+1)\leq [p_X(k)]^2[/math] for &amp;lt;math&amp;gt;1\leq k \leq n-1&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Discrete random variable) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (I)&amp;lt;/strong&amp;gt;] Every package of some intrinsically dull commodity includes a small and exciting plastic object. There are [math]c[/math] different types of object, and each package is equally likely to contain any given type. You buy one package each day.&amp;lt;/p&amp;gt;&lt;br /&gt;
    &amp;lt;ol&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse between the acquisitions of the [math]j[/math]-th new type of object and the [math](j + 1)[/math]-th new type.&amp;lt;/li&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse before you have a full set of objects.&amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (II)&amp;lt;/strong&amp;gt;] Prove that geometric distribution is the only discrete memoryless distribution with range values &amp;lt;math&amp;gt;\mathbb{N}_+&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;n_1,n_2 \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;0 \le p \le 1&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Bin}(n_1,p),Y \sim \mathrm{Bin}(n_2,p)&amp;lt;/math&amp;gt; be independent random variables. Prove that &amp;lt;math&amp;gt;X+Y \sim \mathrm{Bin}(n_1+n_2,p)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Negative binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; follows the negative binomial distribution with parameter &amp;lt;math&amp;gt;r \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt;. Calculate &amp;lt;math&amp;gt;\mathbf{Var}[X] = \mathbf{E}[X^2] - \left(\mathbf{E}[X]\right)^2&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Hypergeometric distribution&amp;lt;/strong&amp;gt;] An urn contains [math]N[/math] balls, [math]b[/math] of which are blue and [math]r = N -b[/math] of which are red. A random sample of [math]n[/math] balls is drawn &amp;lt;strong&amp;gt;without replacement&amp;lt;/strong&amp;gt; (无放回) from the urn. Let [math]B[/math] the number of blue balls in this sample. Show that if [math]N, b[/math], and [math]r[/math] approach [math]+\infty[/math] in such a way that [math]b/N \rightarrow  p[/math] and [math]r/N \rightarrow 1 - p[/math], then [math]\mathbf{Pr}(B = k) \rightarrow {n\choose k}p^k(1-p)^{n-k}[/math] for [math]0\leq k \leq n[/math].&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Poisson distribution&amp;lt;/strong&amp;gt;] In your pocket is a random number &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; of coins, where &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; has the Poisson distribution with parameter &amp;lt;math&amp;gt;\lambda&amp;lt;/math&amp;gt;. You toss each coin once, with heads showing with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; each time. Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the (random) number of heads outcomes and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the (also random) number of tails.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the joint mass function of &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find PMF of the marginal distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; in &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;. Are &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; independent?&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Conditional distribution (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be independent &amp;lt;math&amp;gt;\text{Bin}(n, p)&amp;lt;/math&amp;gt; random variables, and let &amp;lt;math&amp;gt;Z = X + Y&amp;lt;/math&amp;gt;. Show that the conditional distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; given &amp;lt;math&amp;gt;Z = N&amp;lt;/math&amp;gt; is the hypergeometric distribution.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Conditional distribution (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;\lambda,\mu &amp;gt; 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;n \in \mathbb{N}&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Pois}(\lambda), Y \sim \mathrm{Pois}(\mu)&amp;lt;/math&amp;gt; be independent random variables. Find out the conditional distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, given &amp;lt;math&amp;gt;X+Y = n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 4 (Linearity of Expectation, 12 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Streak&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Suppose we flip a fair coin &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; times independently to obtain a sequence of flips &amp;lt;math&amp;gt;X_1, X_2, \ldots , X_n&amp;lt;/math&amp;gt;. A streak of flips is a consecutive subsequence of flips that are all the same. For example, if &amp;lt;math&amp;gt;X_3&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;X_4&amp;lt;/math&amp;gt;, and &amp;lt;math&amp;gt;X_5&amp;lt;/math&amp;gt; are all heads, there is a streak of length &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; starting at the third lip. (If &amp;lt;math&amp;gt;X_6&amp;lt;/math&amp;gt; is also heads, then there is also a streak of length &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; starting at the third lip.) Find the expected number of streaks of length &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; for some integer &amp;lt;math&amp;gt;k \ge 1&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Number of cycles&amp;lt;/strong&amp;gt;] &lt;br /&gt;
At a banquet, there are &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people who shake hands according to the following process: In each round, two idle hands are randomly selected and shaken (&amp;lt;strong&amp;gt;these two hands are no longer idle&amp;lt;/strong&amp;gt;). After &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds, there will be no idle hands left, and the &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people will form several cycles. For example, when &amp;lt;math&amp;gt;n=3&amp;lt;/math&amp;gt;, the following situation may occur: the left and right hands of the first person are held together, the left hand of the second person and the right hand of the third person are held together, and the right hand of the second person and the left hand of the third person are held together. In this case, three people form two cycles. How many cycles are expected to be formed after &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds?&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Expected Mex&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X_1,X_2,\ldots,X_{100} \sim \mathrm{Geo}(1/2)&amp;lt;/math&amp;gt; be independent random variables. Compute &amp;lt;math&amp;gt;\mathbf{E}[\mathrm{mex}(X_1,X_2,\ldots,X_{100})]&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\mathrm{mex}(a_1,a_2,\ldots,a_n)&amp;lt;/math&amp;gt; is the smallest positive integer that does not appear in &amp;lt;math&amp;gt;a_1,a_2,\ldots,a_n&amp;lt;/math&amp;gt;. Your answer is considered correct if the absolute error does not exceed &amp;lt;math&amp;gt;10^{-6}&amp;lt;/math&amp;gt;. (Hint: Although mathematics will help you arrive at elegant and efficient methods, the use of a computer and programming skills will be required.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 5 (Probability meets graph theory) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random social networks&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G = (V, E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph without isolating vertex. &lt;br /&gt;
Let &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; be the degree of vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Let &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be a uniformly chosen vertex, and &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; a uniformly chosen neighbor of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Show that &amp;lt;math&amp;gt;\mathbf{E}[d_Z] \geq \mathbf{E}[d_Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Interpret this inequality in the context of social networks, in which the vertices represent people, and the edges represent friendship.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Turán&#039;s Theorem&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G=(V,E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph, and write &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; for the degree of the vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Use probablistic method to prove that &amp;lt;math&amp;gt;\alpha(G) \ge \sum_{v \in V} \frac{1}{d_v + 1}&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\alpha(G)&amp;lt;/math&amp;gt; is the size of a maximum independent set. (Hint: Consider the following random procedure for generating an independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; from a graph with vertex set &amp;lt;math&amp;gt;V&amp;lt;/math&amp;gt;: First, generate a random permutation of the vertices, denoted as &amp;lt;math&amp;gt;v_1,v_2,\ldots,v_n&amp;lt;/math&amp;gt;. Then, construct the independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; as follows: For each vertex &amp;lt;math&amp;gt;v_i \in V&amp;lt;/math&amp;gt;, add &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; if and only if none of its predecessors in the permutation, i.e., &amp;lt;math&amp;gt;v_1,\ldots,v_{i-1}&amp;lt;/math&amp;gt;, are neighbors of &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Dominating set&amp;lt;/strong&amp;gt;]&lt;br /&gt;
A &#039;&#039;dominating set&#039;&#039; of vertices in an undirected graph &amp;lt;math&amp;gt;G = (V, E)&amp;lt;/math&amp;gt; is a set &amp;lt;math&amp;gt;S \subseteq V&amp;lt;/math&amp;gt; such that every vertex of&lt;br /&gt;
&amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; belongs to &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; or has a neighbor in &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;. &lt;br /&gt;
&lt;br /&gt;
Let &amp;lt;math&amp;gt;G = (V, E)&amp;lt;/math&amp;gt; be an &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;-vertex graph with minimum degree &amp;lt;math&amp;gt;d &amp;gt; 1&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; has a dominating set with at most &amp;lt;math&amp;gt;\frac{n\left(1+\log(d+1)\right)}{d+1}&amp;lt;/math&amp;gt; vertices. (Hint: Consider a random vertex subset &amp;lt;math&amp;gt;S \subseteq V&amp;lt;/math&amp;gt; by including each vertex independently with&lt;br /&gt;
probability &amp;lt;math&amp;gt;p := \log(d + 1)/(d + 1)&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 6 (1D random walk, 8 points) ==&lt;br /&gt;
Let &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt; be a constant, and &amp;lt;math&amp;gt;\{X_n\}_{n \ge 1}&amp;lt;/math&amp;gt; be independent Bernoulli trials with successful probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;.&lt;br /&gt;
Define &amp;lt;math&amp;gt;S_n = 2\sum_{i=1}^n X_i - n&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;S_0 = 0&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Range of random walk&amp;lt;/strong&amp;gt;] The range &amp;lt;math&amp;gt;R_n&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;S_0, S_1, \ldots, S_n&amp;lt;/math&amp;gt; is defined as the number of distinct values taken by the sequence. Show that &amp;lt;math&amp;gt;\mathbf{Pr}\left(R_n = R_{n-1}+1\right) = \mathbf{Pr}\left(\forall 1 \le i \le n, S_i \neq 0\right)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, and deduce that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n]\to \mathbf{Pr}(\forall i \ge 1, S_i \neq 0)&amp;lt;/math&amp;gt;. Hence show that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n] \to |2p-1|&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Symmetric 1D random walk (III)&amp;lt;/strong&amp;gt;] Suppose &amp;lt;math&amp;gt;p = \frac{1}{2}&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{E}[|S_n|] = \Theta(\sqrt{n})&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12948</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 2</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_2&amp;diff=12948"/>
		<updated>2025-03-13T15:06:23Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: /* Problem 1 (Warm-up problems, 12 points) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 2==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Warm-up problems, 12 points) ==&lt;br /&gt;
* [&#039;&#039;&#039;Function of random variable (I)&#039;&#039;&#039;] Let [math]X[/math] be a random variable with distribution function [math]\max(0,\min(1,x))[/math]. Let [math]F[/math] be a distribution function which is continuous and strictly increasing. Show that [math]Y=F^{-1}(X)[/math] be a random variable with distribution function [math]F[/math].&lt;br /&gt;
* [&#039;&#039;&#039;Function of random variable (II)&#039;&#039;&#039;] Express the distribution functions of [math]X^+ = \max\{0,X\}[/math]&lt;br /&gt;
* [&#039;&#039;&#039;Independence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;1\leq r\leq n&amp;lt;/math&amp;gt; be independent random variables which are symmetric about &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt;; that is, &amp;lt;math&amp;gt;X_r&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-X_r&amp;lt;/math&amp;gt; have the same distributions. Show that, for all &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{Pr}[S_n \geq x] = \mathbf{Pr}[S_n \leq -x]&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;S_n = \sum_{r=1}^n X_r&amp;lt;/math&amp;gt;. Is the conclusion true without the assumtion of independence?&lt;br /&gt;
* [&#039;&#039;&#039;Dependence&#039;&#039;&#039;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be discrete random variables with joint mass function &amp;lt;math&amp;gt;f(x,y) = \frac{C}{(x+y-1)(x+y)(x+y+1)}&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;x,y \in \mathbb{N}_+&amp;lt;/math&amp;gt; (in other words, &amp;lt;math&amp;gt;x,y = 1,2,3,\cdots&amp;lt;/math&amp;gt;). Find (1) the value of &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;, (2) marginal mass function of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and (3) &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt;.&lt;br /&gt;
* [&#039;&#039;&#039;Expectation&#039;&#039;&#039;] Is it generally true that [math]\mathbf{E}[1/X] = 1/\mathbf{E}[X][/math]? Is it ever true that [math]\mathbf{E}[1/X] = 1/\mathbf{E}[X][/math]?&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Entropy of discrete random variable&amp;lt;/strong&amp;gt;] Let [math]X[/math] be a discrete random variable with range of values [math][N] = \{1,2,\ldots,N\}[/math] and probability mass function [math]p[/math]. Define [math]H(X) = -\sum_{n \ge 1} p(n) \log p(n)[/math] with convention [math]0\log 0 = 0[/math]. Prove that [math]H(X) \le \log N[/math] using Jensen&#039;s inequality.&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Law of total expectation&amp;lt;/strong&amp;gt;] Let [math]X \sim \mathrm{Geom}(p)[/math] for some parameter [math]p \in (0,1)[/math]. Calculate [math]\mathbf{E}[X][/math] using the law of total expectation.&lt;br /&gt;
&lt;br /&gt;
* [&amp;lt;strong&amp;gt;Random number of random variables&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\{X_n\}_{n \ge 1}&amp;lt;/math&amp;gt; be identically distributed random variable and &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; be a random variable taking values in the non-negative integers and independent of the &amp;lt;math&amp;gt;X_n&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;n \ge 1&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{E}\left[\sum_{i=1}^N X_i\right] = \mathbf{E}[N] \mathbf{E}[X_1]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Distribution of random variable, 8 points) ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Cumulative distribution function (CDF)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a random variable with cumulative distribution function &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Show that &amp;lt;math&amp;gt;Y = aX+b&amp;lt;/math&amp;gt; is a random variable where &amp;lt;math&amp;gt;a&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;b&amp;lt;/math&amp;gt; are real constants, and express the CDF of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; by &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;. (Hint: Try expressing the event [math]Y=aX+b\le y[/math] by countably many set operations on the events defined on [math]X[/math].)&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Let &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; be the CDF of random variable &amp;lt;math&amp;gt;Z:\Omega\rightarrow \mathbb{R}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;0\leq \lambda \leq 1&amp;lt;/math&amp;gt;, show that &lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;math&amp;gt;\lambda F + (1-\lambda)G&amp;lt;/math&amp;gt; is a CDF function. &amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;The product &amp;lt;math&amp;gt;FG&amp;lt;/math&amp;gt; is a CDF function, and if &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; are independent, then &amp;lt;math&amp;gt;FG&amp;lt;/math&amp;gt; is the CDF of &amp;lt;math&amp;gt;\max\{X,Z\}&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Probability mass function (PMF)&amp;lt;/strong&amp;gt;] We toss &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; coins, and each one shows heads with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;, independently of each of the others. Each coin which shows head is tossed again. (If the coin shows tail, it won&#039;t be tossed again.) Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of heads resulting from the &amp;lt;strong&amp;gt;second&amp;lt;/strong&amp;gt; round of tosses, and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the number of heads resulting from &amp;lt;strong&amp;gt;all&amp;lt;/strong&amp;gt; tosses, which includes the first and (possible) second round of each toss.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the PMF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find &amp;lt;math&amp;gt;\mathbf{E}[X]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{E}[Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Let &amp;lt;math&amp;gt;p_X&amp;lt;/math&amp;gt; be the PMF of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, show that [math]p_X(k-1)p_X(k+1)\leq [p_X(k)]^2[/math] for &amp;lt;math&amp;gt;1\leq k \leq n-1&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Discrete random variable) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (I)&amp;lt;/strong&amp;gt;] Every package of some intrinsically dull commodity includes a small and exciting plastic object. There are [math]c[/math] different types of object, and each package is equally likely to contain any given type. You buy one package each day.&amp;lt;/p&amp;gt;&lt;br /&gt;
    &amp;lt;ol&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse between the acquisitions of the [math]j[/math]-th new type of object and the [math](j + 1)[/math]-th new type.&amp;lt;/li&amp;gt;&lt;br /&gt;
        &amp;lt;li&amp;gt;Find the expected number of days which elapse before you have a full set of objects.&amp;lt;/li&amp;gt;&lt;br /&gt;
    &amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Geometric distribution (II)&amp;lt;/strong&amp;gt;] Prove that geometric distribution is the only discrete memoryless distribution with range values &amp;lt;math&amp;gt;\mathbb{N}_+&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;n_1,n_2 \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;0 \le p \le 1&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Bin}(n_1,p),Y \sim \mathrm{Bin}(n_2,p)&amp;lt;/math&amp;gt; be independent random variables. Prove that &amp;lt;math&amp;gt;X+Y \sim \mathrm{Bin}(n_1+n_2,p)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Negative binomial distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; follows the negative binomial distribution with parameter &amp;lt;math&amp;gt;r \in \mathbb{N}_+&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt;. Calculate &amp;lt;math&amp;gt;\mathbf{Var}[X] = \mathbf{E}[X^2] - \left(\mathbf{E}[X]\right)^2&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Hypergeometric distribution&amp;lt;/strong&amp;gt;] An urn contains [math]N[/math] balls, [math]b[/math] of which are blue and [math]r = N -b[/math] of which are red. A random sample of [math]n[/math] balls is drawn &amp;lt;strong&amp;gt;without replacement&amp;lt;/strong&amp;gt; (无放回) from the urn. Let [math]B[/math] the number of blue balls in this sample. Show that if [math]N, b[/math], and [math]r[/math] approach [math]+\infty[/math] in such a way that [math]b/N \rightarrow  p[/math] and [math]r/N \rightarrow 1 - p[/math], then [math]\mathbf{Pr}(B = k) \rightarrow {n\choose k}p^k(1-p)^{n-k}[/math] for [math]0\leq k \leq n[/math].&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Poisson distribution&amp;lt;/strong&amp;gt;] In your pocket is a random number &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; of coins, where &amp;lt;math&amp;gt;N&amp;lt;/math&amp;gt; has the Poisson distribution with parameter &amp;lt;math&amp;gt;\lambda&amp;lt;/math&amp;gt;. You toss each coin once, with heads showing with probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; each time. Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the (random) number of heads outcomes and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be the (also random) number of tails.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find the joint mass function of &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Find PMF of the marginal distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; in &amp;lt;math&amp;gt;(X,Y)&amp;lt;/math&amp;gt;. Are &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; independent?&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Conditional distribution (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be independent &amp;lt;math&amp;gt;\text{Bin}(n, p)&amp;lt;/math&amp;gt; random variables, and let &amp;lt;math&amp;gt;Z = X + Y&amp;lt;/math&amp;gt;. Show that the conditional distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; given &amp;lt;math&amp;gt;Z = N&amp;lt;/math&amp;gt; is the hypergeometric distribution.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Conditional distribution (II)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;\lambda,\mu &amp;gt; 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;n \in \mathbb{N}&amp;lt;/math&amp;gt; be parameters, and &amp;lt;math&amp;gt;X \sim \mathrm{Pois}(\lambda), Y \sim \mathrm{Pois}(\mu)&amp;lt;/math&amp;gt; be independent random variables. Find out the conditional distribution of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;, given &amp;lt;math&amp;gt;X+Y = n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 4 (Linearity of Expectation, 12 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Streak&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Suppose we flip a fair coin &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; times independently to obtain a sequence of flips &amp;lt;math&amp;gt;X_1, X_2, \ldots , X_n&amp;lt;/math&amp;gt;. A streak of flips is a consecutive subsequence of flips that are all the same. For example, if &amp;lt;math&amp;gt;X_3&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;X_4&amp;lt;/math&amp;gt;, and &amp;lt;math&amp;gt;X_5&amp;lt;/math&amp;gt; are all heads, there is a streak of length &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; starting at the third lip. (If &amp;lt;math&amp;gt;X_6&amp;lt;/math&amp;gt; is also heads, then there is also a streak of length &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; starting at the third lip.) Find the expected number of streaks of length &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; for some integer &amp;lt;math&amp;gt;k \ge 1&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Number of cycles&amp;lt;/strong&amp;gt;] &lt;br /&gt;
At a banquet, there are &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people who shake hands according to the following process: In each round, two idle hands are randomly selected and shaken (&amp;lt;strong&amp;gt;these two hands are no longer idle&amp;lt;/strong&amp;gt;). After &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds, there will be no idle hands left, and the &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; people will form several cycles. For example, when &amp;lt;math&amp;gt;n=3&amp;lt;/math&amp;gt;, the following situation may occur: the left and right hands of the first person are held together, the left hand of the second person and the right hand of the third person are held together, and the right hand of the second person and the left hand of the third person are held together. In this case, three people form two cycles. How many cycles are expected to be formed after &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds?&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Expected Mex&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X_1,X_2,\ldots,X_{100} \sim \mathrm{Geo}(1/2)&amp;lt;/math&amp;gt; be independent random variables. Compute &amp;lt;math&amp;gt;\mathbf{E}[\mathrm{mex}(X_1,X_2,\ldots,X_{100})]&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\mathrm{mex}(a_1,a_2,\ldots,a_n)&amp;lt;/math&amp;gt; is the smallest positive integer that does not appear in &amp;lt;math&amp;gt;a_1,a_2,\ldots,a_n&amp;lt;/math&amp;gt;. Your answer is considered correct if the absolute error does not exceed &amp;lt;math&amp;gt;10^{-6}&amp;lt;/math&amp;gt;. (Hint: Although mathematics will help you arrive at elegant and efficient methods, the use of a computer and programming skills will be required.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 5 (Probability meets graph theory) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random social networks&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G = (V, E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph without isolating vertex. &lt;br /&gt;
Let &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; be the degree of vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Let &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be a uniformly chosen vertex, and &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; a uniformly chosen neighbor of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Show that &amp;lt;math&amp;gt;\mathbf{E}[d_Z] \geq \mathbf{E}[d_Y]&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Interpret this inequality in the context of social networks, in which the vertices represent people, and the edges represent friendship.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Turán&#039;s Theorem&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;G=(V,E)&amp;lt;/math&amp;gt; be a &amp;lt;strong&amp;gt;fixed&amp;lt;/strong&amp;gt; undirected graph, and write &amp;lt;math&amp;gt;d_v&amp;lt;/math&amp;gt; for the degree of the vertex &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt;. Use probablistic method to prove that &amp;lt;math&amp;gt;\alpha(G) \ge \sum_{v \in V} \frac{1}{d_v + 1}&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\alpha(G)&amp;lt;/math&amp;gt; is the size of a maximum independent set. (Hint: Consider the following random procedure for generating an independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; from a graph with vertex set &amp;lt;math&amp;gt;V&amp;lt;/math&amp;gt;: First, generate a random permutation of the vertices, denoted as &amp;lt;math&amp;gt;v_1,v_2,\ldots,v_n&amp;lt;/math&amp;gt;. Then, construct the independent set &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; as follows: For each vertex &amp;lt;math&amp;gt;v_i \in V&amp;lt;/math&amp;gt;, add &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;I&amp;lt;/math&amp;gt; if and only if none of its predecessors in the permutation, i.e., &amp;lt;math&amp;gt;v_1,\ldots,v_{i-1}&amp;lt;/math&amp;gt;, are neighbors of &amp;lt;math&amp;gt;v_i&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Dominating set&amp;lt;/strong&amp;gt;]&lt;br /&gt;
A &#039;&#039;dominating set&#039;&#039; of vertices in an undirected graph &amp;lt;math&amp;gt;G = (V, E)&amp;lt;/math&amp;gt; is a set &amp;lt;math&amp;gt;S \subseteq V&amp;lt;/math&amp;gt; such that every vertex of&lt;br /&gt;
&amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; belongs to &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; or has a neighbor in &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;. &lt;br /&gt;
&lt;br /&gt;
Let &amp;lt;math&amp;gt;G = (V, E)&amp;lt;/math&amp;gt; be an &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;-vertex graph with minimum degree &amp;lt;math&amp;gt;d &amp;gt; 1&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; has a dominating set with at most &amp;lt;math&amp;gt;\frac{n\left(1+\log(d+1)\right)}{d+1}&amp;lt;/math&amp;gt; vertices. (Hint: Consider a random vertex subset &amp;lt;math&amp;gt;S \subseteq V&amp;lt;/math&amp;gt; by including each vertex independently with&lt;br /&gt;
probability &amp;lt;math&amp;gt;p := \log(d + 1)/(d + 1)&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 6 (1D random walk, 8 points) ==&lt;br /&gt;
Let &amp;lt;math&amp;gt;p \in (0,1)&amp;lt;/math&amp;gt; be a constant, and &amp;lt;math&amp;gt;\{X_n\}_{n \ge 1}&amp;lt;/math&amp;gt; be independent Bernoulli trials with successful probability &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt;.&lt;br /&gt;
Define &amp;lt;math&amp;gt;S_n = 2\sum_{i=1}^n X_i - n&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;S_0 = 0&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Range of random walk&amp;lt;/strong&amp;gt;] The range &amp;lt;math&amp;gt;R_n&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;S_0, S_1, \ldots, S_n&amp;lt;/math&amp;gt; is defined as the number of distinct values taken by the sequence. Show that &amp;lt;math&amp;gt;\mathbf{Pr}\left(R_n = R_{n-1}+1\right) = \mathbf{Pr}\left(\forall 1 \le i \le n, S_i \neq 0\right)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, and deduce that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n]\to \mathbf{Pr}(\forall i \ge 1, S_i \neq 0)&amp;lt;/math&amp;gt;. Hence show that &amp;lt;math&amp;gt;n^{-1} \mathbf{E}[R_n] \to |2p-1|&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Symmetric 1D random walk (III)&amp;lt;/strong&amp;gt;] Suppose &amp;lt;math&amp;gt;p = \frac{1}{2}&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;\mathbf{E}[|S_n|] = \Theta(\sqrt{n})&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/%E7%AC%AC%E4%B8%80%E6%AC%A1%E4%BD%9C%E4%B8%9A%E6%8F%90%E4%BA%A4%E5%90%8D%E5%8D%95&amp;diff=12937</id>
		<title>概率论与数理统计 (Spring 2025)/第一次作业提交名单</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/%E7%AC%AC%E4%B8%80%E6%AC%A1%E4%BD%9C%E4%B8%9A%E6%8F%90%E4%BA%A4%E5%90%8D%E5%8D%95&amp;diff=12937"/>
		<updated>2025-03-10T12:09:00Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: Created page with &amp;quot;如有错漏邮件请及时联系助教。 &amp;lt;center&amp;gt; {| class=&amp;quot;wikitable&amp;quot; |- ! 学号 !! 姓名 |- | 231200035 || 葛翰飞  |- | 231880394 || 翟笑晨  |- | 231840166 || 赵少凡  |- | 231880123 || 陆博文  |- | 231840058 || 王润希  |- | 231240002 || 余孟凡  |- | 231240060 || 侯佳松  |- | 231240055 || 魏全烨  |- | 231098046 || 李方泽  |- | 231240015 || 邱文宣  |- | 231240011 || 陈力琰  |- | 231240038 || 肖金宇  |- | 231230102 || 庄铸锴  |-...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;如有错漏邮件请及时联系助教。&lt;br /&gt;
&amp;lt;center&amp;gt;&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! 学号 !! 姓名&lt;br /&gt;
|-&lt;br /&gt;
| 231200035 || 葛翰飞 &lt;br /&gt;
|-&lt;br /&gt;
| 231880394 || 翟笑晨 &lt;br /&gt;
|-&lt;br /&gt;
| 231840166 || 赵少凡 &lt;br /&gt;
|-&lt;br /&gt;
| 231880123 || 陆博文 &lt;br /&gt;
|-&lt;br /&gt;
| 231840058 || 王润希 &lt;br /&gt;
|-&lt;br /&gt;
| 231240002 || 余孟凡 &lt;br /&gt;
|-&lt;br /&gt;
| 231240060 || 侯佳松 &lt;br /&gt;
|-&lt;br /&gt;
| 231240055 || 魏全烨 &lt;br /&gt;
|-&lt;br /&gt;
| 231098046 || 李方泽 &lt;br /&gt;
|-&lt;br /&gt;
| 231240015 || 邱文宣 &lt;br /&gt;
|-&lt;br /&gt;
| 231240011 || 陈力琰 &lt;br /&gt;
|-&lt;br /&gt;
| 231240038 || 肖金宇 &lt;br /&gt;
|-&lt;br /&gt;
| 231230102 || 庄铸锴 &lt;br /&gt;
|-&lt;br /&gt;
| 231880140 || 桂天麟 &lt;br /&gt;
|-&lt;br /&gt;
| 231880152 || 李佳骏 &lt;br /&gt;
|-&lt;br /&gt;
| 231240047 || 姜淮仁 &lt;br /&gt;
|-&lt;br /&gt;
| 231240018 || 渠翔凯 &lt;br /&gt;
|-&lt;br /&gt;
| 231240029 || 朱非凡 &lt;br /&gt;
|-&lt;br /&gt;
| 231840242 || 李昊凌 &lt;br /&gt;
|-&lt;br /&gt;
| 231240056 || 靳濡搏 &lt;br /&gt;
|-&lt;br /&gt;
| 231240007 || 高亮节 &lt;br /&gt;
|-&lt;br /&gt;
| 231220001 || 陈梓睿 &lt;br /&gt;
|-&lt;br /&gt;
| 231240058 || 陈小川 &lt;br /&gt;
|-&lt;br /&gt;
| 231240019 || 邵与乔 &lt;br /&gt;
|-&lt;br /&gt;
| 231240025 || 李熠辰 &lt;br /&gt;
|-&lt;br /&gt;
| 231240004 || 何梓杨 &lt;br /&gt;
|-&lt;br /&gt;
| 231240051 || 王昱霏 &lt;br /&gt;
|-&lt;br /&gt;
| 231240023 || 叶宸见 &lt;br /&gt;
|-&lt;br /&gt;
| 231240045 || 杨俊炜 &lt;br /&gt;
|-&lt;br /&gt;
| 231240046 || 陈昊奕 &lt;br /&gt;
|-&lt;br /&gt;
| 231220005 || 樊书岩 &lt;br /&gt;
|-&lt;br /&gt;
| 231240013 || 南晨曦 &lt;br /&gt;
|-&lt;br /&gt;
| 231240024 || 陈越 &lt;br /&gt;
|-&lt;br /&gt;
| 231240021 || 陈奕帆 &lt;br /&gt;
|-&lt;br /&gt;
| 231240043 || 朱宇曦 &lt;br /&gt;
|-&lt;br /&gt;
| 231830123 || 刘逸杨 &lt;br /&gt;
|-&lt;br /&gt;
| 231240016 || 徐冰冰 &lt;br /&gt;
|-&lt;br /&gt;
| 231220002 || 潘谟天 &lt;br /&gt;
|-&lt;br /&gt;
| 231240027 || 彭浩楠 &lt;br /&gt;
|-&lt;br /&gt;
| 231240053 || 王艺文 &lt;br /&gt;
|-&lt;br /&gt;
| 231220090 || 邓淇 &lt;br /&gt;
|-&lt;br /&gt;
| 231240009 || 陈心怡 &lt;br /&gt;
|-&lt;br /&gt;
| 211250032 || 吕卉森 &lt;br /&gt;
|-&lt;br /&gt;
| 231240026 || 蒙航 &lt;br /&gt;
|-&lt;br /&gt;
| 231240026 || 蒙航 &lt;br /&gt;
|-&lt;br /&gt;
| 211840320 || 华徐涛 &lt;br /&gt;
|-&lt;br /&gt;
| 231880432 || 向灿 &lt;br /&gt;
|}&lt;br /&gt;
&amp;lt;/center&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)&amp;diff=12936</id>
		<title>概率论与数理统计 (Spring 2025)</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)&amp;diff=12936"/>
		<updated>2025-03-10T12:08:11Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: /* Assignments */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Infobox&lt;br /&gt;
|name         = Infobox&lt;br /&gt;
|bodystyle    = &lt;br /&gt;
|title        = &amp;lt;font size=3&amp;gt;&#039;&#039;&#039;概率论与数理统计&#039;&#039;&#039;&amp;lt;br&amp;gt;&lt;br /&gt;
&#039;&#039;&#039;Probability Theory&#039;&#039;&#039; &amp;lt;br&amp;gt; &amp;amp; &#039;&#039;&#039;Mathematical Statistics&#039;&#039;&#039;&amp;lt;/font&amp;gt;&lt;br /&gt;
|titlestyle   = &lt;br /&gt;
&lt;br /&gt;
|image        = &lt;br /&gt;
|imagestyle   = &lt;br /&gt;
|caption      = &lt;br /&gt;
|captionstyle = &lt;br /&gt;
|headerstyle  = background:#ccf;&lt;br /&gt;
|labelstyle   = background:#ddf;&lt;br /&gt;
|datastyle    = &lt;br /&gt;
&lt;br /&gt;
|header1 =Instructor&lt;br /&gt;
|label1  = &lt;br /&gt;
|data1   = &lt;br /&gt;
|header2 = &lt;br /&gt;
|label2  = &lt;br /&gt;
|data2   = &#039;&#039;&#039;尹一通&#039;&#039;&#039;&lt;br /&gt;
|header3 = &lt;br /&gt;
|label3  = Email&lt;br /&gt;
|data3   = yinyt@nju.edu.cn  &lt;br /&gt;
|header4 =&lt;br /&gt;
|label4  = office&lt;br /&gt;
|data4   = 计算机系 804&lt;br /&gt;
|header5 = &lt;br /&gt;
|label5  = &lt;br /&gt;
|data5   = &#039;&#039;&#039;刘景铖&#039;&#039;&#039;&lt;br /&gt;
|header6 = &lt;br /&gt;
|label6  = Email&lt;br /&gt;
|data6   = liu@nju.edu.cn  &lt;br /&gt;
|header7 =&lt;br /&gt;
|label7  = office&lt;br /&gt;
|data7   = 计算机系 516&lt;br /&gt;
|header8 = Class&lt;br /&gt;
|label8  = &lt;br /&gt;
|data8   = &lt;br /&gt;
|header9 =&lt;br /&gt;
|label9  = Class meeting&lt;br /&gt;
|data9   = Monday, 10am-12pm&amp;lt;br&amp;gt;&lt;br /&gt;
Wednesday (单), 10am-12pm&amp;lt;br&amp;gt;&lt;br /&gt;
仙Ⅰ-204&lt;br /&gt;
|header10=&lt;br /&gt;
|label10 = Office hour&lt;br /&gt;
|data10  = Friday, 2pm-3pm &amp;lt;br&amp;gt;计算机系 804（尹一通）&amp;lt;br&amp;gt;计算机系 516（刘景铖）&lt;br /&gt;
|header11= Textbook&lt;br /&gt;
|label11 = &lt;br /&gt;
|data11  = &lt;br /&gt;
|header12=&lt;br /&gt;
|label12 = &lt;br /&gt;
|data12  = [[File:概率导论.jpeg|border|100px]]&lt;br /&gt;
|header13=&lt;br /&gt;
|label13 = &lt;br /&gt;
|data13  = &#039;&#039;&#039;概率导论&#039;&#039;&#039;（第2版·修订版）&amp;lt;br&amp;gt; Dimitri P. Bertsekas and John N. Tsitsiklis&amp;lt;br&amp;gt; 郑忠国 童行伟 译；人民邮电出版社 (2022)&lt;br /&gt;
|header14=&lt;br /&gt;
|label14 = &lt;br /&gt;
|data14  = [[File:Grimmett_probability.jpg|border|100px]]&lt;br /&gt;
|header15=&lt;br /&gt;
|label15 = &lt;br /&gt;
|data15  = &#039;&#039;&#039;Probability and Random Processes&#039;&#039;&#039; (4E) &amp;lt;br&amp;gt; Geoffrey Grimmett and David Stirzaker &amp;lt;br&amp;gt;  Oxford University Press (2020)&lt;br /&gt;
|header16=&lt;br /&gt;
|label16 = &lt;br /&gt;
|data16  = [[File:Probability_and_Computing_2ed.jpg|border|100px]]&lt;br /&gt;
|header17=&lt;br /&gt;
|label17 = &lt;br /&gt;
|data17  = &#039;&#039;&#039;Probability and Computing&#039;&#039;&#039; (2E) &amp;lt;br&amp;gt; Michael Mitzenmacher and Eli Upfal &amp;lt;br&amp;gt;   Cambridge University Press (2017)&lt;br /&gt;
|belowstyle = background:#ddf;&lt;br /&gt;
|below = &lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
This is the webpage for the &#039;&#039;Probability Theory and Mathematical Statistics&#039;&#039; (概率论与数理统计) class of Spring 2025. Students who take this class should check this page periodically for content updates and new announcements. &lt;br /&gt;
&lt;br /&gt;
= Announcement =&lt;br /&gt;
* TBA&lt;br /&gt;
&lt;br /&gt;
= Course info =&lt;br /&gt;
* &#039;&#039;&#039;Instructor &#039;&#039;&#039;: &lt;br /&gt;
:* [http://tcs.nju.edu.cn/yinyt/ 尹一通]：[mailto:yinyt@nju.edu.cn &amp;lt;yinyt@nju.edu.cn&amp;gt;]，计算机系 804 &lt;br /&gt;
:* [https://liuexp.github.io 刘景铖]：[mailto:liu@nju.edu.cn &amp;lt;liu@nju.edu.cn&amp;gt;]，计算机系 516 &lt;br /&gt;
* &#039;&#039;&#039;Teaching assistant&#039;&#039;&#039;:&lt;br /&gt;
** [https://sites.google.com/view/xinyuanzhang 张昕渊]：[mailto:zhangxy@smail.nju.edu.cn &amp;lt;zhangxy@smail.nju.edu.cn&amp;gt;]，计算机系 410&lt;br /&gt;
** 邹宗瑞：[mailto:zou.zongrui@smail.nju.edu.cn &amp;lt;zou.zongrui@smail.nju.edu.cn&amp;gt;]，计算机系 410&lt;br /&gt;
* &#039;&#039;&#039;Class meeting&#039;&#039;&#039;:&lt;br /&gt;
** 周一：10am-12pm，仙Ⅰ-204&lt;br /&gt;
** 周三（单）：10am-12pm，仙Ⅰ-204&lt;br /&gt;
* &#039;&#039;&#039;Office hour&#039;&#039;&#039;: &lt;br /&gt;
:* 周五 2pm-3pm, 计算机系 804（尹一通）&lt;br /&gt;
:* 周五 2pm-3pm, 计算机系 516（刘景铖）&lt;br /&gt;
:* &#039;&#039;&#039;QQ群&#039;&#039;&#039;: 974742320（申请加入需提供姓名、院系、学号）&lt;br /&gt;
&lt;br /&gt;
= Syllabus =&lt;br /&gt;
课程内容分为三大部分：&lt;br /&gt;
* &#039;&#039;&#039;经典概率论&#039;&#039;&#039;：概率空间、随机变量及其数字特征、多维与连续随机变量、极限定理等内容&lt;br /&gt;
* &#039;&#039;&#039;概率与计算&#039;&#039;&#039;：测度集中现象 (concentration of measure)、概率法 (the probabilistic method)、离散随机过程的相关专题&lt;br /&gt;
* &#039;&#039;&#039;数理统计&#039;&#039;&#039;：参数估计、假设检验、贝叶斯估计、线性回归等统计推断等概念&lt;br /&gt;
&lt;br /&gt;
对于第一和第二部分，要求清楚掌握基本概念，深刻理解关键的现象与规律以及背后的原理，并可以灵活运用所学方法求解相关问题。对于第三部分，要求熟悉数理统计的若干基本概念，以及典型的统计模型和统计推断问题。&lt;br /&gt;
&lt;br /&gt;
经过本课程的训练，力求使学生能够熟悉掌握概率的语言，并会利用概率思维来理解客观世界并对其建模，以及驾驭概率的数学工具来分析和求解专业问题。&lt;br /&gt;
&lt;br /&gt;
=== 教材与参考书 Course Materials ===&lt;br /&gt;
* &#039;&#039;&#039;[BT]&#039;&#039;&#039; 概率导论（第2版·修订版），[美]伯特瑟卡斯（Dimitri P.Bertsekas）[美]齐齐克利斯（John N.Tsitsiklis）著，郑忠国 童行伟 译，人民邮电出版社（2022）。&lt;br /&gt;
* &#039;&#039;&#039;[GS]&#039;&#039;&#039; &#039;&#039;Probability and Random Processes&#039;&#039;, by Geoffrey Grimmett and David Stirzaker; Oxford University Press; 4th edition (2020).&lt;br /&gt;
* &#039;&#039;&#039;[MU]&#039;&#039;&#039; &#039;&#039;Probability and Computing: Randomization and Probabilistic Techniques in Algorithms and Data Analysis&#039;&#039;, by Michael Mitzenmacher, Eli Upfal; Cambridge University Press; 2nd edition (2017).&lt;br /&gt;
&lt;br /&gt;
=== 成绩 Grading Policy ===&lt;br /&gt;
* 课程成绩：本课程将会有若干次作业和一次期末考试。最终成绩将由平时作业成绩和期末考试成绩综合得出。&lt;br /&gt;
* 迟交：如果有特殊的理由，无法按时完成作业，请提前联系授课老师，给出正当理由。否则迟交的作业将不被接受。&lt;br /&gt;
&lt;br /&gt;
=== &amp;lt;font color=red&amp;gt; 学术诚信 Academic Integrity &amp;lt;/font&amp;gt;===&lt;br /&gt;
学术诚信是所有从事学术活动的学生和学者最基本的职业道德底线，本课程将不遗余力的维护学术诚信规范，违反这一底线的行为将不会被容忍。&lt;br /&gt;
&lt;br /&gt;
作业完成的原则：署你名字的工作必须是你个人的贡献。在完成作业的过程中，允许讨论，前提是讨论的所有参与者均处于同等完成度。但关键想法的执行、以及作业文本的写作必须独立完成，并在作业中致谢（acknowledge）所有参与讨论的人。符合规则的讨论与致谢将不会影响得分。不允许其他任何形式的合作——尤其是与已经完成作业的同学“讨论”。&lt;br /&gt;
&lt;br /&gt;
本课程将对剽窃行为采取零容忍的态度。在完成作业过程中，对他人工作（出版物、互联网资料、其他人的作业等）直接的文本抄袭和对关键思想、关键元素的抄袭，按照 [http://www.acm.org/publications/policies/plagiarism_policy ACM Policy on Plagiarism]的解释，都将视为剽窃。剽窃者成绩将被取消。如果发现互相抄袭行为，&amp;lt;font color=red&amp;gt; 抄袭和被抄袭双方的成绩都将被取消&amp;lt;/font&amp;gt;。因此请主动防止自己的作业被他人抄袭。&lt;br /&gt;
&lt;br /&gt;
学术诚信影响学生个人的品行，也关乎整个教育系统的正常运转。为了一点分数而做出学术不端的行为，不仅使自己沦为一个欺骗者，也使他人的诚实努力失去意义。让我们一起努力维护一个诚信的环境。&lt;br /&gt;
&lt;br /&gt;
= Assignments =&lt;br /&gt;
*[[概率论与数理统计 (Spring 2025)/Problem Set 1|Problem Set 1]]  请在 2025/3/10 上课之前(10am UTC+8)提交到 [mailto:pr_nju@163.com pr_nju@163.com] (文件名为&#039;&amp;lt;font color=red &amp;gt;学号_姓名_A1.pdf&amp;lt;/font&amp;gt;&#039;).&lt;br /&gt;
** [[概率论与数理统计 (Spring 2025)/第一次作业提交名单|第一次作业提交名单]]&lt;br /&gt;
&lt;br /&gt;
= Lectures =&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2025/Intro.pdf 课程简介]&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2025/ProbSpace.pdf 概率空间]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 第1章&#039;&#039;&#039; 或 &#039;&#039;&#039;[GS] Chapter 1&#039;&#039;&#039;&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2025)/Entropy and volume of Hamming balls|Entropy and volume of Hamming balls]]&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2025/RandVar.pdf 随机变量]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 第2章&#039;&#039;&#039; 或 &#039;&#039;&#039;[GS] Chapter 2, Sections 3.1~3.5, 3.7&#039;&#039;&#039;&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[MU] Chapter 2&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= Concepts =&lt;br /&gt;
* [https://plato.stanford.edu/entries/probability-interpret/ Interpretations of probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/History_of_probability History of probability]&lt;br /&gt;
* Example problems:&lt;br /&gt;
** [https://dornsifecms.usc.edu/assets/sites/520/docs/VonNeumann-ams12p36-38.pdf von Neumann&#039;s Bernoulli factory] and other [https://peteroupc.github.io/bernoulli.html Bernoulli factory algorithms]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Boy_or_Girl_paradox Boy or Girl paradox]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Monty_Hall_problem Monty Hall problem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bertrand_paradox_(probability) Bertrand paradox]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hard_spheres Hard spheres model] and [https://en.wikipedia.org/wiki/Ising_model Ising model]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/PageRank &#039;&#039;PageRank&#039;&#039;] and stationary [https://en.wikipedia.org/wiki/Random_walk random walk]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Diffusion_process Diffusion process] and [https://en.wikipedia.org/wiki/Diffusion_model diffusion model]&lt;br /&gt;
*[https://en.wikipedia.org/wiki/Probability_space Probability space]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Sample_space Sample space]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Event_(probability_theory) Event] and [https://en.wikipedia.org/wiki/Σ-algebra &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-algebra]&lt;br /&gt;
** Kolmogorov&#039;s [https://en.wikipedia.org/wiki/Probability_axioms axioms of probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Discrete_uniform_distribution Classical] and [https://en.wikipedia.org/wiki/Geometric_probability goemetric probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Boole%27s_inequality Union bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Inclusion%E2%80%93exclusion_principle Inclusion-Exclusion principle]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Boole%27s_inequality#Bonferroni_inequalities Bonferroni inequalities]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Conditional_probability Conditional probability]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chain_rule_(probability) Chain rule]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_total_probability Law of total probability]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bayes%27_theorem Bayes&#039; law]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Independence_(probability_theory) Independence] &lt;br /&gt;
** [https://en.wikipedia.org/wiki/Pairwise_independence Pairwise independence]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Random_variable Random variable]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cumulative_distribution_function Cumulative distribution function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_mass_function Probability mass function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_density_function Probability density function]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Multivariate_random_variable Random vector]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Joint_probability_distribution Joint probability distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Conditional_probability_distribution Conditional probability distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Marginal_distribution Marginal distribution]&lt;br /&gt;
* Some &#039;&#039;&#039;discrete&#039;&#039;&#039; probability distributions&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bernoulli_trial Bernoulli trial] and [https://en.wikipedia.org/wiki/Bernoulli_distribution Bernoulli distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Discrete_uniform_distribution Discrete uniform distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Binomial_distribution Binomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Geometric_distribution Geometric distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Negative_binomial_distribution Negative binomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hypergeometric_distribution Hypergeometric distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Poisson_distribution Poisson distribution]&lt;br /&gt;
** and [https://en.wikipedia.org/wiki/List_of_probability_distributions#Discrete_distributions others]&lt;br /&gt;
* Balls into bins model&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Multinomial_distribution Multinomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Birthday_problem Birthday problem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Coupon_collector%27s_problem Coupon collector]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Balls_into_bins_problem Occupancy problem]&lt;br /&gt;
* Random graphs&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Erd%C5%91s%E2%80%93R%C3%A9nyi_model Erdős–Rényi random graph model]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Galton%E2%80%93Watson_process Galton–Watson branching process]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Expected_value Expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_the_unconscious_statistician Law of the unconscious statistician, &#039;&#039;LOTUS&#039;&#039;]&lt;br /&gt;
** [https://dlsun.github.io/probability/linearity.html Linearity of expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Conditional_expectation Conditional expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_total_expectation Law of total expectation]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Markov%27s_inequality Markov&#039;s inequality]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Chebyshev%27s_inequality Chebyshev&#039;s inequality]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Moment_(mathematics) Moment] related&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Central_moment Central moment]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Variance Variance]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Standard_deviation Standard deviation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Covariance Covariance]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Correlation Correlation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Correlation_does_not_imply_causation Correlation does not imply causation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Skewness Skewness]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Kurtosis Kurtosis]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Moment_problem Moment problem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Measure_(mathematics) Measure]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Borel_set Borel set]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Lebesgue_integration Lebesgue integration]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Measurable_function Measurable function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cantor_set Cantor set]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Non-measurable_set Non-measurable set]&lt;br /&gt;
*[https://en.wikipedia.org/wiki/Probability_density_function Probability density function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_distribution#Absolutely_continuous_probability_distribution Continuous probability distribution]&lt;br /&gt;
**[https://en.wikipedia.org/wiki/Conditional_probability_distribution#Conditional_continuous_distributions Conditional continuous distributions]&lt;br /&gt;
**[https://en.wikipedia.org/wiki/Convolution_of_probability_distributions Convolution of probability distributions] and [https://en.wikipedia.org/wiki/List_of_convolutions_of_probability_distributions List of convolutions of probability distributions]&lt;br /&gt;
* Some &#039;&#039;&#039;continuous&#039;&#039;&#039; probability distributions&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Continuous_uniform_distribution Continuous uniform distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Exponential_distribution Exponential distribution] and [https://en.wikipedia.org/wiki/Poisson_point_process Poisson point process]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Normal_distribution Normal (Gaussion) distribution]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Gaussian_function Gaussian function] and [https://en.wikipedia.org/wiki/Gaussian_integral Gaussian integral]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Standard_normal_table Standard normal table]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/68%E2%80%9395%E2%80%9399.7_rule 68–95–99.7 rule]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Sum_of_normally_distributed_random_variables Sum of normally distributed random variables]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Multivariate_normal_distribution Multivariate normal distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chi-squared_distribution Chi-squared distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Gamma_distribution Gamma distribution] and [https://en.wikipedia.org/wiki/Gamma_function Gamma function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Beta_distribution Beta distribution] and [https://en.wikipedia.org/wiki/Beta_function Beta function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cauchy_distribution Cauchy distribution]&lt;br /&gt;
** and [https://en.wikipedia.org/wiki/List_of_probability_distributions#Absolutely_continuous_distributions others]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Inverse_transform_sampling Inverse transform sampling]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stochastic_dominance Stochastic dominance] and [https://en.wikipedia.org/wiki/Coupling_(probability) Coupling] &lt;br /&gt;
* [https://en.wikipedia.org/wiki/Moment-generating_function Moment-generating function]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Convergence_of_random_variables Convergence of random variables]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Pointwise_convergence Pointwise convergence]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Convergence_of_measures#Weak_convergence_of_measures Weak convergence of measures]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Convergence_in_measure Convergence in measure]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Skorokhod%27s_representation_theorem Skorokhod&#039;s representation theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Continuous_mapping_theorem Continuous mapping theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Borel%E2%80%93Cantelli_lemma Borel–Cantelli lemma] and [https://en.wikipedia.org/wiki/Zero%E2%80%93one_law zero-one laws]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Law_of_large_numbers Law of large numbers]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Central_limit_theorem Central limit theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/De_Moivre%E2%80%93Laplace_theorem De Moivre–Laplace theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Berry%E2%80%93Esseen_theorem Berry–Esseen theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Characteristic_function_(probability_theory) Characteristic function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Lévy%27s_continuity_theorem Lévy&#039;s continuity theorem]&lt;br /&gt;
* Concentration of measures:&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Moment-generating_function Moment-generating function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chernoff_bound Chernoff bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hoeffding%27s_inequality Hoeffding&#039;s bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Doob_martingale#McDiarmid&#039;s_inequality McDiarmid&#039;s inequality]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Doob_martingale Doob martingale] and [https://en.wikipedia.org/wiki/Azuma%27s_inequality Azuma&#039;s inequality]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Sub-Gaussian_distribution Sub-Gaussian distribution]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stochastic_process Stochastic process]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stopping_time Stopping time]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Martingale_(probability_theory) Martingale]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Optional_stopping_theorem Optional stopping theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Wald%27s_equation Wald&#039;s equation]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Markov_chain Markov chain]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Markov_property Markov property] and [https://en.wikipedia.org/wiki/Memorylessness memorylessness]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Stochastic_matrix Stochastic matrix]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Stationary_distribution Stationary distribution]&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)&amp;diff=12911</id>
		<title>概率论与数理统计 (Spring 2025)</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)&amp;diff=12911"/>
		<updated>2025-02-21T10:39:23Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: /* Assignments */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Infobox&lt;br /&gt;
|name         = Infobox&lt;br /&gt;
|bodystyle    = &lt;br /&gt;
|title        = &amp;lt;font size=3&amp;gt;&#039;&#039;&#039;概率论与数理统计&#039;&#039;&#039;&amp;lt;br&amp;gt;&lt;br /&gt;
&#039;&#039;&#039;Probability Theory&#039;&#039;&#039; &amp;lt;br&amp;gt; &amp;amp; &#039;&#039;&#039;Mathematical Statistics&#039;&#039;&#039;&amp;lt;/font&amp;gt;&lt;br /&gt;
|titlestyle   = &lt;br /&gt;
&lt;br /&gt;
|image        = &lt;br /&gt;
|imagestyle   = &lt;br /&gt;
|caption      = &lt;br /&gt;
|captionstyle = &lt;br /&gt;
|headerstyle  = background:#ccf;&lt;br /&gt;
|labelstyle   = background:#ddf;&lt;br /&gt;
|datastyle    = &lt;br /&gt;
&lt;br /&gt;
|header1 =Instructor&lt;br /&gt;
|label1  = &lt;br /&gt;
|data1   = &lt;br /&gt;
|header2 = &lt;br /&gt;
|label2  = &lt;br /&gt;
|data2   = &#039;&#039;&#039;尹一通&#039;&#039;&#039;&lt;br /&gt;
|header3 = &lt;br /&gt;
|label3  = Email&lt;br /&gt;
|data3   = yinyt@nju.edu.cn  &lt;br /&gt;
|header4 =&lt;br /&gt;
|label4  = office&lt;br /&gt;
|data4   = 计算机系 804&lt;br /&gt;
|header5 = &lt;br /&gt;
|label5  = &lt;br /&gt;
|data5   = &#039;&#039;&#039;刘景铖&#039;&#039;&#039;&lt;br /&gt;
|header6 = &lt;br /&gt;
|label6  = Email&lt;br /&gt;
|data6   = liu@nju.edu.cn  &lt;br /&gt;
|header7 =&lt;br /&gt;
|label7  = office&lt;br /&gt;
|data7   = 计算机系 516&lt;br /&gt;
|header8 = Class&lt;br /&gt;
|label8  = &lt;br /&gt;
|data8   = &lt;br /&gt;
|header9 =&lt;br /&gt;
|label9  = Class meeting&lt;br /&gt;
|data9   = Monday, 10am-12pm&amp;lt;br&amp;gt;&lt;br /&gt;
Wednesday (单), 10am-12pm&amp;lt;br&amp;gt;&lt;br /&gt;
仙Ⅰ-204&lt;br /&gt;
|header10=&lt;br /&gt;
|label10 = Office hour&lt;br /&gt;
|data10  = Friday, 2pm-3pm &amp;lt;br&amp;gt;计算机系 804（尹一通）&amp;lt;br&amp;gt;计算机系 516（刘景铖）&lt;br /&gt;
|header11= Textbook&lt;br /&gt;
|label11 = &lt;br /&gt;
|data11  = &lt;br /&gt;
|header12=&lt;br /&gt;
|label12 = &lt;br /&gt;
|data12  = [[File:概率导论.jpeg|border|100px]]&lt;br /&gt;
|header13=&lt;br /&gt;
|label13 = &lt;br /&gt;
|data13  = &#039;&#039;&#039;概率导论&#039;&#039;&#039;（第2版·修订版）&amp;lt;br&amp;gt; Dimitri P. Bertsekas and John N. Tsitsiklis&amp;lt;br&amp;gt; 郑忠国 童行伟 译；人民邮电出版社 (2022)&lt;br /&gt;
|header14=&lt;br /&gt;
|label14 = &lt;br /&gt;
|data14  = [[File:Grimmett_probability.jpg|border|100px]]&lt;br /&gt;
|header15=&lt;br /&gt;
|label15 = &lt;br /&gt;
|data15  = &#039;&#039;&#039;Probability and Random Processes&#039;&#039;&#039; (4E) &amp;lt;br&amp;gt; Geoffrey Grimmett and David Stirzaker &amp;lt;br&amp;gt;  Oxford University Press (2020)&lt;br /&gt;
|header16=&lt;br /&gt;
|label16 = &lt;br /&gt;
|data16  = [[File:Probability_and_Computing_2ed.jpg|border|100px]]&lt;br /&gt;
|header17=&lt;br /&gt;
|label17 = &lt;br /&gt;
|data17  = &#039;&#039;&#039;Probability and Computing&#039;&#039;&#039; (2E) &amp;lt;br&amp;gt; Michael Mitzenmacher and Eli Upfal &amp;lt;br&amp;gt;   Cambridge University Press (2017)&lt;br /&gt;
|belowstyle = background:#ddf;&lt;br /&gt;
|below = &lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
This is the webpage for the &#039;&#039;Probability Theory and Mathematical Statistics&#039;&#039; (概率论与数理统计) class of Spring 2025. Students who take this class should check this page periodically for content updates and new announcements. &lt;br /&gt;
&lt;br /&gt;
= Announcement =&lt;br /&gt;
* TBA&lt;br /&gt;
&lt;br /&gt;
= Course info =&lt;br /&gt;
* &#039;&#039;&#039;Instructor &#039;&#039;&#039;: &lt;br /&gt;
:* [http://tcs.nju.edu.cn/yinyt/ 尹一通]：[mailto:yinyt@nju.edu.cn &amp;lt;yinyt@nju.edu.cn&amp;gt;]，计算机系 804 &lt;br /&gt;
:* [https://liuexp.github.io 刘景铖]：[mailto:liu@nju.edu.cn &amp;lt;liu@nju.edu.cn&amp;gt;]，计算机系 516 &lt;br /&gt;
* &#039;&#039;&#039;Teaching assistant&#039;&#039;&#039;:&lt;br /&gt;
** [https://sites.google.com/view/xinyuanzhang 张昕渊]：[mailto:zhangxy@smail.nju.edu.cn &amp;lt;zhangxy@smail.nju.edu.cn&amp;gt;]，计算机系 410&lt;br /&gt;
** 邹宗瑞：[mailto:zou.zongrui@smail.nju.edu.cn &amp;lt;zou.zongrui@smail.nju.edu.cn&amp;gt;]，计算机系 410&lt;br /&gt;
* &#039;&#039;&#039;Class meeting&#039;&#039;&#039;:&lt;br /&gt;
** 周一：10am-12pm，仙Ⅰ-204&lt;br /&gt;
** 周三（单）：10am-12pm，仙Ⅰ-204&lt;br /&gt;
* &#039;&#039;&#039;Office hour&#039;&#039;&#039;: &lt;br /&gt;
:* 周五 2pm-3pm, 计算机系 804（尹一通）&lt;br /&gt;
:* 周五 2pm-3pm, 计算机系 516（刘景铖）&lt;br /&gt;
:* &#039;&#039;&#039;QQ群&#039;&#039;&#039;: 974742320（申请加入需提供姓名、院系、学号）&lt;br /&gt;
&lt;br /&gt;
= Syllabus =&lt;br /&gt;
课程内容分为三大部分：&lt;br /&gt;
* &#039;&#039;&#039;经典概率论&#039;&#039;&#039;：概率空间、随机变量及其数字特征、多维与连续随机变量、极限定理等内容&lt;br /&gt;
* &#039;&#039;&#039;概率与计算&#039;&#039;&#039;：测度集中现象 (concentration of measure)、概率法 (the probabilistic method)、离散随机过程的相关专题&lt;br /&gt;
* &#039;&#039;&#039;数理统计&#039;&#039;&#039;：参数估计、假设检验、贝叶斯估计、线性回归等统计推断等概念&lt;br /&gt;
&lt;br /&gt;
对于第一和第二部分，要求清楚掌握基本概念，深刻理解关键的现象与规律以及背后的原理，并可以灵活运用所学方法求解相关问题。对于第三部分，要求熟悉数理统计的若干基本概念，以及典型的统计模型和统计推断问题。&lt;br /&gt;
&lt;br /&gt;
经过本课程的训练，力求使学生能够熟悉掌握概率的语言，并会利用概率思维来理解客观世界并对其建模，以及驾驭概率的数学工具来分析和求解专业问题。&lt;br /&gt;
&lt;br /&gt;
=== 教材与参考书 Course Materials ===&lt;br /&gt;
* &#039;&#039;&#039;[BT]&#039;&#039;&#039; 概率导论（第2版·修订版），[美]伯特瑟卡斯（Dimitri P.Bertsekas）[美]齐齐克利斯（John N.Tsitsiklis）著，郑忠国 童行伟 译，人民邮电出版社（2022）。&lt;br /&gt;
* &#039;&#039;&#039;[GS]&#039;&#039;&#039; &#039;&#039;Probability and Random Processes&#039;&#039;, by Geoffrey Grimmett and David Stirzaker; Oxford University Press; 4th edition (2020).&lt;br /&gt;
* &#039;&#039;&#039;[MU]&#039;&#039;&#039; &#039;&#039;Probability and Computing: Randomization and Probabilistic Techniques in Algorithms and Data Analysis&#039;&#039;, by Michael Mitzenmacher, Eli Upfal; Cambridge University Press; 2nd edition (2017).&lt;br /&gt;
&lt;br /&gt;
=== 成绩 Grading Policy ===&lt;br /&gt;
* 课程成绩：本课程将会有若干次作业和一次期末考试。最终成绩将由平时作业成绩和期末考试成绩综合得出。&lt;br /&gt;
* 迟交：如果有特殊的理由，无法按时完成作业，请提前联系授课老师，给出正当理由。否则迟交的作业将不被接受。&lt;br /&gt;
&lt;br /&gt;
=== &amp;lt;font color=red&amp;gt; 学术诚信 Academic Integrity &amp;lt;/font&amp;gt;===&lt;br /&gt;
学术诚信是所有从事学术活动的学生和学者最基本的职业道德底线，本课程将不遗余力的维护学术诚信规范，违反这一底线的行为将不会被容忍。&lt;br /&gt;
&lt;br /&gt;
作业完成的原则：署你名字的工作必须是你个人的贡献。在完成作业的过程中，允许讨论，前提是讨论的所有参与者均处于同等完成度。但关键想法的执行、以及作业文本的写作必须独立完成，并在作业中致谢（acknowledge）所有参与讨论的人。符合规则的讨论与致谢将不会影响得分。不允许其他任何形式的合作——尤其是与已经完成作业的同学“讨论”。&lt;br /&gt;
&lt;br /&gt;
本课程将对剽窃行为采取零容忍的态度。在完成作业过程中，对他人工作（出版物、互联网资料、其他人的作业等）直接的文本抄袭和对关键思想、关键元素的抄袭，按照 [http://www.acm.org/publications/policies/plagiarism_policy ACM Policy on Plagiarism]的解释，都将视为剽窃。剽窃者成绩将被取消。如果发现互相抄袭行为，&amp;lt;font color=red&amp;gt; 抄袭和被抄袭双方的成绩都将被取消&amp;lt;/font&amp;gt;。因此请主动防止自己的作业被他人抄袭。&lt;br /&gt;
&lt;br /&gt;
学术诚信影响学生个人的品行，也关乎整个教育系统的正常运转。为了一点分数而做出学术不端的行为，不仅使自己沦为一个欺骗者，也使他人的诚实努力失去意义。让我们一起努力维护一个诚信的环境。&lt;br /&gt;
&lt;br /&gt;
= Assignments =&lt;br /&gt;
*[[概率论与数理统计 (Spring 2025)/Problem Set 1|Problem Set 1]]  请在 2025/3/10 上课之前(10am UTC+8)提交到 [mailto:pr_nju@163.com pr_nju@163.com] (文件名为&#039;&amp;lt;font color=red &amp;gt;学号_姓名_A1.pdf&amp;lt;/font&amp;gt;&#039;).&lt;br /&gt;
&lt;br /&gt;
= Lectures =&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2025/Intro.pdf 课程简介]&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2025/ProbSpace.pdf 概率空间]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 第1章&#039;&#039;&#039; 或 &#039;&#039;&#039;[GS] Chapter 1&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= Concepts =&lt;br /&gt;
* [https://plato.stanford.edu/entries/probability-interpret/ Interpretations of probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/History_of_probability History of probability]&lt;br /&gt;
* Example problems:&lt;br /&gt;
** [https://dornsifecms.usc.edu/assets/sites/520/docs/VonNeumann-ams12p36-38.pdf von Neumann&#039;s Bernoulli factory] and other [https://peteroupc.github.io/bernoulli.html Bernoulli factory algorithms]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Boy_or_Girl_paradox Boy or Girl paradox]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Monty_Hall_problem Monty Hall problem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bertrand_paradox_(probability) Bertrand paradox]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hard_spheres Hard spheres model] and [https://en.wikipedia.org/wiki/Ising_model Ising model]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/PageRank &#039;&#039;PageRank&#039;&#039;] and stationary [https://en.wikipedia.org/wiki/Random_walk random walk]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Diffusion_process Diffusion process] and [https://en.wikipedia.org/wiki/Diffusion_model diffusion model]&lt;br /&gt;
*[https://en.wikipedia.org/wiki/Probability_space Probability space]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Sample_space Sample space]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Event_(probability_theory) Event] and [https://en.wikipedia.org/wiki/Σ-algebra &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-algebra]&lt;br /&gt;
** Kolmogorov&#039;s [https://en.wikipedia.org/wiki/Probability_axioms axioms of probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Discrete_uniform_distribution Classical] and [https://en.wikipedia.org/wiki/Geometric_probability goemetric probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Boole%27s_inequality Union bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Inclusion%E2%80%93exclusion_principle Inclusion-Exclusion principle]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Boole%27s_inequality#Bonferroni_inequalities Bonferroni inequalities]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Conditional_probability Conditional probability]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chain_rule_(probability) Chain rule]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_total_probability Law of total probability]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bayes%27_theorem Bayes&#039; law]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Independence_(probability_theory) Independence] &lt;br /&gt;
** [https://en.wikipedia.org/wiki/Pairwise_independence Pairwise independence]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Random_variable Random variable]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cumulative_distribution_function Cumulative distribution function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_mass_function Probability mass function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_density_function Probability density function]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Multivariate_random_variable Random vector]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Joint_probability_distribution Joint probability distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Conditional_probability_distribution Conditional probability distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Marginal_distribution Marginal distribution]&lt;br /&gt;
* Some &#039;&#039;&#039;discrete&#039;&#039;&#039; probability distributions&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bernoulli_trial Bernoulli trial] and [https://en.wikipedia.org/wiki/Bernoulli_distribution Bernoulli distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Discrete_uniform_distribution Discrete uniform distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Binomial_distribution Binomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Geometric_distribution Geometric distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Negative_binomial_distribution Negative binomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hypergeometric_distribution Hypergeometric distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Poisson_distribution Poisson distribution]&lt;br /&gt;
** and [https://en.wikipedia.org/wiki/List_of_probability_distributions#Discrete_distributions others]&lt;br /&gt;
* Balls into bins model&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Multinomial_distribution Multinomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Birthday_problem Birthday problem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Coupon_collector%27s_problem Coupon collector]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Balls_into_bins_problem Occupancy problem]&lt;br /&gt;
* Random graphs&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Erd%C5%91s%E2%80%93R%C3%A9nyi_model Erdős–Rényi random graph model]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Galton%E2%80%93Watson_process Galton–Watson branching process]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Expected_value Expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_the_unconscious_statistician Law of the unconscious statistician, &#039;&#039;LOTUS&#039;&#039;]&lt;br /&gt;
** [https://dlsun.github.io/probability/linearity.html Linearity of expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Conditional_expectation Conditional expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_total_expectation Law of total expectation]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Markov%27s_inequality Markov&#039;s inequality]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Chebyshev%27s_inequality Chebyshev&#039;s inequality]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Moment_(mathematics) Moment] related&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Central_moment Central moment]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Variance Variance]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Standard_deviation Standard deviation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Covariance Covariance]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Correlation Correlation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Correlation_does_not_imply_causation Correlation does not imply causation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Skewness Skewness]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Kurtosis Kurtosis]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Moment_problem Moment problem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Measure_(mathematics) Measure]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Borel_set Borel set]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Lebesgue_integration Lebesgue integration]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Measurable_function Measurable function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cantor_set Cantor set]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Non-measurable_set Non-measurable set]&lt;br /&gt;
*[https://en.wikipedia.org/wiki/Probability_density_function Probability density function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_distribution#Absolutely_continuous_probability_distribution Continuous probability distribution]&lt;br /&gt;
**[https://en.wikipedia.org/wiki/Conditional_probability_distribution#Conditional_continuous_distributions Conditional continuous distributions]&lt;br /&gt;
**[https://en.wikipedia.org/wiki/Convolution_of_probability_distributions Convolution of probability distributions] and [https://en.wikipedia.org/wiki/List_of_convolutions_of_probability_distributions List of convolutions of probability distributions]&lt;br /&gt;
* Some &#039;&#039;&#039;continuous&#039;&#039;&#039; probability distributions&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Continuous_uniform_distribution Continuous uniform distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Exponential_distribution Exponential distribution] and [https://en.wikipedia.org/wiki/Poisson_point_process Poisson point process]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Normal_distribution Normal (Gaussion) distribution]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Gaussian_function Gaussian function] and [https://en.wikipedia.org/wiki/Gaussian_integral Gaussian integral]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Standard_normal_table Standard normal table]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/68%E2%80%9395%E2%80%9399.7_rule 68–95–99.7 rule]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Sum_of_normally_distributed_random_variables Sum of normally distributed random variables]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Multivariate_normal_distribution Multivariate normal distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chi-squared_distribution Chi-squared distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Gamma_distribution Gamma distribution] and [https://en.wikipedia.org/wiki/Gamma_function Gamma function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Beta_distribution Beta distribution] and [https://en.wikipedia.org/wiki/Beta_function Beta function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cauchy_distribution Cauchy distribution]&lt;br /&gt;
** and [https://en.wikipedia.org/wiki/List_of_probability_distributions#Absolutely_continuous_distributions others]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Inverse_transform_sampling Inverse transform sampling]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stochastic_dominance Stochastic dominance] and [https://en.wikipedia.org/wiki/Coupling_(probability) Coupling] &lt;br /&gt;
* [https://en.wikipedia.org/wiki/Moment-generating_function Moment-generating function]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Convergence_of_random_variables Convergence of random variables]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Pointwise_convergence Pointwise convergence]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Convergence_of_measures#Weak_convergence_of_measures Weak convergence of measures]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Convergence_in_measure Convergence in measure]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Skorokhod%27s_representation_theorem Skorokhod&#039;s representation theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Continuous_mapping_theorem Continuous mapping theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Borel%E2%80%93Cantelli_lemma Borel–Cantelli lemma] and [https://en.wikipedia.org/wiki/Zero%E2%80%93one_law zero-one laws]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Law_of_large_numbers Law of large numbers]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Central_limit_theorem Central limit theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/De_Moivre%E2%80%93Laplace_theorem De Moivre–Laplace theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Berry%E2%80%93Esseen_theorem Berry–Esseen theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Characteristic_function_(probability_theory) Characteristic function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Lévy%27s_continuity_theorem Lévy&#039;s continuity theorem]&lt;br /&gt;
* Concentration of measures:&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Moment-generating_function Moment-generating function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chernoff_bound Chernoff bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hoeffding%27s_inequality Hoeffding&#039;s bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Doob_martingale#McDiarmid&#039;s_inequality McDiarmid&#039;s inequality]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Doob_martingale Doob martingale] and [https://en.wikipedia.org/wiki/Azuma%27s_inequality Azuma&#039;s inequality]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Sub-Gaussian_distribution Sub-Gaussian distribution]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stochastic_process Stochastic process]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stopping_time Stopping time]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Martingale_(probability_theory) Martingale]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Optional_stopping_theorem Optional stopping theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Wald%27s_equation Wald&#039;s equation]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Markov_chain Markov chain]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Markov_property Markov property] and [https://en.wikipedia.org/wiki/Memorylessness memorylessness]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Stochastic_matrix Stochastic matrix]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Stationary_distribution Stationary distribution]&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_1&amp;diff=12901</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 1</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_1&amp;diff=12901"/>
		<updated>2025-02-19T12:37:15Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: /* Assumption throughout Problem Set 1 */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 1==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-2-principle-of-inclusion-and-exclusion&amp;quot;&amp;gt;Problem 1 (Principle of Inclusion and Exclusion)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Let &amp;lt;math&amp;gt;n\ge 1&amp;lt;/math&amp;gt; be a positive integer and &amp;lt;math&amp;gt;A_1,A_2,\ldots,A_n&amp;lt;/math&amp;gt; be &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; events.&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Union bound&amp;lt;/strong&amp;gt;] Prove &amp;lt;math&amp;gt;\mathbf{Pr}\left( \bigcup_{i=1}^{n} A_i \right) \le \sum_{i=1}^{n} \mathbf{Pr}\left(A_i\right)&amp;lt;/math&amp;gt;&amp;lt;strong&amp;gt; using the definition of probability space&amp;lt;/strong&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Principle of Inclusion and Exclusion (PIE)&amp;lt;/strong&amp;gt;] Prove that &amp;lt;math&amp;gt;\mathbf{Pr}\left( \bigcup_{i=1}^n A_i\right) = \sum_{\emptyset \neq S \subseteq [n]} (-1)^{|S|-1} \mathbf{Pr}\left( \bigcap_{i \in S} A_i \right)&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;[n]=\{1,2,\ldots,n\}&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Surjection&amp;lt;/strong&amp;gt;] For positive integers &amp;lt;math&amp;gt;m\ge n&amp;lt;/math&amp;gt;, prove that the probability of a uniform random function &amp;lt;math&amp;gt;f:[m]\to[n]&amp;lt;/math&amp;gt; to be surjective (满射) is &amp;lt;math&amp;gt;\sum_{k=1}^n(-1)^{n-k}{n\choose k}\left(\frac{k}{n}\right)^m&amp;lt;/math&amp;gt;. &amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Coprime integers&amp;lt;/strong&amp;gt;] Given positive integers &amp;lt;math&amp;gt;n \ge 2&amp;lt;/math&amp;gt;, calculate the number of integer pairs &amp;lt;math&amp;gt;(x,y)&amp;lt;/math&amp;gt; satisfying &amp;lt;math&amp;gt;1 \le x &amp;lt; y \le n&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathrm{gcd}(x,y) = 1&amp;lt;/math&amp;gt;. &lt;br /&gt;
(Remark: It suggests the probability that two randomly chosen integers from &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; are coprime tends to &amp;lt;math&amp;gt;\frac{6}{\pi^2}&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Bonferroni&amp;amp;#39;s inequality and Kounias&amp;amp;#39; inequality&amp;lt;/strong&amp;gt;] Prove that &lt;br /&gt;
&amp;lt;math&amp;gt;&lt;br /&gt;
\sum_{i=1}^n \mathbf{Pr}(A_i) - \sum_{1 \le i&amp;lt; j \le n} \mathbf{Pr}(A_i \cap A_j)\le \mathbf{Pr}\left(\bigcup_{i=1}^n A_i\right) \le \sum_{i=1}^n \mathbf{Pr} \left( A_i\right) - \sum_{i=2}^n \mathbf{Pr}(A_1 \cap A_i).&lt;br /&gt;
&amp;lt;/math&amp;gt; &lt;br /&gt;
(Hint: This is sometimes called Kounias&#039; inequality which is weaker than the Bonferroni&#039;s inequality. You can try using Venn diagram to understand these inequalities.)&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-1-classic-examples&amp;quot;&amp;gt;Problem 2 (Symmetric 1D random walk)&amp;lt;/h2&amp;gt;&lt;br /&gt;
A gambler plays a fair gambling game: At each round, he flips a fair coin, earns &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; point if it&#039;s HEADs, and loses &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; point if otherwise.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Symmetric 1D random walk (I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;A_i&amp;lt;/math&amp;gt; be the event that the gambler earns &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; points after playing &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; rounds of the game, that is, the number of times the coin lands on heads is equal to the number of times it lands on tails. Compute &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Pr}(A_i)&amp;lt;/math&amp;gt;. (Hint: You may use Stirling&#039;s approximation to estimate [math]\mathbf{Pr}(A_i)[/math] and derive that &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Pr}(A_i) = +\infty&amp;lt;/math&amp;gt;.)&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Symmetric 1D random walk (II)&amp;lt;/strong&amp;gt;] Suppose that the game ends upon that the gambler loses all his &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt; points. Let &amp;lt;math&amp;gt;B_i&amp;lt;/math&amp;gt; be the event that the game ends within &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; rounds. Compute &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Pr}(\overline{B_i})&amp;lt;/math&amp;gt;. &lt;br /&gt;
(Hint: You may first consider [math] \displaystyle{m = 1}[/math] case. Let [math]\displaystyle{C_i}[/math] be the event that the game ends at the [math]\displaystyle{i}[/math]-th round. (i) Prove that &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Pr}(\overline{B_i}) = \sum_{i=1}^{+\infty} (i-1) \mathbf{Pr}(C_i)&amp;lt;/math&amp;gt;. (ii) Compute &amp;lt;math&amp;gt;\mathbf{Pr}(C_i)&amp;lt;/math&amp;gt;, which is a special case of the ballot problem.  (iii) Finally, use Stirling&#039;s approximation to derive that &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Pr}(\overline{B_i}) = +\infty&amp;lt;/math&amp;gt;.)&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-3-probability-space&amp;quot;&amp;gt;Problem 3 (Probability space)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Nonexistence of probability space&amp;lt;/strong&amp;gt;] Prove that it is impossible to define a uniform probability law on natural numbers &amp;lt;math&amp;gt;\mathbb{N}&amp;lt;/math&amp;gt;. More precisely, prove that there does not exist a probability space &amp;lt;math&amp;gt;(\mathbb{N},2^{\mathbb{N}},\mathbf{Pr})&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;\mathbf{Pr}(\{i\}) = \mathbf{Pr}(\{j\})&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;i, j \in \mathbb{N}&amp;lt;/math&amp;gt;. &lt;br /&gt;
Please explain why the same argument fails to prove that there is no uniform probability law on the real interval &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt;, that is, there is no such probability space &amp;lt;math&amp;gt;([0,1],\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt; that for any interval &amp;lt;math&amp;gt;(l,r] \subseteq [0,1]&amp;lt;/math&amp;gt;, it holds that &amp;lt;math&amp;gt;(l,r] \in \mathcal{F}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{Pr}( (l,r] ) = r-l&amp;lt;/math&amp;gt;. (Actually, such probability measure does exist and is called the Lebesgue measure on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt;). &amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Smallest &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field&amp;lt;/strong&amp;gt;] For any subset &amp;lt;math&amp;gt;S \subseteq 2^\Omega&amp;lt;/math&amp;gt;, prove that the smallest &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field containing &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; is given by &amp;lt;math&amp;gt;\bigcap_{\substack{S \subseteq \mathcal{F} \subseteq 2^\Omega\\ \mathcal{F} \text{ is a } \sigma\text{-field }} } \mathcal{F}&amp;lt;/math&amp;gt;. (Hint: You should show that it is indeed a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field and also it is the smallest one containing &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;.)&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Union of &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\mathcal{F}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathcal{G}&amp;lt;/math&amp;gt; be &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-fields of subsets of &amp;lt;math&amp;gt;\Omega&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathcal{F} \cup \mathcal{G}&amp;lt;/math&amp;gt; is not necessarily a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Probability space?&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\Omega = \mathbb{R}&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathcal{F}&amp;lt;/math&amp;gt; is the set of all subsets &amp;lt;math&amp;gt;A \subseteq \Omega&amp;lt;/math&amp;gt; so that &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;\overline{A}&amp;lt;/math&amp;gt; (complement of &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;) is countable, &amp;lt;math&amp;gt;P(A) = 0&amp;lt;/math&amp;gt; in the first case and &amp;lt;math&amp;gt;P(A) = 1&amp;lt;/math&amp;gt; in the second. Is &amp;lt;math&amp;gt;(\Omega,\mathcal{F},P)&amp;lt;/math&amp;gt; a probability space? Please explain your answer.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;&amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field?&amp;lt;/strong&amp;gt;] A set &amp;lt;math&amp;gt;A \subseteq \mathbb{N}&amp;lt;/math&amp;gt; is said to have asymptotic density &amp;lt;math&amp;gt;\theta&amp;lt;/math&amp;gt; if &amp;lt;math&amp;gt;\lim_{n \to \infty} |A \cap \{1,2,\ldots,n\}| / n = \theta&amp;lt;/math&amp;gt;. Let &amp;lt;math&amp;gt;\mathcal{A}&amp;lt;/math&amp;gt; be the collection of sets for which the asymptotic density exists. Is &amp;lt;math&amp;gt;\mathcal{A}&amp;lt;/math&amp;gt; a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-algebra? Please explain your answer.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-4-conditional-probability&amp;quot;&amp;gt;Problem 4 (Conditional probability)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Positive correlation&amp;lt;/strong&amp;gt;] We say that events &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; gives &#039;&#039;positive information&#039;&#039; about event &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; if &amp;lt;math&amp;gt;\mathbf{Pr}(A|B) &amp;gt; \mathbf{Pr}(A)&amp;lt;/math&amp;gt;, that is, the occurrence of &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; makes the occurrence of &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; more likely. Now suppose that &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; gives positive information about &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;.&lt;br /&gt;
# Does &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; give positive information about &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;?&lt;br /&gt;
# Does &amp;lt;math&amp;gt;\overline{B}&amp;lt;/math&amp;gt; give negative information about &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;, that is, is it true that &amp;lt;math&amp;gt;\mathbf{Pr}(A|\overline{B}) &amp;lt; \mathbf{Pr}(A)&amp;lt;/math&amp;gt;?&lt;br /&gt;
# Does &amp;lt;math&amp;gt;\overline{B}&amp;lt;/math&amp;gt; give positive information or negative information about &amp;lt;math&amp;gt;\overline{A}&amp;lt;/math&amp;gt;?&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Balls in urns (I)&amp;lt;/strong&amp;gt;] There are &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; urns of which the &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;-th contains &amp;lt;math&amp;gt;r-1&amp;lt;/math&amp;gt; white balls and &amp;lt;math&amp;gt;n-r&amp;lt;/math&amp;gt; black balls. You pick an urn uniformly at random (here, &amp;quot;uniformly&amp;quot; means that each urn has equal probability of being chosen) and remove two balls from that urn, uniformly at random without replacement (which means that each of the &amp;lt;math&amp;gt;{n-1\choose 2}&amp;lt;/math&amp;gt; pairs of balls are chosen to be removed with equal probability). Find the following probabilities:&lt;br /&gt;
# the second ball is black;&lt;br /&gt;
# the second ball is black, given that the first is black.&lt;br /&gt;
&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Balls in urns (II)&amp;lt;/strong&amp;gt;] Suppose that an urn contains &amp;lt;math&amp;gt;w&amp;lt;/math&amp;gt; white balls and &amp;lt;math&amp;gt;b&amp;lt;/math&amp;gt; black balls. The balls are drawn from the urn one by one, each time uniformly and independently at random, without replacement (which means we do not put the chosen ball back after each drawing). Find the probabilities of the events:&lt;br /&gt;
# the first white ball drawn is the &amp;lt;math&amp;gt;(k+1)&amp;lt;/math&amp;gt;th ball;&lt;br /&gt;
# the last ball drawn is white.&lt;br /&gt;
&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-5-independence&amp;quot;&amp;gt;Problem 5 (Independence)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Let&amp;amp;#39;s consider a series of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; outputs &amp;lt;math&amp;gt;(X_1, X_2, \cdots, X_n) \in \{0,1\}^n&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; independent Bernoulli trials, where each trial succeeds with the same probability  &amp;lt;math&amp;gt;0 &amp;lt; p &amp;lt; 1&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Limited independence&amp;lt;/strong&amp;gt;] Construct three events &amp;lt;math&amp;gt;A,B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; Bernoulli trials such that &amp;lt;math&amp;gt;A, B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; are pairwise independent but are not (mutually) independent. You need to prove that the constructed events &amp;lt;math&amp;gt;A, B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; satisfy this. (Hint: Consider the case where &amp;lt;math&amp;gt;n = 2&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;p = 1/2&amp;lt;/math&amp;gt;.)&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Product distribution&amp;lt;/strong&amp;gt;] Suppose someone has observed the output of the &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; trials, and she told you that precisely &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; trials succeeded for some &amp;lt;math&amp;gt;0&amp;lt; k&amp;lt; n&amp;lt;/math&amp;gt;. Now you want to predict the output of the &amp;lt;math&amp;gt;(n+1)&amp;lt;/math&amp;gt;-th trial while the parameter &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; of the Bernoulli trial is unknown. One way to estimate &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; is to find such &amp;lt;math&amp;gt;\hat{p}&amp;lt;/math&amp;gt; that makes the observed outcomes most probable, namely you need to solve &lt;br /&gt;
  &amp;lt;math&amp;gt;\arg \max_{\hat{p}\in(0,1)} \mathbf{Pr}_{\hat{p}} [k \text{ out of } n\text{ trials succeed}].&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Estimate &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; by solving the above optimization problem.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;If someone tells you exactly which &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; trials succeed (in addition to just telling you the number of successful trials, which is &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;), would it help you to estimate &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; more accurately? Why?&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-6-probabilistic-method&amp;quot;&amp;gt;Problem 6 (Probabilistic method)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Tournaments&amp;lt;/strong&amp;gt;] A [[wikipedia:Tournament_(graph_theory)|tournament]] can be interpreted as the outcome of a round-robin tournament in which every player faces every other player exactly once, and no draws occur. Given two players (vertices) &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;, we draw an&lt;br /&gt;
arrow from &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; if &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; beats &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; (and vice versa). We say a tournament has property &amp;lt;math&amp;gt;S_k&amp;lt;/math&amp;gt; if for every &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; players, there exists another player &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt; who defeats all of them. For example, a triangle &amp;lt;math&amp;gt;x\rightarrow y \rightarrow z \rightarrow x&amp;lt;/math&amp;gt; has property &amp;lt;math&amp;gt;S_1&amp;lt;/math&amp;gt; but not &amp;lt;math&amp;gt;S_2&amp;lt;/math&amp;gt;. Prove that if &amp;lt;math&amp;gt;\binom n k(1-2^{-k})^{n-k}&amp;lt;1&amp;lt;/math&amp;gt;, then there  is a tournament on &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; vertices that has the property &amp;lt;math&amp;gt;S_k&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_1&amp;diff=12900</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 1</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_1&amp;diff=12900"/>
		<updated>2025-02-19T12:35:40Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: /* Assumption throughout Problem Set 1 */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 1==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-2-principle-of-inclusion-and-exclusion&amp;quot;&amp;gt;Problem 1 (Principle of Inclusion and Exclusion)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Let &amp;lt;math&amp;gt;n\ge 1&amp;lt;/math&amp;gt; be a positive integer and &amp;lt;math&amp;gt;A_1,A_2,\ldots,A_n&amp;lt;/math&amp;gt; be &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; events.&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Union bound&amp;lt;/strong&amp;gt;] Prove &amp;lt;math&amp;gt;\mathbf{Pr}\left( \bigcup_{i=1}^{n} A_i \right) \le \sum_{i=1}^{n} \mathbf{Pr}\left(A_i\right)&amp;lt;/math&amp;gt;&amp;lt;strong&amp;gt; using the definition of probability space&amp;lt;/strong&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Principle of Inclusion and Exclusion (PIE)&amp;lt;/strong&amp;gt;] Prove that &amp;lt;math&amp;gt;\mathbf{Pr}\left( \bigcup_{i=1}^n A_i\right) = \sum_{\emptyset \neq S \subseteq [n]} (-1)^{|S|-1} \mathbf{Pr}\left( \bigcap_{i \in S} A_i \right)&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;[n]=\{1,2,\ldots,n\}&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Surjection&amp;lt;/strong&amp;gt;] For positive integers &amp;lt;math&amp;gt;m\ge n&amp;lt;/math&amp;gt;, prove that the probability of a uniform random function &amp;lt;math&amp;gt;f:[m]\to[n]&amp;lt;/math&amp;gt; to be surjective (满射) is &amp;lt;math&amp;gt;\sum_{k=1}^n(-1)^{n-k}{n\choose k}\left(\frac{k}{n}\right)^m&amp;lt;/math&amp;gt;. &amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Coprime integers&amp;lt;/strong&amp;gt;] Given positive integers &amp;lt;math&amp;gt;n \ge 2&amp;lt;/math&amp;gt;, calculate the number of integer pairs &amp;lt;math&amp;gt;(x,y)&amp;lt;/math&amp;gt; satisfying &amp;lt;math&amp;gt;1 \le x &amp;lt; y \le n&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathrm{gcd}(x,y) = 1&amp;lt;/math&amp;gt;. &lt;br /&gt;
(Remark: It suggests the probability that two randomly chosen integers from &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; are coprime tends to &amp;lt;math&amp;gt;\frac{6}{\pi^2}&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Bonferroni&amp;amp;#39;s inequality and Kounias&amp;amp;#39; inequality&amp;lt;/strong&amp;gt;] Prove that &lt;br /&gt;
&amp;lt;math&amp;gt;&lt;br /&gt;
\sum_{i=1}^n \mathbf{Pr}(A_i) - \sum_{1 \le i&amp;lt; j \le n} \mathbf{Pr}(A_i \cap A_j)\le \mathbf{Pr}\left(\bigcup_{i=1}^n A_i\right) \le \sum_{i=1}^n \mathbf{Pr} \left( A_i\right) - \sum_{i=2}^n \mathbf{Pr}(A_1 \cap A_i).&lt;br /&gt;
&amp;lt;/math&amp;gt; &lt;br /&gt;
(Hint: This is sometimes called Kounias&#039; inequality which is weaker than the Bonferroni&#039;s inequality. You can try using Venn diagram to understand these inequalities.)&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-1-classic-examples&amp;quot;&amp;gt;Problem 2 (Symmetric 1D random walk)&amp;lt;/h2&amp;gt;&lt;br /&gt;
A gambler plays a fair gambling game: At each round, he flips a fair coin, earns &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; point if it&#039;s HEADs, and loses &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; point if otherwise.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Symmetric 1D random walk (I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;A_i&amp;lt;/math&amp;gt; be the event that the gambler earns &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; points after playing &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; rounds of the game, that is, the number of times the coin lands on heads is equal to the number of times it lands on tails. Compute &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Pr}(A_i)&amp;lt;/math&amp;gt;. (Hint: You may use Stirling&#039;s approximation to estimate [math]\mathbf{Pr}(A_i)[/math] and derive that &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Pr}(A_i) = +\infty&amp;lt;/math&amp;gt;.)&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Symmetric 1D random walk (II)&amp;lt;/strong&amp;gt;] Suppose that the game ends upon that the gambler loses all his &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt; points. Let &amp;lt;math&amp;gt;B_i&amp;lt;/math&amp;gt; be the event that the game ends within &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; rounds. Compute &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Pr}(\overline{B_i})&amp;lt;/math&amp;gt;. &lt;br /&gt;
(Hint: You may first consider [math] \displaystyle{m = 1}[/math] case. Let [math]\displaystyle{C_i}[/math] be the event that the game ends at the [math]\displaystyle{i}[/math]-th round. (i) Prove that &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Pr}(\overline{B_i}) = \sum_{i=1}^{+\infty} (i-1) \mathbf{Pr}(C_i)&amp;lt;/math&amp;gt;. (ii) Compute &amp;lt;math&amp;gt;\mathbf{Pr}(C_i)&amp;lt;/math&amp;gt;, which is a special case of the ballot problem.  (iii) Finally, use Stirling&#039;s approximation to derive that &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Pr}(\overline{B_i}) = +\infty&amp;lt;/math&amp;gt;.)&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-3-probability-space&amp;quot;&amp;gt;Problem 3 (Probability space)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Nonexistence of probability space&amp;lt;/strong&amp;gt;] Prove that it is impossible to define a uniform probability law on natural numbers &amp;lt;math&amp;gt;\mathbb{N}&amp;lt;/math&amp;gt;. More precisely, prove that there does not exist a probability space &amp;lt;math&amp;gt;(\mathbb{N},2^{\mathbb{N}},\mathbf{Pr})&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;\mathbf{Pr}(\{i\}) = \mathbf{Pr}(\{j\})&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;i, j \in \mathbb{N}&amp;lt;/math&amp;gt;. &lt;br /&gt;
Please explain why the same argument fails to prove that there is no uniform probability law on the real interval &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt;, that is, there is no such probability space &amp;lt;math&amp;gt;([0,1],\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt; that for any interval &amp;lt;math&amp;gt;(l,r] \subseteq [0,1]&amp;lt;/math&amp;gt;, it holds that &amp;lt;math&amp;gt;(l,r] \in \mathcal{F}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{Pr}( (l,r] ) = r-l&amp;lt;/math&amp;gt;. (Actually, such probability measure does exist and is called the Lebesgue measure on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt;). &amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Smallest &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field&amp;lt;/strong&amp;gt;] For any subset &amp;lt;math&amp;gt;S \subseteq 2^\Omega&amp;lt;/math&amp;gt;, prove that the smallest &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field containing &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; is given by &amp;lt;math&amp;gt;\bigcap_{\substack{S \subseteq \mathcal{F} \subseteq 2^\Omega\\ \mathcal{F} \text{ is a } \sigma\text{-field }} } \mathcal{F}&amp;lt;/math&amp;gt;. (Hint: You should show that it is indeed a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field and also it is the smallest one containing &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;.)&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Union of &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\mathcal{F}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathcal{G}&amp;lt;/math&amp;gt; be &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-fields of subsets of &amp;lt;math&amp;gt;\Omega&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathcal{F} \cup \mathcal{G}&amp;lt;/math&amp;gt; is not necessarily a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Probability space?&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\Omega = \mathbb{R}&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathcal{F}&amp;lt;/math&amp;gt; is the set of all subsets &amp;lt;math&amp;gt;A \subseteq \Omega&amp;lt;/math&amp;gt; so that &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;\overline{A}&amp;lt;/math&amp;gt; (complement of &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;) is countable, &amp;lt;math&amp;gt;P(A) = 0&amp;lt;/math&amp;gt; in the first case and &amp;lt;math&amp;gt;P(A) = 1&amp;lt;/math&amp;gt; in the second. Is &amp;lt;math&amp;gt;(\Omega,\mathcal{F},P)&amp;lt;/math&amp;gt; a probability space? Please explain your answer.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;&amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field?&amp;lt;/strong&amp;gt;] A set &amp;lt;math&amp;gt;A \subseteq \mathbb{N}&amp;lt;/math&amp;gt; is said to have asymptotic density &amp;lt;math&amp;gt;\theta&amp;lt;/math&amp;gt; if &amp;lt;math&amp;gt;\lim_{n \to \infty} |A \cap \{1,2,\ldots,n\}| / n = \theta&amp;lt;/math&amp;gt;. Let &amp;lt;math&amp;gt;\mathcal{A}&amp;lt;/math&amp;gt; be the collection of sets for which the asymptotic density exists. Is &amp;lt;math&amp;gt;\mathcal{A}&amp;lt;/math&amp;gt; a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-algebra? Please explain your answer.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-4-conditional-probability&amp;quot;&amp;gt;Problem 4 (Conditional probability)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Positive correlation&amp;lt;/strong&amp;gt;] We say that events &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; gives &#039;&#039;positive information&#039;&#039; about event &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; if &amp;lt;math&amp;gt;\mathbf{Pr}(A|B) &amp;gt; \mathbf{Pr}(A)&amp;lt;/math&amp;gt;, that is, the occurrence of &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; makes the occurrence of &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; more likely. Now suppose that &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; gives positive information about &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;.&lt;br /&gt;
# Does &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; give positive information about &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;?&lt;br /&gt;
# Does &amp;lt;math&amp;gt;\overline{B}&amp;lt;/math&amp;gt; give negative information about &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;, that is, is it true that &amp;lt;math&amp;gt;\mathbf{Pr}(A|\overline{B}) &amp;lt; \mathbf{Pr}(A)&amp;lt;/math&amp;gt;?&lt;br /&gt;
# Does &amp;lt;math&amp;gt;\overline{B}&amp;lt;/math&amp;gt; give positive information or negative information about &amp;lt;math&amp;gt;\overline{A}&amp;lt;/math&amp;gt;?&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Balls in urns (I)&amp;lt;/strong&amp;gt;] There are &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; urns of which the &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;-th contains &amp;lt;math&amp;gt;r-1&amp;lt;/math&amp;gt; white balls and &amp;lt;math&amp;gt;n-r&amp;lt;/math&amp;gt; black balls. You pick an urn uniformly at random (here, &amp;quot;uniformly&amp;quot; means that each urn has equal probability of being chosen) and remove two balls from that urn, uniformly at random without replacement (which means that each of the &amp;lt;math&amp;gt;{n-1\choose 2}&amp;lt;/math&amp;gt; pairs of balls are chosen to be removed with equal probability). Find the following probabilities:&lt;br /&gt;
# the second ball is black;&lt;br /&gt;
# the second ball is black, given that the first is black.&lt;br /&gt;
&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Balls in urns (II)&amp;lt;/strong&amp;gt;] Suppose that an urn contains &amp;lt;math&amp;gt;w&amp;lt;/math&amp;gt; white balls and &amp;lt;math&amp;gt;b&amp;lt;/math&amp;gt; black balls. The balls are drawn from the urn one by one, each time uniformly and independently at random, without replacement (which means we do not put the chosen ball back after each drawing). Find the probabilities of the events:&lt;br /&gt;
# the first white ball drawn is the &amp;lt;math&amp;gt;(k+1)&amp;lt;/math&amp;gt;th ball;&lt;br /&gt;
# the last ball drawn is white.&lt;br /&gt;
&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&#039;&#039;&#039;Noisy channel&#039;&#039;&#039;] When coded messages are sent, there are sometimes errors in transmission. In particular, Morse code uses &amp;quot;dots&amp;quot; and &amp;quot;dashes&amp;quot;, which are known to occur in the proportion of 3:4. In other words, the probability of a &amp;quot;dot&amp;quot; being sent is &amp;lt;math&amp;gt;3/7&amp;lt;/math&amp;gt;, and the probability of a &amp;quot;dash&amp;quot; being sent is &amp;lt;math&amp;gt;4/7&amp;lt;/math&amp;gt;. Suppose due to interference on the transmission channel, with probability &amp;lt;math&amp;gt;1/8&amp;lt;/math&amp;gt;, a dot will be mistakenly received as a dash, and vice versa. If we receive a dot, what is the probability that the transmitted symbol is indeed a dot? &lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-5-independence&amp;quot;&amp;gt;Problem 5 (Independence)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Let&amp;amp;#39;s consider a series of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; outputs &amp;lt;math&amp;gt;(X_1, X_2, \cdots, X_n) \in \{0,1\}^n&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; independent Bernoulli trials, where each trial succeeds with the same probability  &amp;lt;math&amp;gt;0 &amp;lt; p &amp;lt; 1&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Limited independence&amp;lt;/strong&amp;gt;] Construct three events &amp;lt;math&amp;gt;A,B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; Bernoulli trials such that &amp;lt;math&amp;gt;A, B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; are pairwise independent but are not (mutually) independent. You need to prove that the constructed events &amp;lt;math&amp;gt;A, B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; satisfy this. (Hint: Consider the case where &amp;lt;math&amp;gt;n = 2&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;p = 1/2&amp;lt;/math&amp;gt;.)&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Product distribution&amp;lt;/strong&amp;gt;] Suppose someone has observed the output of the &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; trials, and she told you that precisely &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; trials succeeded for some &amp;lt;math&amp;gt;0&amp;lt; k&amp;lt; n&amp;lt;/math&amp;gt;. Now you want to predict the output of the &amp;lt;math&amp;gt;(n+1)&amp;lt;/math&amp;gt;-th trial while the parameter &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; of the Bernoulli trial is unknown. One way to estimate &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; is to find such &amp;lt;math&amp;gt;\hat{p}&amp;lt;/math&amp;gt; that makes the observed outcomes most probable, namely you need to solve &lt;br /&gt;
  &amp;lt;math&amp;gt;\arg \max_{\hat{p}\in(0,1)} \mathbf{Pr}_{\hat{p}} [k \text{ out of } n\text{ trials succeed}].&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Estimate &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; by solving the above optimization problem.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;If someone tells you exactly which &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; trials succeed (in addition to just telling you the number of successful trials, which is &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;), would it help you to estimate &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; more accurately? Why?&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-6-probabilistic-method&amp;quot;&amp;gt;Problem 6 (Probabilistic method)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Tournaments&amp;lt;/strong&amp;gt;] A [[wikipedia:Tournament_(graph_theory)|tournament]] can be interpreted as the outcome of a round-robin tournament in which every player faces every other player exactly once, and no draws occur. Given two players (vertices) &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;, we draw an&lt;br /&gt;
arrow from &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; if &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; beats &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; (and vice versa). We say a tournament has property &amp;lt;math&amp;gt;S_k&amp;lt;/math&amp;gt; if for every &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; players, there exists another player &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt; who defeats all of them. For example, a triangle &amp;lt;math&amp;gt;x\rightarrow y \rightarrow z \rightarrow x&amp;lt;/math&amp;gt; has property &amp;lt;math&amp;gt;S_1&amp;lt;/math&amp;gt; but not &amp;lt;math&amp;gt;S_2&amp;lt;/math&amp;gt;. Prove that if &amp;lt;math&amp;gt;\binom n k(1-2^{-k})^{n-k}&amp;lt;1&amp;lt;/math&amp;gt;, then there  is a tournament on &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; vertices that has the property &amp;lt;math&amp;gt;S_k&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_1&amp;diff=12899</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 1</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_1&amp;diff=12899"/>
		<updated>2025-02-19T12:34:35Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: /* Assumption throughout Problem Set 1 */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 1==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-2-principle-of-inclusion-and-exclusion&amp;quot;&amp;gt;Problem 1 (Principle of Inclusion and Exclusion)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Let &amp;lt;math&amp;gt;n\ge 1&amp;lt;/math&amp;gt; be a positive integer and &amp;lt;math&amp;gt;A_1,A_2,\ldots,A_n&amp;lt;/math&amp;gt; be &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; events.&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Union bound&amp;lt;/strong&amp;gt;] Prove &amp;lt;math&amp;gt;\mathbf{Pr}\left( \bigcup_{i=1}^{n} A_i \right) \le \sum_{i=1}^{n} \mathbf{Pr}\left(A_i\right)&amp;lt;/math&amp;gt;&amp;lt;strong&amp;gt; using the definition of probability space&amp;lt;/strong&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Principle of Inclusion and Exclusion (PIE)&amp;lt;/strong&amp;gt;] Prove that &amp;lt;math&amp;gt;\mathbf{Pr}\left( \bigcup_{i=1}^n A_i\right) = \sum_{\emptyset \neq S \subseteq [n]} (-1)^{|S|-1} \mathbf{Pr}\left( \bigcap_{i \in S} A_i \right)&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;[n]=\{1,2,\ldots,n\}&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Surjection&amp;lt;/strong&amp;gt;] For positive integers &amp;lt;math&amp;gt;m\ge n&amp;lt;/math&amp;gt;, prove that the probability of a uniform random function &amp;lt;math&amp;gt;f:[m]\to[n]&amp;lt;/math&amp;gt; to be surjective (满射) is &amp;lt;math&amp;gt;\sum_{k=1}^n(-1)^{n-k}{n\choose k}\left(\frac{k}{n}\right)^m&amp;lt;/math&amp;gt;. &amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Coprime integers&amp;lt;/strong&amp;gt;] Given positive integers &amp;lt;math&amp;gt;n \ge 2&amp;lt;/math&amp;gt;, calculate the number of integer pairs &amp;lt;math&amp;gt;(x,y)&amp;lt;/math&amp;gt; satisfying &amp;lt;math&amp;gt;1 \le x &amp;lt; y \le n&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathrm{gcd}(x,y) = 1&amp;lt;/math&amp;gt;. &lt;br /&gt;
(Remark: It suggests the probability that two randomly chosen integers from &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; are coprime tends to &amp;lt;math&amp;gt;\frac{6}{\pi^2}&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Bonferroni&amp;amp;#39;s inequality and Kounias&amp;amp;#39; inequality&amp;lt;/strong&amp;gt;] Prove that &lt;br /&gt;
&amp;lt;math&amp;gt;&lt;br /&gt;
\sum_{i=1}^n \mathbf{Pr}(A_i) - \sum_{1 \le i&amp;lt; j \le n} \mathbf{Pr}(A_i \cap A_j)\le \mathbf{Pr}\left(\bigcup_{i=1}^n A_i\right) \le \sum_{i=1}^n \mathbf{Pr} \left( A_i\right) - \sum_{i=2}^n \mathbf{Pr}(A_1 \cap A_i).&lt;br /&gt;
&amp;lt;/math&amp;gt; &lt;br /&gt;
(Hint: This is sometimes called Kounias&#039; inequality which is weaker than the Bonferroni&#039;s inequality. You can try using Venn diagram to understand these inequalities.)&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-1-classic-examples&amp;quot;&amp;gt;Problem 2 (Symmetric 1D random walk)&amp;lt;/h2&amp;gt;&lt;br /&gt;
A gambler plays a fair gambling game: At each round, he flips a fair coin, earns &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; point if it&#039;s HEADs, and loses &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; point if otherwise.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Symmetric 1D random walk (I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;A_i&amp;lt;/math&amp;gt; be the event that the gambler earns &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; points after playing &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; rounds of the game, that is, the number of times the coin lands on heads is equal to the number of times it lands on tails. Compute &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Pr}(A_i)&amp;lt;/math&amp;gt;. (Hint: You may use Stirling&#039;s approximation to estimate [math]\mathbf{Pr}(A_i)[/math] and derive that &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Pr}(A_i) = +\infty&amp;lt;/math&amp;gt;.)&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Symmetric 1D random walk (II)&amp;lt;/strong&amp;gt;] Suppose that the game ends upon that the gambler loses all his &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt; points. Let &amp;lt;math&amp;gt;B_i&amp;lt;/math&amp;gt; be the event that the game ends within &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; rounds. Compute &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Pr}(\overline{B_i})&amp;lt;/math&amp;gt;. &lt;br /&gt;
(Hint: You may first consider [math] \displaystyle{m = 1}[/math] case. Let [math]\displaystyle{C_i}[/math] be the event that the game ends at the [math]\displaystyle{i}[/math]-th round. (i) Prove that &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Pr}(\overline{B_i}) = \sum_{i=1}^{+\infty} (i-1) \mathbf{Pr}(C_i)&amp;lt;/math&amp;gt;. (ii) Compute &amp;lt;math&amp;gt;\mathbf{Pr}(C_i)&amp;lt;/math&amp;gt;, which is a special case of the ballot problem.  (iii) Finally, use Stirling&#039;s approximation to derive that &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Pr}(\overline{B_i}) = +\infty&amp;lt;/math&amp;gt;.)&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-3-probability-space&amp;quot;&amp;gt;Problem 3 (Probability space)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Nonexistence of probability space&amp;lt;/strong&amp;gt;] Prove that it is impossible to define a uniform probability law on natural numbers &amp;lt;math&amp;gt;\mathbb{N}&amp;lt;/math&amp;gt;. More precisely, prove that there does not exist a probability space &amp;lt;math&amp;gt;(\mathbb{N},2^{\mathbb{N}},\mathbf{Pr})&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;\mathbf{Pr}(\{i\}) = \mathbf{Pr}(\{j\})&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;i, j \in \mathbb{N}&amp;lt;/math&amp;gt;. &lt;br /&gt;
Please explain why the same argument fails to prove that there is no uniform probability law on the real interval &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt;, that is, there is no such probability space &amp;lt;math&amp;gt;([0,1],\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt; that for any interval &amp;lt;math&amp;gt;(l,r] \subseteq [0,1]&amp;lt;/math&amp;gt;, it holds that &amp;lt;math&amp;gt;(l,r] \in \mathcal{F}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{Pr}( (l,r] ) = r-l&amp;lt;/math&amp;gt;. (Actually, such probability measure does exist and is called the Lebesgue measure on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt;). &amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Smallest &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field&amp;lt;/strong&amp;gt;] For any subset &amp;lt;math&amp;gt;S \subseteq 2^\Omega&amp;lt;/math&amp;gt;, prove that the smallest &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field containing &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; is given by &amp;lt;math&amp;gt;\bigcap_{\substack{S \subseteq \mathcal{F} \subseteq 2^\Omega\\ \mathcal{F} \text{ is a } \sigma\text{-field }} } \mathcal{F}&amp;lt;/math&amp;gt;. (Hint: You should show that it is indeed a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field and also it is the smallest one containing &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;.)&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Union of &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\mathcal{F}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathcal{G}&amp;lt;/math&amp;gt; be &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-fields of subsets of &amp;lt;math&amp;gt;\Omega&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathcal{F} \cup \mathcal{G}&amp;lt;/math&amp;gt; is not necessarily a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Probability space?&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\Omega = \mathbb{R}&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathcal{F}&amp;lt;/math&amp;gt; is the set of all subsets &amp;lt;math&amp;gt;A \subseteq \Omega&amp;lt;/math&amp;gt; so that &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;\overline{A}&amp;lt;/math&amp;gt; (complement of &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;) is countable, &amp;lt;math&amp;gt;P(A) = 0&amp;lt;/math&amp;gt; in the first case and &amp;lt;math&amp;gt;P(A) = 1&amp;lt;/math&amp;gt; in the second. Is &amp;lt;math&amp;gt;(\Omega,\mathcal{F},P)&amp;lt;/math&amp;gt; a probability space? Please explain your answer.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;&amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field?&amp;lt;/strong&amp;gt;] A set &amp;lt;math&amp;gt;A \subseteq \mathbb{N}&amp;lt;/math&amp;gt; is said to have asymptotic density &amp;lt;math&amp;gt;\theta&amp;lt;/math&amp;gt; if &amp;lt;math&amp;gt;\lim_{n \to \infty} |A \cap \{1,2,\ldots,n\}| / n = \theta&amp;lt;/math&amp;gt;. Let &amp;lt;math&amp;gt;\mathcal{A}&amp;lt;/math&amp;gt; be the collection of sets for which the asymptotic density exists. Is &amp;lt;math&amp;gt;\mathcal{A}&amp;lt;/math&amp;gt; a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-algebra? Please explain your answer.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-4-conditional-probability&amp;quot;&amp;gt;Problem 4 (Conditional probability)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Positive correlation&amp;lt;/strong&amp;gt;] We say that events &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; gives &#039;&#039;positive information&#039;&#039; about event &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; if &amp;lt;math&amp;gt;\mathbf{Pr}(A|B) &amp;gt; \mathbf{Pr}(A)&amp;lt;/math&amp;gt;, that is, the occurrence of &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; makes the occurrence of &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; more likely. Now suppose that &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; gives positive information about &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;.&lt;br /&gt;
# Does &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; give positive information about &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;?&lt;br /&gt;
# Does &amp;lt;math&amp;gt;\overline{B}&amp;lt;/math&amp;gt; give negative information about &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;, that is, is it true that &amp;lt;math&amp;gt;\mathbf{Pr}(A|\overline{B}) &amp;lt; \mathbf{Pr}(A)&amp;lt;/math&amp;gt;?&lt;br /&gt;
# Does &amp;lt;math&amp;gt;\overline{B}&amp;lt;/math&amp;gt; give positive information or negative information about &amp;lt;math&amp;gt;\overline{A}&amp;lt;/math&amp;gt;?&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Balls in urns (I)&amp;lt;/strong&amp;gt;] There are &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; urns of which the &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;-th contains &amp;lt;math&amp;gt;r-1&amp;lt;/math&amp;gt; white balls and &amp;lt;math&amp;gt;n-r&amp;lt;/math&amp;gt; black balls. You pick an urn uniformly at random (here, &amp;quot;uniformly&amp;quot; means that each urn has equal probability of being chosen) and remove two balls from that urn, uniformly at random without replacement (which means that each of the &amp;lt;math&amp;gt;{n-1\choose 2}&amp;lt;/math&amp;gt; pairs of balls are chosen to be removed with equal probability). Find the following probabilities:&lt;br /&gt;
# the second ball is black;&lt;br /&gt;
# the second ball is black, given that the first is black.&lt;br /&gt;
&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Balls in urns (II)&amp;lt;/strong&amp;gt;] Suppose that an urn contains &amp;lt;math&amp;gt;w&amp;lt;/math&amp;gt; white balls and &amp;lt;math&amp;gt;b&amp;lt;/math&amp;gt; black balls. The balls are drawn from the urn one by one, each time uniformly and independently at random, without replacement (which means we do not put the chosen ball back after each drawing). Find the probabilities of the events:&lt;br /&gt;
# the first white ball drawn is the &amp;lt;math&amp;gt;(k+1)&amp;lt;/math&amp;gt;th ball;&lt;br /&gt;
# the last ball drawn is white.&lt;br /&gt;
&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&#039;&#039;&#039;Noisy channel&#039;&#039;&#039;] When coded messages are sent, there are sometimes errors in transmission. In particular, Morse code uses &amp;quot;dots&amp;quot; and &amp;quot;dashes&amp;quot;, which are known to occur in the proportion of 3:4. In other words, the probability of a &amp;quot;dot&amp;quot; being sent is &amp;lt;math&amp;gt;3/7&amp;lt;/math&amp;gt;, and the probability of a &amp;quot;dash&amp;quot; being sent is &amp;lt;math&amp;gt;4/7&amp;lt;/math&amp;gt;. Suppose due to interference on the transmission channel, with probability &amp;lt;math&amp;gt;1/8&amp;lt;/math&amp;gt;, a dot will be mistakenly received as a dash, and vice versa. If we receive a dot, what is the probability that the transmitted symbol is indeed a dot? &lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-5-independence&amp;quot;&amp;gt;Problem 5 (Independence)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Let&amp;amp;#39;s consider a series of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; outputs &amp;lt;math&amp;gt;(X_1, X_2, \cdots, X_n) \in \{0,1\}^n&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; independent Bernoulli trials, where each trial succeeds with the same probability  &amp;lt;math&amp;gt;0 &amp;lt; p &amp;lt; 1&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Limited independence&amp;lt;/strong&amp;gt;] Construct three events &amp;lt;math&amp;gt;A,B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; Bernoulli trials such that &amp;lt;math&amp;gt;A, B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; are pairwise independent but are not (mutually) independent. You need to prove that the constructed events &amp;lt;math&amp;gt;A, B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; satisfy this. (Hint: Consider the case where &amp;lt;math&amp;gt;n = 2&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;p = 1/2&amp;lt;/math&amp;gt;.)&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Product distribution&amp;lt;/strong&amp;gt;] Suppose someone has observed the output of the &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; trials, and she told you that precisely &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; trials succeeded for some &amp;lt;math&amp;gt;0&amp;lt; k&amp;lt; n&amp;lt;/math&amp;gt;. Now you want to predict the output of the &amp;lt;math&amp;gt;(n+1)&amp;lt;/math&amp;gt;-th trial while the parameter &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; of the Bernoulli trial is unknown. One way to estimate &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; is to find such &amp;lt;math&amp;gt;\hat{p}&amp;lt;/math&amp;gt; that makes the observed outcomes most probable, namely you need to solve &lt;br /&gt;
  &amp;lt;math&amp;gt;\arg \max_{\hat{p}\in(0,1)} \mathbf{Pr}_{\hat{p}} [k \text{ out of } n\text{ trials succeed}].&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Estimate &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; by solving the above optimization problem.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;If someone tells you exactly which &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; trials succeed (in addition to just telling you the number of successful trials, which is &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;), would it help you to estimate &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; more accurately? Why?&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-6-probabilistic-method&amp;quot;&amp;gt;Problem 6 (Probabilistic method)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Tournaments&amp;lt;/strong&amp;gt;] A [[wikipedia:Tournament_(graph_theory)|tournament]] can be interpreted as the outcome of a round-robin tournament in which every player faces every other player exactly once, and no draws occur. Given two players &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;, we draw an&lt;br /&gt;
arrow from &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; if &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; beats &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;. We say a tournament has property &amp;lt;math&amp;gt;S_k&amp;lt;/math&amp;gt; if for every &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; players, there exists another player &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt; who defeats all of them. For example, a triangle &amp;lt;math&amp;gt;x\rightarrow y \rightarrow z \rightarrow x&amp;lt;/math&amp;gt; has property &amp;lt;math&amp;gt;S_1&amp;lt;/math&amp;gt; but not &amp;lt;math&amp;gt;S_2&amp;lt;/math&amp;gt;. Prove that if &amp;lt;math&amp;gt;\binom n k(1-2^{-k})^{n-k}&amp;lt;1&amp;lt;/math&amp;gt;, then there  is a tournament on &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; vertices that has the property &amp;lt;math&amp;gt;S_k&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_1&amp;diff=12898</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 1</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_1&amp;diff=12898"/>
		<updated>2025-02-19T12:34:22Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: /* Assumption throughout Problem Set 1 */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 1==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-2-principle-of-inclusion-and-exclusion&amp;quot;&amp;gt;Problem 1 (Principle of Inclusion and Exclusion)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Let &amp;lt;math&amp;gt;n\ge 1&amp;lt;/math&amp;gt; be a positive integer and &amp;lt;math&amp;gt;A_1,A_2,\ldots,A_n&amp;lt;/math&amp;gt; be &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; events.&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Union bound&amp;lt;/strong&amp;gt;] Prove &amp;lt;math&amp;gt;\mathbf{Pr}\left( \bigcup_{i=1}^{n} A_i \right) \le \sum_{i=1}^{n} \mathbf{Pr}\left(A_i\right)&amp;lt;/math&amp;gt;&amp;lt;strong&amp;gt; using the definition of probability space&amp;lt;/strong&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Principle of Inclusion and Exclusion (PIE)&amp;lt;/strong&amp;gt;] Prove that &amp;lt;math&amp;gt;\mathbf{Pr}\left( \bigcup_{i=1}^n A_i\right) = \sum_{\emptyset \neq S \subseteq [n]} (-1)^{|S|-1} \mathbf{Pr}\left( \bigcap_{i \in S} A_i \right)&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;[n]=\{1,2,\ldots,n\}&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Surjection&amp;lt;/strong&amp;gt;] For positive integers &amp;lt;math&amp;gt;m\ge n&amp;lt;/math&amp;gt;, prove that the probability of a uniform random function &amp;lt;math&amp;gt;f:[m]\to[n]&amp;lt;/math&amp;gt; to be surjective (满射) is &amp;lt;math&amp;gt;\sum_{k=1}^n(-1)^{n-k}{n\choose k}\left(\frac{k}{n}\right)^m&amp;lt;/math&amp;gt;. &amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Coprime integers&amp;lt;/strong&amp;gt;] Given positive integers &amp;lt;math&amp;gt;n \ge 2&amp;lt;/math&amp;gt;, calculate the number of integer pairs &amp;lt;math&amp;gt;(x,y)&amp;lt;/math&amp;gt; satisfying &amp;lt;math&amp;gt;1 \le x &amp;lt; y \le n&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathrm{gcd}(x,y) = 1&amp;lt;/math&amp;gt;. &lt;br /&gt;
(Remark: It suggests the probability that two randomly chosen integers from &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; are coprime tends to &amp;lt;math&amp;gt;\frac{6}{\pi^2}&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Bonferroni&amp;amp;#39;s inequality and Kounias&amp;amp;#39; inequality&amp;lt;/strong&amp;gt;] Prove that &lt;br /&gt;
&amp;lt;math&amp;gt;&lt;br /&gt;
\sum_{i=1}^n \mathbf{Pr}(A_i) - \sum_{1 \le i&amp;lt; j \le n} \mathbf{Pr}(A_i \cap A_j)\le \mathbf{Pr}\left(\bigcup_{i=1}^n A_i\right) \le \sum_{i=1}^n \mathbf{Pr} \left( A_i\right) - \sum_{i=2}^n \mathbf{Pr}(A_1 \cap A_i).&lt;br /&gt;
&amp;lt;/math&amp;gt; &lt;br /&gt;
(Hint: This is sometimes called Kounias&#039; inequality which is weaker than the Bonferroni&#039;s inequality. You can try using Venn diagram to understand these inequalities.)&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-1-classic-examples&amp;quot;&amp;gt;Problem 2 (Symmetric 1D random walk)&amp;lt;/h2&amp;gt;&lt;br /&gt;
A gambler plays a fair gambling game: At each round, he flips a fair coin, earns &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; point if it&#039;s HEADs, and loses &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; point if otherwise.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Symmetric 1D random walk (I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;A_i&amp;lt;/math&amp;gt; be the event that the gambler earns &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; points after playing &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; rounds of the game, that is, the number of times the coin lands on heads is equal to the number of times it lands on tails. Compute &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Pr}(A_i)&amp;lt;/math&amp;gt;. (Hint: You may use Stirling&#039;s approximation to estimate [math]\mathbf{Pr}(A_i)[/math] and derive that &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Pr}(A_i) = +\infty&amp;lt;/math&amp;gt;.)&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Symmetric 1D random walk (II)&amp;lt;/strong&amp;gt;] Suppose that the game ends upon that the gambler loses all his &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt; points. Let &amp;lt;math&amp;gt;B_i&amp;lt;/math&amp;gt; be the event that the game ends within &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; rounds. Compute &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Pr}(\overline{B_i})&amp;lt;/math&amp;gt;. &lt;br /&gt;
(Hint: You may first consider [math] \displaystyle{m = 1}[/math] case. Let [math]\displaystyle{C_i}[/math] be the event that the game ends at the [math]\displaystyle{i}[/math]-th round. (i) Prove that &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Pr}(\overline{B_i}) = \sum_{i=1}^{+\infty} (i-1) \mathbf{Pr}(C_i)&amp;lt;/math&amp;gt;. (ii) Compute &amp;lt;math&amp;gt;\mathbf{Pr}(C_i)&amp;lt;/math&amp;gt;, which is a special case of the ballot problem.  (iii) Finally, use Stirling&#039;s approximation to derive that &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Pr}(\overline{B_i}) = +\infty&amp;lt;/math&amp;gt;.)&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-3-probability-space&amp;quot;&amp;gt;Problem 3 (Probability space)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Nonexistence of probability space&amp;lt;/strong&amp;gt;] Prove that it is impossible to define a uniform probability law on natural numbers &amp;lt;math&amp;gt;\mathbb{N}&amp;lt;/math&amp;gt;. More precisely, prove that there does not exist a probability space &amp;lt;math&amp;gt;(\mathbb{N},2^{\mathbb{N}},\mathbf{Pr})&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;\mathbf{Pr}(\{i\}) = \mathbf{Pr}(\{j\})&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;i, j \in \mathbb{N}&amp;lt;/math&amp;gt;. &lt;br /&gt;
Please explain why the same argument fails to prove that there is no uniform probability law on the real interval &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt;, that is, there is no such probability space &amp;lt;math&amp;gt;([0,1],\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt; that for any interval &amp;lt;math&amp;gt;(l,r] \subseteq [0,1]&amp;lt;/math&amp;gt;, it holds that &amp;lt;math&amp;gt;(l,r] \in \mathcal{F}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{Pr}( (l,r] ) = r-l&amp;lt;/math&amp;gt;. (Actually, such probability measure does exist and is called the Lebesgue measure on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt;). &amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Smallest &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field&amp;lt;/strong&amp;gt;] For any subset &amp;lt;math&amp;gt;S \subseteq 2^\Omega&amp;lt;/math&amp;gt;, prove that the smallest &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field containing &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; is given by &amp;lt;math&amp;gt;\bigcap_{\substack{S \subseteq \mathcal{F} \subseteq 2^\Omega\\ \mathcal{F} \text{ is a } \sigma\text{-field }} } \mathcal{F}&amp;lt;/math&amp;gt;. (Hint: You should show that it is indeed a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field and also it is the smallest one containing &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;.)&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Union of &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\mathcal{F}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathcal{G}&amp;lt;/math&amp;gt; be &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-fields of subsets of &amp;lt;math&amp;gt;\Omega&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathcal{F} \cup \mathcal{G}&amp;lt;/math&amp;gt; is not necessarily a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Probability space?&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\Omega = \mathbb{R}&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathcal{F}&amp;lt;/math&amp;gt; is the set of all subsets &amp;lt;math&amp;gt;A \subseteq \Omega&amp;lt;/math&amp;gt; so that &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;\overline{A}&amp;lt;/math&amp;gt; (complement of &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;) is countable, &amp;lt;math&amp;gt;P(A) = 0&amp;lt;/math&amp;gt; in the first case and &amp;lt;math&amp;gt;P(A) = 1&amp;lt;/math&amp;gt; in the second. Is &amp;lt;math&amp;gt;(\Omega,\mathcal{F},P)&amp;lt;/math&amp;gt; a probability space? Please explain your answer.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;&amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field?&amp;lt;/strong&amp;gt;] A set &amp;lt;math&amp;gt;A \subseteq \mathbb{N}&amp;lt;/math&amp;gt; is said to have asymptotic density &amp;lt;math&amp;gt;\theta&amp;lt;/math&amp;gt; if &amp;lt;math&amp;gt;\lim_{n \to \infty} |A \cap \{1,2,\ldots,n\}| / n = \theta&amp;lt;/math&amp;gt;. Let &amp;lt;math&amp;gt;\mathcal{A}&amp;lt;/math&amp;gt; be the collection of sets for which the asymptotic density exists. Is &amp;lt;math&amp;gt;\mathcal{A}&amp;lt;/math&amp;gt; a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-algebra? Please explain your answer.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-4-conditional-probability&amp;quot;&amp;gt;Problem 4 (Conditional probability)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Positive correlation&amp;lt;/strong&amp;gt;] We say that events &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; gives &#039;&#039;positive information&#039;&#039; about event &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; if &amp;lt;math&amp;gt;\mathbf{Pr}(A|B) &amp;gt; \mathbf{Pr}(A)&amp;lt;/math&amp;gt;, that is, the occurrence of &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; makes the occurrence of &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; more likely. Now suppose that &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; gives positive information about &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;.&lt;br /&gt;
# Does &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; give positive information about &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;?&lt;br /&gt;
# Does &amp;lt;math&amp;gt;\overline{B}&amp;lt;/math&amp;gt; give negative information about &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;, that is, is it true that &amp;lt;math&amp;gt;\mathbf{Pr}(A|\overline{B}) &amp;lt; \mathbf{Pr}(A)&amp;lt;/math&amp;gt;?&lt;br /&gt;
# Does &amp;lt;math&amp;gt;\overline{B}&amp;lt;/math&amp;gt; give positive information or negative information about &amp;lt;math&amp;gt;\overline{A}&amp;lt;/math&amp;gt;?&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Balls in urns (I)&amp;lt;/strong&amp;gt;] There are &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; urns of which the &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;-th contains &amp;lt;math&amp;gt;r-1&amp;lt;/math&amp;gt; white balls and &amp;lt;math&amp;gt;n-r&amp;lt;/math&amp;gt; black balls. You pick an urn uniformly at random (here, &amp;quot;uniformly&amp;quot; means that each urn has equal probability of being chosen) and remove two balls from that urn, uniformly at random without replacement (which means that each of the &amp;lt;math&amp;gt;{n-1\choose 2}&amp;lt;/math&amp;gt; pairs of balls are chosen to be removed with equal probability). Find the following probabilities:&lt;br /&gt;
# the second ball is black;&lt;br /&gt;
# the second ball is black, given that the first is black.&lt;br /&gt;
&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Balls in urns (II)&amp;lt;/strong&amp;gt;] Suppose that an urn contains &amp;lt;math&amp;gt;w&amp;lt;/math&amp;gt; white balls and &amp;lt;math&amp;gt;b&amp;lt;/math&amp;gt; black balls. The balls are drawn from the urn one by one, each time uniformly and independently at random, without replacement (which means we do not put the chosen ball back after each drawing). Find the probabilities of the events:&lt;br /&gt;
# the first white ball drawn is the &amp;lt;math&amp;gt;(k+1)&amp;lt;/math&amp;gt;th ball;&lt;br /&gt;
# the last ball drawn is white.&lt;br /&gt;
&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&#039;&#039;&#039;Noisy channel&#039;&#039;&#039;] When coded messages are sent, there are sometimes errors in transmission. In particular, Morse code uses &amp;quot;dots&amp;quot; and &amp;quot;dashes&amp;quot;, which are known to occur in the proportion of 3:4. In other words, the probability of a &amp;quot;dot&amp;quot; being sent is &amp;lt;math&amp;gt;3/7&amp;lt;/math&amp;gt;, and the probability of a &amp;quot;dash&amp;quot; being sent is &amp;lt;math&amp;gt;4/7&amp;lt;/math&amp;gt;. Suppose due to interference on the transmission channel, with probability &amp;lt;math&amp;gt;1/8&amp;lt;/math&amp;gt;, a dot will be mistakenly received as a dash, and vice versa. If we receive a dot, what is the probability that the transmitted symbol is indeed a dot? &lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-5-independence&amp;quot;&amp;gt;Problem 5 (Independence)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Let&amp;amp;#39;s consider a series of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; outputs &amp;lt;math&amp;gt;(X_1, X_2, \cdots, X_n) \in \{0,1\}^n&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; independent Bernoulli trials, where each trial succeeds with the same probability  &amp;lt;math&amp;gt;0 &amp;lt; p &amp;lt; 1&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Limited independence&amp;lt;/strong&amp;gt;] Construct three events &amp;lt;math&amp;gt;A,B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; Bernoulli trials such that &amp;lt;math&amp;gt;A, B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; are pairwise independent but are not (mutually) independent. You need to prove that the constructed events &amp;lt;math&amp;gt;A, B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; satisfy this. (Hint: Consider the case where &amp;lt;math&amp;gt;n = 2&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;p = 1/2&amp;lt;/math&amp;gt;.)&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Product distribution&amp;lt;/strong&amp;gt;] Suppose someone has observed the output of the &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; trials, and she told you that precisely &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; trials succeeded for some &amp;lt;math&amp;gt;0&amp;lt; k&amp;lt; n&amp;lt;/math&amp;gt;. Now you want to predict the output of the &amp;lt;math&amp;gt;(n+1)&amp;lt;/math&amp;gt;-th trial while the parameter &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; of the Bernoulli trial is unknown. One way to estimate &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; is to find such &amp;lt;math&amp;gt;\hat{p}&amp;lt;/math&amp;gt; that makes the observed outcomes most probable, namely you need to solve &lt;br /&gt;
  &amp;lt;math&amp;gt;\arg \max_{\hat{p}\in(0,1)} \mathbf{Pr}_{\hat{p}} [k \text{ out of } n\text{ trials succeed}].&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Estimate &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; by solving the above optimization problem.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;If someone tells you exactly which &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; trials succeed (in addition to just telling you the number of successful trials, which is &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;), would it help you to estimate &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; more accurately? Why?&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-6-probabilistic-method&amp;quot;&amp;gt;Problem 6 (Probabilistic method)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Tournaments&amp;lt;/strong&amp;gt;] A [[wikipedia:Tournament_(graph_theory)|tournament]] can be interpreted as the outcome of a round-robin tournament in which every player faces every other player exactly once, and no draws occur. Given two players &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;, we draw an&lt;br /&gt;
arrow from &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; if &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; beats &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;. We say a tournament has property &amp;lt;math&amp;gt;S_k&amp;lt;/math&amp;gt; if for every &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; players, there exists another player &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt; who defeats all of them. For example, a triangle &amp;lt;math&amp;gt;x\rightarrow y \rightarrow \rightarrow z \rightarrow x&amp;lt;/math&amp;gt; has property &amp;lt;math&amp;gt;S_1&amp;lt;/math&amp;gt; but not &amp;lt;math&amp;gt;S_2&amp;lt;/math&amp;gt;. Prove that if &amp;lt;math&amp;gt;\binom n k(1-2^{-k})^{n-k}&amp;lt;1&amp;lt;/math&amp;gt;, then there  is a tournament on &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; vertices that has the property &amp;lt;math&amp;gt;S_k&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_1&amp;diff=12897</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 1</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_1&amp;diff=12897"/>
		<updated>2025-02-19T12:33:56Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: /* Assumption throughout Problem Set 1 */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 1==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-2-principle-of-inclusion-and-exclusion&amp;quot;&amp;gt;Problem 1 (Principle of Inclusion and Exclusion)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Let &amp;lt;math&amp;gt;n\ge 1&amp;lt;/math&amp;gt; be a positive integer and &amp;lt;math&amp;gt;A_1,A_2,\ldots,A_n&amp;lt;/math&amp;gt; be &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; events.&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Union bound&amp;lt;/strong&amp;gt;] Prove &amp;lt;math&amp;gt;\mathbf{Pr}\left( \bigcup_{i=1}^{n} A_i \right) \le \sum_{i=1}^{n} \mathbf{Pr}\left(A_i\right)&amp;lt;/math&amp;gt;&amp;lt;strong&amp;gt; using the definition of probability space&amp;lt;/strong&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Principle of Inclusion and Exclusion (PIE)&amp;lt;/strong&amp;gt;] Prove that &amp;lt;math&amp;gt;\mathbf{Pr}\left( \bigcup_{i=1}^n A_i\right) = \sum_{\emptyset \neq S \subseteq [n]} (-1)^{|S|-1} \mathbf{Pr}\left( \bigcap_{i \in S} A_i \right)&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;[n]=\{1,2,\ldots,n\}&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Surjection&amp;lt;/strong&amp;gt;] For positive integers &amp;lt;math&amp;gt;m\ge n&amp;lt;/math&amp;gt;, prove that the probability of a uniform random function &amp;lt;math&amp;gt;f:[m]\to[n]&amp;lt;/math&amp;gt; to be surjective (满射) is &amp;lt;math&amp;gt;\sum_{k=1}^n(-1)^{n-k}{n\choose k}\left(\frac{k}{n}\right)^m&amp;lt;/math&amp;gt;. &amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Coprime integers&amp;lt;/strong&amp;gt;] Given positive integers &amp;lt;math&amp;gt;n \ge 2&amp;lt;/math&amp;gt;, calculate the number of integer pairs &amp;lt;math&amp;gt;(x,y)&amp;lt;/math&amp;gt; satisfying &amp;lt;math&amp;gt;1 \le x &amp;lt; y \le n&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathrm{gcd}(x,y) = 1&amp;lt;/math&amp;gt;. &lt;br /&gt;
(Remark: It suggests the probability that two randomly chosen integers from &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; are coprime tends to &amp;lt;math&amp;gt;\frac{6}{\pi^2}&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Bonferroni&amp;amp;#39;s inequality and Kounias&amp;amp;#39; inequality&amp;lt;/strong&amp;gt;] Prove that &lt;br /&gt;
&amp;lt;math&amp;gt;&lt;br /&gt;
\sum_{i=1}^n \mathbf{Pr}(A_i) - \sum_{1 \le i&amp;lt; j \le n} \mathbf{Pr}(A_i \cap A_j)\le \mathbf{Pr}\left(\bigcup_{i=1}^n A_i\right) \le \sum_{i=1}^n \mathbf{Pr} \left( A_i\right) - \sum_{i=2}^n \mathbf{Pr}(A_1 \cap A_i).&lt;br /&gt;
&amp;lt;/math&amp;gt; &lt;br /&gt;
(Hint: This is sometimes called Kounias&#039; inequality which is weaker than the Bonferroni&#039;s inequality. You can try using Venn diagram to understand these inequalities.)&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-1-classic-examples&amp;quot;&amp;gt;Problem 2 (Symmetric 1D random walk)&amp;lt;/h2&amp;gt;&lt;br /&gt;
A gambler plays a fair gambling game: At each round, he flips a fair coin, earns &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; point if it&#039;s HEADs, and loses &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; point if otherwise.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Symmetric 1D random walk (I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;A_i&amp;lt;/math&amp;gt; be the event that the gambler earns &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; points after playing &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; rounds of the game, that is, the number of times the coin lands on heads is equal to the number of times it lands on tails. Compute &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Pr}(A_i)&amp;lt;/math&amp;gt;. (Hint: You may use Stirling&#039;s approximation to estimate [math]\mathbf{Pr}(A_i)[/math] and derive that &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Pr}(A_i) = +\infty&amp;lt;/math&amp;gt;.)&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Symmetric 1D random walk (II)&amp;lt;/strong&amp;gt;] Suppose that the game ends upon that the gambler loses all his &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt; points. Let &amp;lt;math&amp;gt;B_i&amp;lt;/math&amp;gt; be the event that the game ends within &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; rounds. Compute &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Pr}(\overline{B_i})&amp;lt;/math&amp;gt;. &lt;br /&gt;
(Hint: You may first consider [math] \displaystyle{m = 1}[/math] case. Let [math]\displaystyle{C_i}[/math] be the event that the game ends at the [math]\displaystyle{i}[/math]-th round. (i) Prove that &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Pr}(\overline{B_i}) = \sum_{i=1}^{+\infty} (i-1) \mathbf{Pr}(C_i)&amp;lt;/math&amp;gt;. (ii) Compute &amp;lt;math&amp;gt;\mathbf{Pr}(C_i)&amp;lt;/math&amp;gt;, which is a special case of the ballot problem.  (iii) Finally, use Stirling&#039;s approximation to derive that &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Pr}(\overline{B_i}) = +\infty&amp;lt;/math&amp;gt;.)&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-3-probability-space&amp;quot;&amp;gt;Problem 3 (Probability space)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Nonexistence of probability space&amp;lt;/strong&amp;gt;] Prove that it is impossible to define a uniform probability law on natural numbers &amp;lt;math&amp;gt;\mathbb{N}&amp;lt;/math&amp;gt;. More precisely, prove that there does not exist a probability space &amp;lt;math&amp;gt;(\mathbb{N},2^{\mathbb{N}},\mathbf{Pr})&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;\mathbf{Pr}(\{i\}) = \mathbf{Pr}(\{j\})&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;i, j \in \mathbb{N}&amp;lt;/math&amp;gt;. &lt;br /&gt;
Please explain why the same argument fails to prove that there is no uniform probability law on the real interval &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt;, that is, there is no such probability space &amp;lt;math&amp;gt;([0,1],\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt; that for any interval &amp;lt;math&amp;gt;(l,r] \subseteq [0,1]&amp;lt;/math&amp;gt;, it holds that &amp;lt;math&amp;gt;(l,r] \in \mathcal{F}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{Pr}( (l,r] ) = r-l&amp;lt;/math&amp;gt;. (Actually, such probability measure does exist and is called the Lebesgue measure on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt;). &amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Smallest &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field&amp;lt;/strong&amp;gt;] For any subset &amp;lt;math&amp;gt;S \subseteq 2^\Omega&amp;lt;/math&amp;gt;, prove that the smallest &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field containing &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; is given by &amp;lt;math&amp;gt;\bigcap_{\substack{S \subseteq \mathcal{F} \subseteq 2^\Omega\\ \mathcal{F} \text{ is a } \sigma\text{-field }} } \mathcal{F}&amp;lt;/math&amp;gt;. (Hint: You should show that it is indeed a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field and also it is the smallest one containing &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;.)&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Union of &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\mathcal{F}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathcal{G}&amp;lt;/math&amp;gt; be &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-fields of subsets of &amp;lt;math&amp;gt;\Omega&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathcal{F} \cup \mathcal{G}&amp;lt;/math&amp;gt; is not necessarily a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Probability space?&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\Omega = \mathbb{R}&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathcal{F}&amp;lt;/math&amp;gt; is the set of all subsets &amp;lt;math&amp;gt;A \subseteq \Omega&amp;lt;/math&amp;gt; so that &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;\overline{A}&amp;lt;/math&amp;gt; (complement of &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;) is countable, &amp;lt;math&amp;gt;P(A) = 0&amp;lt;/math&amp;gt; in the first case and &amp;lt;math&amp;gt;P(A) = 1&amp;lt;/math&amp;gt; in the second. Is &amp;lt;math&amp;gt;(\Omega,\mathcal{F},P)&amp;lt;/math&amp;gt; a probability space? Please explain your answer.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;&amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field?&amp;lt;/strong&amp;gt;] A set &amp;lt;math&amp;gt;A \subseteq \mathbb{N}&amp;lt;/math&amp;gt; is said to have asymptotic density &amp;lt;math&amp;gt;\theta&amp;lt;/math&amp;gt; if &amp;lt;math&amp;gt;\lim_{n \to \infty} |A \cap \{1,2,\ldots,n\}| / n = \theta&amp;lt;/math&amp;gt;. Let &amp;lt;math&amp;gt;\mathcal{A}&amp;lt;/math&amp;gt; be the collection of sets for which the asymptotic density exists. Is &amp;lt;math&amp;gt;\mathcal{A}&amp;lt;/math&amp;gt; a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-algebra? Please explain your answer.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-4-conditional-probability&amp;quot;&amp;gt;Problem 4 (Conditional probability)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Positive correlation&amp;lt;/strong&amp;gt;] We say that events &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; gives &#039;&#039;positive information&#039;&#039; about event &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; if &amp;lt;math&amp;gt;\mathbf{Pr}(A|B) &amp;gt; \mathbf{Pr}(A)&amp;lt;/math&amp;gt;, that is, the occurrence of &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; makes the occurrence of &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; more likely. Now suppose that &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; gives positive information about &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;.&lt;br /&gt;
# Does &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; give positive information about &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;?&lt;br /&gt;
# Does &amp;lt;math&amp;gt;\overline{B}&amp;lt;/math&amp;gt; give negative information about &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;, that is, is it true that &amp;lt;math&amp;gt;\mathbf{Pr}(A|\overline{B}) &amp;lt; \mathbf{Pr}(A)&amp;lt;/math&amp;gt;?&lt;br /&gt;
# Does &amp;lt;math&amp;gt;\overline{B}&amp;lt;/math&amp;gt; give positive information or negative information about &amp;lt;math&amp;gt;\overline{A}&amp;lt;/math&amp;gt;?&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Balls in urns (I)&amp;lt;/strong&amp;gt;] There are &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; urns of which the &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;-th contains &amp;lt;math&amp;gt;r-1&amp;lt;/math&amp;gt; white balls and &amp;lt;math&amp;gt;n-r&amp;lt;/math&amp;gt; black balls. You pick an urn uniformly at random (here, &amp;quot;uniformly&amp;quot; means that each urn has equal probability of being chosen) and remove two balls from that urn, uniformly at random without replacement (which means that each of the &amp;lt;math&amp;gt;{n-1\choose 2}&amp;lt;/math&amp;gt; pairs of balls are chosen to be removed with equal probability). Find the following probabilities:&lt;br /&gt;
# the second ball is black;&lt;br /&gt;
# the second ball is black, given that the first is black.&lt;br /&gt;
&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Balls in urns (II)&amp;lt;/strong&amp;gt;] Suppose that an urn contains &amp;lt;math&amp;gt;w&amp;lt;/math&amp;gt; white balls and &amp;lt;math&amp;gt;b&amp;lt;/math&amp;gt; black balls. The balls are drawn from the urn one by one, each time uniformly and independently at random, without replacement (which means we do not put the chosen ball back after each drawing). Find the probabilities of the events:&lt;br /&gt;
# the first white ball drawn is the &amp;lt;math&amp;gt;(k+1)&amp;lt;/math&amp;gt;th ball;&lt;br /&gt;
# the last ball drawn is white.&lt;br /&gt;
&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&#039;&#039;&#039;Noisy channel&#039;&#039;&#039;] When coded messages are sent, there are sometimes errors in transmission. In particular, Morse code uses &amp;quot;dots&amp;quot; and &amp;quot;dashes&amp;quot;, which are known to occur in the proportion of 3:4. In other words, the probability of a &amp;quot;dot&amp;quot; being sent is &amp;lt;math&amp;gt;3/7&amp;lt;/math&amp;gt;, and the probability of a &amp;quot;dash&amp;quot; being sent is &amp;lt;math&amp;gt;4/7&amp;lt;/math&amp;gt;. Suppose due to interference on the transmission channel, with probability &amp;lt;math&amp;gt;1/8&amp;lt;/math&amp;gt;, a dot will be mistakenly received as a dash, and vice versa. If we receive a dot, what is the probability that the transmitted symbol is indeed a dot? &lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-5-independence&amp;quot;&amp;gt;Problem 5 (Independence)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Let&amp;amp;#39;s consider a series of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; outputs &amp;lt;math&amp;gt;(X_1, X_2, \cdots, X_n) \in \{0,1\}^n&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; independent Bernoulli trials, where each trial succeeds with the same probability  &amp;lt;math&amp;gt;0 &amp;lt; p &amp;lt; 1&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Limited independence&amp;lt;/strong&amp;gt;] Construct three events &amp;lt;math&amp;gt;A,B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; Bernoulli trials such that &amp;lt;math&amp;gt;A, B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; are pairwise independent but are not (mutually) independent. You need to prove that the constructed events &amp;lt;math&amp;gt;A, B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; satisfy this. (Hint: Consider the case where &amp;lt;math&amp;gt;n = 2&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;p = 1/2&amp;lt;/math&amp;gt;.)&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Product distribution&amp;lt;/strong&amp;gt;] Suppose someone has observed the output of the &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; trials, and she told you that precisely &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; trials succeeded for some &amp;lt;math&amp;gt;0&amp;lt; k&amp;lt; n&amp;lt;/math&amp;gt;. Now you want to predict the output of the &amp;lt;math&amp;gt;(n+1)&amp;lt;/math&amp;gt;-th trial while the parameter &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; of the Bernoulli trial is unknown. One way to estimate &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; is to find such &amp;lt;math&amp;gt;\hat{p}&amp;lt;/math&amp;gt; that makes the observed outcomes most probable, namely you need to solve &lt;br /&gt;
  &amp;lt;math&amp;gt;\arg \max_{\hat{p}\in(0,1)} \mathbf{Pr}_{\hat{p}} [k \text{ out of } n\text{ trials succeed}].&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Estimate &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; by solving the above optimization problem.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;If someone tells you exactly which &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; trials succeed (in addition to just telling you the number of successful trials, which is &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;), would it help you to estimate &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; more accurately? Why?&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-6-probabilistic-method&amp;quot;&amp;gt;Problem 6 (Probabilistic method)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Tournaments&amp;lt;/strong&amp;gt;] A [[wikipedia:Tournament_(graph_theory)|tournament]] can be interpreted as the outcome of a round-robin tournament in which every player faces every other player exactly once, and no draws occur. Given two players &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;, we draw an&lt;br /&gt;
arrow from &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; if &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; beats &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; (we say “ &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; dominates &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;”). We say a tournament has property &amp;lt;math&amp;gt;S_k&amp;lt;/math&amp;gt; if for every &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; players, there exists another player &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt; who defeats all of them. For example, a triangle &amp;lt;math&amp;gt;x\rightarrow y \rightarrow \rightarrow z \rightarrow x&amp;lt;/math&amp;gt; has property &amp;lt;math&amp;gt;S_1&amp;lt;/math&amp;gt; but not &amp;lt;math&amp;gt;S_2&amp;lt;/math&amp;gt;. Prove that if &amp;lt;math&amp;gt;\binom n k(1-2^{-k})^{n-k}&amp;lt;1&amp;lt;/math&amp;gt;, then there  is a tournament on &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; vertices that has the property &amp;lt;math&amp;gt;S_k&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_1&amp;diff=12896</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 1</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_1&amp;diff=12896"/>
		<updated>2025-02-19T12:30:00Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: /* Assumption throughout Problem Set 1 */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 1==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-2-principle-of-inclusion-and-exclusion&amp;quot;&amp;gt;Problem 1 (Principle of Inclusion and Exclusion)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Let &amp;lt;math&amp;gt;n\ge 1&amp;lt;/math&amp;gt; be a positive integer and &amp;lt;math&amp;gt;A_1,A_2,\ldots,A_n&amp;lt;/math&amp;gt; be &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; events.&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Union bound&amp;lt;/strong&amp;gt;] Prove &amp;lt;math&amp;gt;\mathbf{Pr}\left( \bigcup_{i=1}^{n} A_i \right) \le \sum_{i=1}^{n} \mathbf{Pr}\left(A_i\right)&amp;lt;/math&amp;gt;&amp;lt;strong&amp;gt; using the definition of probability space&amp;lt;/strong&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Principle of Inclusion and Exclusion (PIE)&amp;lt;/strong&amp;gt;] Prove that &amp;lt;math&amp;gt;\mathbf{Pr}\left( \bigcup_{i=1}^n A_i\right) = \sum_{\emptyset \neq S \subseteq [n]} (-1)^{|S|-1} \mathbf{Pr}\left( \bigcap_{i \in S} A_i \right)&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;[n]=\{1,2,\ldots,n\}&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Surjection&amp;lt;/strong&amp;gt;] For positive integers &amp;lt;math&amp;gt;m\ge n&amp;lt;/math&amp;gt;, prove that the probability of a uniform random function &amp;lt;math&amp;gt;f:[m]\to[n]&amp;lt;/math&amp;gt; to be surjective (满射) is &amp;lt;math&amp;gt;\sum_{k=1}^n(-1)^{n-k}{n\choose k}\left(\frac{k}{n}\right)^m&amp;lt;/math&amp;gt;. &amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Coprime integers&amp;lt;/strong&amp;gt;] Given positive integers &amp;lt;math&amp;gt;n \ge 2&amp;lt;/math&amp;gt;, calculate the number of integer pairs &amp;lt;math&amp;gt;(x,y)&amp;lt;/math&amp;gt; satisfying &amp;lt;math&amp;gt;1 \le x &amp;lt; y \le n&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathrm{gcd}(x,y) = 1&amp;lt;/math&amp;gt;. &lt;br /&gt;
(Remark: It suggests the probability that two randomly chosen integers from &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; are coprime tends to &amp;lt;math&amp;gt;\frac{6}{\pi^2}&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Bonferroni&amp;amp;#39;s inequality and Kounias&amp;amp;#39; inequality&amp;lt;/strong&amp;gt;] Prove that &lt;br /&gt;
&amp;lt;math&amp;gt;&lt;br /&gt;
\sum_{i=1}^n \mathbf{Pr}(A_i) - \sum_{1 \le i&amp;lt; j \le n} \mathbf{Pr}(A_i \cap A_j)\le \mathbf{Pr}\left(\bigcup_{i=1}^n A_i\right) \le \sum_{i=1}^n \mathbf{Pr} \left( A_i\right) - \sum_{i=2}^n \mathbf{Pr}(A_1 \cap A_i).&lt;br /&gt;
&amp;lt;/math&amp;gt; &lt;br /&gt;
(Hint: This is sometimes called Kounias&#039; inequality which is weaker than the Bonferroni&#039;s inequality. You can try using Venn diagram to understand these inequalities.)&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-1-classic-examples&amp;quot;&amp;gt;Problem 2 (Symmetric 1D random walk)&amp;lt;/h2&amp;gt;&lt;br /&gt;
A gambler plays a fair gambling game: At each round, he flips a fair coin, earns &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; point if it&#039;s HEADs, and loses &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; point if otherwise.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Symmetric 1D random walk (I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;A_i&amp;lt;/math&amp;gt; be the event that the gambler earns &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; points after playing &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; rounds of the game, that is, the number of times the coin lands on heads is equal to the number of times it lands on tails. Compute &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Pr}(A_i)&amp;lt;/math&amp;gt;. (Hint: You may use Stirling&#039;s approximation to estimate [math]\mathbf{Pr}(A_i)[/math] and derive that &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Pr}(A_i) = +\infty&amp;lt;/math&amp;gt;.)&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Symmetric 1D random walk (II)&amp;lt;/strong&amp;gt;] Suppose that the game ends upon that the gambler loses all his &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt; points. Let &amp;lt;math&amp;gt;B_i&amp;lt;/math&amp;gt; be the event that the game ends within &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; rounds. Compute &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Pr}(\overline{B_i})&amp;lt;/math&amp;gt;. &lt;br /&gt;
(Hint: You may first consider [math] \displaystyle{m = 1}[/math] case. Let [math]\displaystyle{C_i}[/math] be the event that the game ends at the [math]\displaystyle{i}[/math]-th round. (i) Prove that &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Pr}(\overline{B_i}) = \sum_{i=1}^{+\infty} (i-1) \mathbf{Pr}(C_i)&amp;lt;/math&amp;gt;. (ii) Compute &amp;lt;math&amp;gt;\mathbf{Pr}(C_i)&amp;lt;/math&amp;gt;, which is a special case of the ballot problem.  (iii) Finally, use Stirling&#039;s approximation to derive that &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Pr}(\overline{B_i}) = +\infty&amp;lt;/math&amp;gt;.)&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-3-probability-space&amp;quot;&amp;gt;Problem 3 (Probability space)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Nonexistence of probability space&amp;lt;/strong&amp;gt;] Prove that it is impossible to define a uniform probability law on natural numbers &amp;lt;math&amp;gt;\mathbb{N}&amp;lt;/math&amp;gt;. More precisely, prove that there does not exist a probability space &amp;lt;math&amp;gt;(\mathbb{N},2^{\mathbb{N}},\mathbf{Pr})&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;\mathbf{Pr}(\{i\}) = \mathbf{Pr}(\{j\})&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;i, j \in \mathbb{N}&amp;lt;/math&amp;gt;. &lt;br /&gt;
Please explain why the same argument fails to prove that there is no uniform probability law on the real interval &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt;, that is, there is no such probability space &amp;lt;math&amp;gt;([0,1],\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt; that for any interval &amp;lt;math&amp;gt;(l,r] \subseteq [0,1]&amp;lt;/math&amp;gt;, it holds that &amp;lt;math&amp;gt;(l,r] \in \mathcal{F}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{Pr}( (l,r] ) = r-l&amp;lt;/math&amp;gt;. (Actually, such probability measure does exist and is called the Lebesgue measure on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt;). &amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Smallest &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field&amp;lt;/strong&amp;gt;] For any subset &amp;lt;math&amp;gt;S \subseteq 2^\Omega&amp;lt;/math&amp;gt;, prove that the smallest &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field containing &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; is given by &amp;lt;math&amp;gt;\bigcap_{\substack{S \subseteq \mathcal{F} \subseteq 2^\Omega\\ \mathcal{F} \text{ is a } \sigma\text{-field }} } \mathcal{F}&amp;lt;/math&amp;gt;. (Hint: You should show that it is indeed a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field and also it is the smallest one containing &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;.)&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Union of &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\mathcal{F}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathcal{G}&amp;lt;/math&amp;gt; be &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-fields of subsets of &amp;lt;math&amp;gt;\Omega&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathcal{F} \cup \mathcal{G}&amp;lt;/math&amp;gt; is not necessarily a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Probability space?&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\Omega = \mathbb{R}&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathcal{F}&amp;lt;/math&amp;gt; is the set of all subsets &amp;lt;math&amp;gt;A \subseteq \Omega&amp;lt;/math&amp;gt; so that &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;\overline{A}&amp;lt;/math&amp;gt; (complement of &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;) is countable, &amp;lt;math&amp;gt;P(A) = 0&amp;lt;/math&amp;gt; in the first case and &amp;lt;math&amp;gt;P(A) = 1&amp;lt;/math&amp;gt; in the second. Is &amp;lt;math&amp;gt;(\Omega,\mathcal{F},P)&amp;lt;/math&amp;gt; a probability space? Please explain your answer.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;&amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field?&amp;lt;/strong&amp;gt;] A set &amp;lt;math&amp;gt;A \subseteq \mathbb{N}&amp;lt;/math&amp;gt; is said to have asymptotic density &amp;lt;math&amp;gt;\theta&amp;lt;/math&amp;gt; if &amp;lt;math&amp;gt;\lim_{n \to \infty} |A \cap \{1,2,\ldots,n\}| / n = \theta&amp;lt;/math&amp;gt;. Let &amp;lt;math&amp;gt;\mathcal{A}&amp;lt;/math&amp;gt; be the collection of sets for which the asymptotic density exists. Is &amp;lt;math&amp;gt;\mathcal{A}&amp;lt;/math&amp;gt; a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-algebra? Please explain your answer.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-4-conditional-probability&amp;quot;&amp;gt;Problem 4 (Conditional probability)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Positive correlation&amp;lt;/strong&amp;gt;] We say that events &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; gives &#039;&#039;positive information&#039;&#039; about event &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; if &amp;lt;math&amp;gt;\mathbf{Pr}(A|B) &amp;gt; \mathbf{Pr}(A)&amp;lt;/math&amp;gt;, that is, the occurrence of &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; makes the occurrence of &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; more likely. Now suppose that &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; gives positive information about &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;.&lt;br /&gt;
# Does &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; give positive information about &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;?&lt;br /&gt;
# Does &amp;lt;math&amp;gt;\overline{B}&amp;lt;/math&amp;gt; give negative information about &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;, that is, is it true that &amp;lt;math&amp;gt;\mathbf{Pr}(A|\overline{B}) &amp;lt; \mathbf{Pr}(A)&amp;lt;/math&amp;gt;?&lt;br /&gt;
# Does &amp;lt;math&amp;gt;\overline{B}&amp;lt;/math&amp;gt; give positive information or negative information about &amp;lt;math&amp;gt;\overline{A}&amp;lt;/math&amp;gt;?&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Balls in urns (I)&amp;lt;/strong&amp;gt;] There are &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; urns of which the &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;-th contains &amp;lt;math&amp;gt;r-1&amp;lt;/math&amp;gt; white balls and &amp;lt;math&amp;gt;n-r&amp;lt;/math&amp;gt; black balls. You pick an urn uniformly at random (here, &amp;quot;uniformly&amp;quot; means that each urn has equal probability of being chosen) and remove two balls from that urn, uniformly at random without replacement (which means that each of the &amp;lt;math&amp;gt;{n-1\choose 2}&amp;lt;/math&amp;gt; pairs of balls are chosen to be removed with equal probability). Find the following probabilities:&lt;br /&gt;
# the second ball is black;&lt;br /&gt;
# the second ball is black, given that the first is black.&lt;br /&gt;
&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Balls in urns (II)&amp;lt;/strong&amp;gt;] Suppose that an urn contains &amp;lt;math&amp;gt;w&amp;lt;/math&amp;gt; white balls and &amp;lt;math&amp;gt;b&amp;lt;/math&amp;gt; black balls. The balls are drawn from the urn one by one, each time uniformly and independently at random, without replacement (which means we do not put the chosen ball back after each drawing). Find the probabilities of the events:&lt;br /&gt;
# the first white ball drawn is the &amp;lt;math&amp;gt;(k+1)&amp;lt;/math&amp;gt;th ball;&lt;br /&gt;
# the last ball drawn is white.&lt;br /&gt;
&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&#039;&#039;&#039;Noisy channel&#039;&#039;&#039;] When coded messages are sent, there are sometimes errors in transmission. In particular, Morse code uses &amp;quot;dots&amp;quot; and &amp;quot;dashes&amp;quot;, which are known to occur in the proportion of 3:4. In other words, the probability of a &amp;quot;dot&amp;quot; being sent is &amp;lt;math&amp;gt;3/7&amp;lt;/math&amp;gt;, and the probability of a &amp;quot;dash&amp;quot; being sent is &amp;lt;math&amp;gt;4/7&amp;lt;/math&amp;gt;. Suppose due to interference on the transmission channel, with probability &amp;lt;math&amp;gt;1/8&amp;lt;/math&amp;gt;, a dot will be mistakenly received as a dash, and vice versa. If we receive a dot, what is the probability that the transmitted symbol is indeed a dot? &lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-5-independence&amp;quot;&amp;gt;Problem 5 (Independence)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Let&amp;amp;#39;s consider a series of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; outputs &amp;lt;math&amp;gt;(X_1, X_2, \cdots, X_n) \in \{0,1\}^n&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; independent Bernoulli trials, where each trial succeeds with the same probability  &amp;lt;math&amp;gt;0 &amp;lt; p &amp;lt; 1&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Limited independence&amp;lt;/strong&amp;gt;] Construct three events &amp;lt;math&amp;gt;A,B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; Bernoulli trials such that &amp;lt;math&amp;gt;A, B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; are pairwise independent but are not (mutually) independent. You need to prove that the constructed events &amp;lt;math&amp;gt;A, B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; satisfy this. (Hint: Consider the case where &amp;lt;math&amp;gt;n = 2&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;p = 1/2&amp;lt;/math&amp;gt;.)&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Product distribution&amp;lt;/strong&amp;gt;] Suppose someone has observed the output of the &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; trials, and she told you that precisely &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; trials succeeded for some &amp;lt;math&amp;gt;0&amp;lt; k&amp;lt; n&amp;lt;/math&amp;gt;. Now you want to predict the output of the &amp;lt;math&amp;gt;(n+1)&amp;lt;/math&amp;gt;-th trial while the parameter &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; of the Bernoulli trial is unknown. One way to estimate &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; is to find such &amp;lt;math&amp;gt;\hat{p}&amp;lt;/math&amp;gt; that makes the observed outcomes most probable, namely you need to solve &lt;br /&gt;
  &amp;lt;math&amp;gt;\arg \max_{\hat{p}\in(0,1)} \mathbf{Pr}_{\hat{p}} [k \text{ out of } n\text{ trials succeed}].&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Estimate &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; by solving the above optimization problem.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;If someone tells you exactly which &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; trials succeed (in addition to just telling you the number of successful trials, which is &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;), would it help you to estimate &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; more accurately? Why?&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-6-probabilistic-method&amp;quot;&amp;gt;Problem 6 (Probabilistic method)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Tournaments&amp;lt;/strong&amp;gt;] A [[wikipedia:Tournament_(graph_theory)|tournament]] can be interpreted as the outcome of a round-robin tournament in which every player faces every other player exactly once. Given two players &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;, we draw an&lt;br /&gt;
arrow from &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; if &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; beats &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; (we say “ &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; dominates &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;”), and we draw an arrow from &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;&lt;br /&gt;
if &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; beats &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; (there are no ties). We say a tournament has property &amp;lt;math&amp;gt;S_k&amp;lt;/math&amp;gt; if for every &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; players, there exists another player &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt; who defeats all of them. Prove that if &amp;lt;math&amp;gt;\binom n k(1-2^{-k})^{n-k}&amp;lt;1&amp;lt;/math&amp;gt;, then there  is a tournament on &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; vertices that has the property &amp;lt;math&amp;gt;S_k&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_1&amp;diff=12895</id>
		<title>概率论与数理统计 (Spring 2025)/Problem Set 1</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2025)/Problem_Set_1&amp;diff=12895"/>
		<updated>2025-02-19T12:27:39Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: /* Assumption throughout Problem Set 1 */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 1==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-2-principle-of-inclusion-and-exclusion&amp;quot;&amp;gt;Problem 1 (Principle of Inclusion and Exclusion)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Let &amp;lt;math&amp;gt;n\ge 1&amp;lt;/math&amp;gt; be a positive integer and &amp;lt;math&amp;gt;A_1,A_2,\ldots,A_n&amp;lt;/math&amp;gt; be &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; events.&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Union bound&amp;lt;/strong&amp;gt;] Prove &amp;lt;math&amp;gt;\mathbf{Pr}\left( \bigcup_{i=1}^{n} A_i \right) \le \sum_{i=1}^{n} \mathbf{Pr}\left(A_i\right)&amp;lt;/math&amp;gt;&amp;lt;strong&amp;gt; using the definition of probability space&amp;lt;/strong&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Principle of Inclusion and Exclusion (PIE)&amp;lt;/strong&amp;gt;] Prove that &amp;lt;math&amp;gt;\mathbf{Pr}\left( \bigcup_{i=1}^n A_i\right) = \sum_{\emptyset \neq S \subseteq [n]} (-1)^{|S|-1} \mathbf{Pr}\left( \bigcap_{i \in S} A_i \right)&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;[n]=\{1,2,\ldots,n\}&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Surjection&amp;lt;/strong&amp;gt;] For positive integers &amp;lt;math&amp;gt;m\ge n&amp;lt;/math&amp;gt;, prove that the probability of a uniform random function &amp;lt;math&amp;gt;f:[m]\to[n]&amp;lt;/math&amp;gt; to be surjective (满射) is &amp;lt;math&amp;gt;\sum_{k=1}^n(-1)^{n-k}{n\choose k}\left(\frac{k}{n}\right)^m&amp;lt;/math&amp;gt;. &amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Coprime integers&amp;lt;/strong&amp;gt;] Given positive integers &amp;lt;math&amp;gt;n \ge 2&amp;lt;/math&amp;gt;, calculate the number of integer pairs &amp;lt;math&amp;gt;(x,y)&amp;lt;/math&amp;gt; satisfying &amp;lt;math&amp;gt;1 \le x &amp;lt; y \le n&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathrm{gcd}(x,y) = 1&amp;lt;/math&amp;gt;. &lt;br /&gt;
(Remark: It suggests the probability that two randomly chosen integers from &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; are coprime tends to &amp;lt;math&amp;gt;\frac{6}{\pi^2}&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; tends to infinity.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Bonferroni&amp;amp;#39;s inequality and Kounias&amp;amp;#39; inequality&amp;lt;/strong&amp;gt;] Prove that &lt;br /&gt;
&amp;lt;math&amp;gt;&lt;br /&gt;
\sum_{i=1}^n \mathbf{Pr}(A_i) - \sum_{1 \le i&amp;lt; j \le n} \mathbf{Pr}(A_i \cap A_j)\le \mathbf{Pr}\left(\bigcup_{i=1}^n A_i\right) \le \sum_{i=1}^n \mathbf{Pr} \left( A_i\right) - \sum_{i=2}^n \mathbf{Pr}(A_1 \cap A_i).&lt;br /&gt;
&amp;lt;/math&amp;gt; &lt;br /&gt;
(Hint: This is sometimes called Kounias&#039; inequality which is weaker than the Bonferroni&#039;s inequality. You can try using Venn diagram to understand these inequalities.)&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-1-classic-examples&amp;quot;&amp;gt;Problem 2 (Symmetric 1D random walk)&amp;lt;/h2&amp;gt;&lt;br /&gt;
A gambler plays a fair gambling game: At each round, he flips a fair coin, earns &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; point if it&#039;s HEADs, and loses &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; point if otherwise.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Symmetric 1D random walk (I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;A_i&amp;lt;/math&amp;gt; be the event that the gambler earns &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; points after playing &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; rounds of the game, that is, the number of times the coin lands on heads is equal to the number of times it lands on tails. Compute &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Pr}(A_i)&amp;lt;/math&amp;gt;. (Hint: You may use Stirling&#039;s approximation to estimate [math]\mathbf{Pr}(A_i)[/math] and derive that &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Pr}(A_i) = +\infty&amp;lt;/math&amp;gt;.)&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Symmetric 1D random walk (II)&amp;lt;/strong&amp;gt;] Suppose that the game ends upon that the gambler loses all his &amp;lt;math&amp;gt;m&amp;lt;/math&amp;gt; points. Let &amp;lt;math&amp;gt;B_i&amp;lt;/math&amp;gt; be the event that the game ends within &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; rounds. Compute &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Pr}(\overline{B_i})&amp;lt;/math&amp;gt;. &lt;br /&gt;
(Hint: You may first consider [math] \displaystyle{m = 1}[/math] case. Let [math]\displaystyle{C_i}[/math] be the event that the game ends at the [math]\displaystyle{i}[/math]-th round. (i) Prove that &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Pr}(\overline{B_i}) = \sum_{i=1}^{+\infty} (i-1) \mathbf{Pr}(C_i)&amp;lt;/math&amp;gt;. (ii) Compute &amp;lt;math&amp;gt;\mathbf{Pr}(C_i)&amp;lt;/math&amp;gt;, which is a special case of the ballot problem.  (iii) Finally, use Stirling&#039;s approximation to derive that &amp;lt;math&amp;gt;\sum_{i=1}^{+\infty} \mathbf{Pr}(\overline{B_i}) = +\infty&amp;lt;/math&amp;gt;.)&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-3-probability-space&amp;quot;&amp;gt;Problem 3 (Probability space)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Nonexistence of probability space&amp;lt;/strong&amp;gt;] Prove that it is impossible to define a uniform probability law on natural numbers &amp;lt;math&amp;gt;\mathbb{N}&amp;lt;/math&amp;gt;. More precisely, prove that there does not exist a probability space &amp;lt;math&amp;gt;(\mathbb{N},2^{\mathbb{N}},\mathbf{Pr})&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;\mathbf{Pr}(\{i\}) = \mathbf{Pr}(\{j\})&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;i, j \in \mathbb{N}&amp;lt;/math&amp;gt;. &lt;br /&gt;
Please explain why the same argument fails to prove that there is no uniform probability law on the real interval &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt;, that is, there is no such probability space &amp;lt;math&amp;gt;([0,1],\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt; that for any interval &amp;lt;math&amp;gt;(l,r] \subseteq [0,1]&amp;lt;/math&amp;gt;, it holds that &amp;lt;math&amp;gt;(l,r] \in \mathcal{F}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{Pr}( (l,r] ) = r-l&amp;lt;/math&amp;gt;. (Actually, such probability measure does exist and is called the Lebesgue measure on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt;). &amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Smallest &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field&amp;lt;/strong&amp;gt;] For any subset &amp;lt;math&amp;gt;S \subseteq 2^\Omega&amp;lt;/math&amp;gt;, prove that the smallest &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field containing &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; is given by &amp;lt;math&amp;gt;\bigcap_{\substack{S \subseteq \mathcal{F} \subseteq 2^\Omega\\ \mathcal{F} \text{ is a } \sigma\text{-field }} } \mathcal{F}&amp;lt;/math&amp;gt;. (Hint: You should show that it is indeed a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field and also it is the smallest one containing &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;.)&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Union of &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\mathcal{F}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathcal{G}&amp;lt;/math&amp;gt; be &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-fields of subsets of &amp;lt;math&amp;gt;\Omega&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathcal{F} \cup \mathcal{G}&amp;lt;/math&amp;gt; is not necessarily a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Probability space?&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;\Omega = \mathbb{R}&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathcal{F}&amp;lt;/math&amp;gt; is the set of all subsets &amp;lt;math&amp;gt;A \subseteq \Omega&amp;lt;/math&amp;gt; so that &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;\overline{A}&amp;lt;/math&amp;gt; (complement of &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;) is countable, &amp;lt;math&amp;gt;P(A) = 0&amp;lt;/math&amp;gt; in the first case and &amp;lt;math&amp;gt;P(A) = 1&amp;lt;/math&amp;gt; in the second. Is &amp;lt;math&amp;gt;(\Omega,\mathcal{F},P)&amp;lt;/math&amp;gt; a probability space? Please explain your answer.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;&amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-field?&amp;lt;/strong&amp;gt;] A set &amp;lt;math&amp;gt;A \subseteq \mathbb{N}&amp;lt;/math&amp;gt; is said to have asymptotic density &amp;lt;math&amp;gt;\theta&amp;lt;/math&amp;gt; if &amp;lt;math&amp;gt;\lim_{n \to \infty} |A \cap \{1,2,\ldots,n\}| / n = \theta&amp;lt;/math&amp;gt;. Let &amp;lt;math&amp;gt;\mathcal{A}&amp;lt;/math&amp;gt; be the collection of sets for which the asymptotic density exists. Is &amp;lt;math&amp;gt;\mathcal{A}&amp;lt;/math&amp;gt; a &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-algebra? Please explain your answer.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-4-conditional-probability&amp;quot;&amp;gt;Problem 4 (Conditional probability)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Positive correlation&amp;lt;/strong&amp;gt;] We say that events &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; gives &#039;&#039;positive information&#039;&#039; about event &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; if &amp;lt;math&amp;gt;\mathbf{Pr}(A|B) &amp;gt; \mathbf{Pr}(A)&amp;lt;/math&amp;gt;, that is, the occurrence of &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; makes the occurrence of &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; more likely. Now suppose that &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; gives positive information about &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;.&lt;br /&gt;
# Does &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; give positive information about &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt;?&lt;br /&gt;
# Does &amp;lt;math&amp;gt;\overline{B}&amp;lt;/math&amp;gt; give negative information about &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;, that is, is it true that &amp;lt;math&amp;gt;\mathbf{Pr}(A|\overline{B}) &amp;lt; \mathbf{Pr}(A)&amp;lt;/math&amp;gt;?&lt;br /&gt;
# Does &amp;lt;math&amp;gt;\overline{B}&amp;lt;/math&amp;gt; give positive information or negative information about &amp;lt;math&amp;gt;\overline{A}&amp;lt;/math&amp;gt;?&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Balls in urns (I)&amp;lt;/strong&amp;gt;] There are &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; urns of which the &amp;lt;math&amp;gt;r&amp;lt;/math&amp;gt;-th contains &amp;lt;math&amp;gt;r-1&amp;lt;/math&amp;gt; white balls and &amp;lt;math&amp;gt;n-r&amp;lt;/math&amp;gt; black balls. You pick an urn uniformly at random (here, &amp;quot;uniformly&amp;quot; means that each urn has equal probability of being chosen) and remove two balls from that urn, uniformly at random without replacement (which means that each of the &amp;lt;math&amp;gt;{n-1\choose 2}&amp;lt;/math&amp;gt; pairs of balls are chosen to be removed with equal probability). Find the following probabilities:&lt;br /&gt;
# the second ball is black;&lt;br /&gt;
# the second ball is black, given that the first is black.&lt;br /&gt;
&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Balls in urns (II)&amp;lt;/strong&amp;gt;] Suppose that an urn contains &amp;lt;math&amp;gt;w&amp;lt;/math&amp;gt; white balls and &amp;lt;math&amp;gt;b&amp;lt;/math&amp;gt; black balls. The balls are drawn from the urn one by one, each time uniformly and independently at random, without replacement (which means we do not put the chosen ball back after each drawing). Find the probabilities of the events:&lt;br /&gt;
# the first white ball drawn is the &amp;lt;math&amp;gt;(k+1)&amp;lt;/math&amp;gt;th ball;&lt;br /&gt;
# the last ball drawn is white.&lt;br /&gt;
&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&#039;&#039;&#039;Noisy channel&#039;&#039;&#039;] When coded messages are sent, there are sometimes errors in transmission. In particular, Morse code uses &amp;quot;dots&amp;quot; and &amp;quot;dashes&amp;quot;, which are known to occur in the proportion of 3:4. In other words, the probability of a &amp;quot;dot&amp;quot; being sent is &amp;lt;math&amp;gt;3/7&amp;lt;/math&amp;gt;, and the probability of a &amp;quot;dash&amp;quot; being sent is &amp;lt;math&amp;gt;4/7&amp;lt;/math&amp;gt;. Suppose due to interference on the transmission channel, with probability &amp;lt;math&amp;gt;1/8&amp;lt;/math&amp;gt;, a dot will be mistakenly received as a dash, and vice versa. If we receive a dot, what is the probability that the transmitted symbol is indeed a dot? &lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-5-independence&amp;quot;&amp;gt;Problem 5 (Independence)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Let&amp;amp;#39;s consider a series of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; outputs &amp;lt;math&amp;gt;(X_1, X_2, \cdots, X_n) \in \{0,1\}^n&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; independent Bernoulli trials, where each trial succeeds with the same probability  &amp;lt;math&amp;gt;0 &amp;lt; p &amp;lt; 1&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Limited independence&amp;lt;/strong&amp;gt;] Construct three events &amp;lt;math&amp;gt;A,B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; Bernoulli trials such that &amp;lt;math&amp;gt;A, B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; are pairwise independent but are not (mutually) independent. You need to prove that the constructed events &amp;lt;math&amp;gt;A, B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; satisfy this. (Hint: Consider the case where &amp;lt;math&amp;gt;n = 2&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;p = 1/2&amp;lt;/math&amp;gt;.)&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;[&amp;lt;strong&amp;gt;Product distribution&amp;lt;/strong&amp;gt;] Suppose someone has observed the output of the &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; trials, and she told you that precisely &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; trials succeeded for some &amp;lt;math&amp;gt;0&amp;lt; k&amp;lt; n&amp;lt;/math&amp;gt;. Now you want to predict the output of the &amp;lt;math&amp;gt;(n+1)&amp;lt;/math&amp;gt;-th trial while the parameter &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; of the Bernoulli trial is unknown. One way to estimate &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; is to find such &amp;lt;math&amp;gt;\hat{p}&amp;lt;/math&amp;gt; that makes the observed outcomes most probable, namely you need to solve &lt;br /&gt;
  &amp;lt;math&amp;gt;\arg \max_{\hat{p}\in(0,1)} \mathbf{Pr}_{\hat{p}} [k \text{ out of } n\text{ trials succeed}].&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Estimate &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; by solving the above optimization problem.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;If someone tells you exactly which &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; trials succeed (in addition to just telling you the number of successful trials, which is &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;), would it help you to estimate &amp;lt;math&amp;gt;p&amp;lt;/math&amp;gt; more accurately? Why?&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;h2 id=&amp;quot;problem-6-probabilistic-method&amp;quot;&amp;gt;Problem 6 (Probabilistic method)&amp;lt;/h2&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Tournaments&amp;lt;/strong&amp;gt;] A tournament can be interpreted as the outcome of a round-robin tournament in which every player faces every other player exactly once. Given two players &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;, we draw an&lt;br /&gt;
arrow from &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; if &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; beats &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; (we say “ &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; dominates &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt;”), and we draw an arrow from &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;&lt;br /&gt;
if &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; beats &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; (there are no ties). We say a tournament has property &amp;lt;math&amp;gt;S_k&amp;lt;/math&amp;gt; if for every &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; players, there exists another player &amp;lt;math&amp;gt;v&amp;lt;/math&amp;gt; who defeats all of them. Prove that if &amp;lt;math&amp;gt;\binom n k(1-2^{-k})^{n-k}&amp;lt;1&amp;lt;/math&amp;gt;, then there  is a tournament on &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; vertices that has the property &amp;lt;math&amp;gt;S_k&amp;lt;/math&amp;gt;.&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2024)/%E7%AC%AC%E5%9B%9B%E6%AC%A1%E4%BD%9C%E4%B8%9A%E6%8F%90%E4%BA%A4%E5%90%8D%E5%8D%95&amp;diff=12526</id>
		<title>概率论与数理统计 (Spring 2024)/第四次作业提交名单</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2024)/%E7%AC%AC%E5%9B%9B%E6%AC%A1%E4%BD%9C%E4%B8%9A%E6%8F%90%E4%BA%A4%E5%90%8D%E5%8D%95&amp;diff=12526"/>
		<updated>2024-06-12T05:53:07Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: Created page with &amp;quot;如有错漏邮件请联系助教. &amp;lt;center&amp;gt; {| class=&amp;quot;wikitable&amp;quot; |- ! 学号 !! 姓名 |- | 221240092 || 杨煜申  |- | 201850065 || 严以恒  |- | 211240020 || 朱睿骐  |- | 221240035 || 李想  |- | 221240009 || 冯雨桐  |- | 221240068 || 周凡淇  |- | 221240007 || 郭宇轩  |- | 221240027 || 唐诗博  |- | 221098038 || 张浩淼  |- | 221240093 || 陈力峥  |- | 221240066 || 张植翔  |- | 201240090 || 陈诺星  |- | 221240065 || 何俊渊  |- | 221180115...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;如有错漏邮件请联系助教.&lt;br /&gt;
&amp;lt;center&amp;gt;&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! 学号 !! 姓名&lt;br /&gt;
|-&lt;br /&gt;
| 221240092 || 杨煜申 &lt;br /&gt;
|-&lt;br /&gt;
| 201850065 || 严以恒 &lt;br /&gt;
|-&lt;br /&gt;
| 211240020 || 朱睿骐 &lt;br /&gt;
|-&lt;br /&gt;
| 221240035 || 李想 &lt;br /&gt;
|-&lt;br /&gt;
| 221240009 || 冯雨桐 &lt;br /&gt;
|-&lt;br /&gt;
| 221240068 || 周凡淇 &lt;br /&gt;
|-&lt;br /&gt;
| 221240007 || 郭宇轩 &lt;br /&gt;
|-&lt;br /&gt;
| 221240027 || 唐诗博 &lt;br /&gt;
|-&lt;br /&gt;
| 221098038 || 张浩淼 &lt;br /&gt;
|-&lt;br /&gt;
| 221240093 || 陈力峥 &lt;br /&gt;
|-&lt;br /&gt;
| 221240066 || 张植翔 &lt;br /&gt;
|-&lt;br /&gt;
| 201240090 || 陈诺星 &lt;br /&gt;
|-&lt;br /&gt;
| 221240065 || 何俊渊 &lt;br /&gt;
|-&lt;br /&gt;
| 221180115 || 黄文睿 &lt;br /&gt;
|-&lt;br /&gt;
| 221900059 || 王齐剑 &lt;br /&gt;
|-&lt;br /&gt;
| 221830206 || 李君羡 &lt;br /&gt;
|-&lt;br /&gt;
| 221240056 || 郭子良 &lt;br /&gt;
|-&lt;br /&gt;
| 221870066 || 陶心悦 &lt;br /&gt;
|-&lt;br /&gt;
| 221240098 || 游书阳 &lt;br /&gt;
|-&lt;br /&gt;
| 211240056 || 赵烜呈 &lt;br /&gt;
|-&lt;br /&gt;
| 221240082 || 李岳翔 &lt;br /&gt;
|-&lt;br /&gt;
| 221240041 || 周越洋 &lt;br /&gt;
|-&lt;br /&gt;
| 221240047 || 孙梓洋 &lt;br /&gt;
|-&lt;br /&gt;
| 221240095 || 王芷若 &lt;br /&gt;
|-&lt;br /&gt;
| 201850051 || 李轲楠 &lt;br /&gt;
|-&lt;br /&gt;
| 221240012 || 吴承迅 &lt;br /&gt;
|-&lt;br /&gt;
| 221900051 || 吴子奕 &lt;br /&gt;
|-&lt;br /&gt;
| 221240022 || 韩瑞 &lt;br /&gt;
|-&lt;br /&gt;
| 221220108 || 黄嘉祺 &lt;br /&gt;
|-&lt;br /&gt;
| 221220108 || 黄嘉祺 &lt;br /&gt;
|-&lt;br /&gt;
| 221870120 || 余明晖 &lt;br /&gt;
|-&lt;br /&gt;
| 221900156 || 韩加瑞 &lt;br /&gt;
|-&lt;br /&gt;
| 221240097 || 徐川 &lt;br /&gt;
|-&lt;br /&gt;
| 215220012 || 吕茄旋 &lt;br /&gt;
|-&lt;br /&gt;
| 221900332 || 王卫东 &lt;br /&gt;
|-&lt;br /&gt;
| 221240083 || 陈正佺 &lt;br /&gt;
|-&lt;br /&gt;
| 221240002 || 季悦宁 &lt;br /&gt;
|-&lt;br /&gt;
| 221240075 || 高歌 &lt;br /&gt;
|-&lt;br /&gt;
| 221240051 || 姜钦宸 &lt;br /&gt;
|-&lt;br /&gt;
| 221240032 || 蔡坤志 &lt;br /&gt;
|-&lt;br /&gt;
| 221240060 || 熊浚丞 &lt;br /&gt;
|-&lt;br /&gt;
| 221240008 || 胡佳昕 &lt;br /&gt;
|-&lt;br /&gt;
| 221240074 || 曹任飞 &lt;br /&gt;
|-&lt;br /&gt;
| 221240033 || 陈宣宇 &lt;br /&gt;
|-&lt;br /&gt;
| 221900500 || 李亦非 &lt;br /&gt;
|-&lt;br /&gt;
| 221900045 || 郑玥轲 &lt;br /&gt;
|-&lt;br /&gt;
| 221240001 || 王炳旭 &lt;br /&gt;
|-&lt;br /&gt;
| 221240090 || 冯浚哲 &lt;br /&gt;
|-&lt;br /&gt;
| 221240023 || 蒋耀瑾 &lt;br /&gt;
|-&lt;br /&gt;
| 221240058 || 熊峥杰 &lt;br /&gt;
|-&lt;br /&gt;
| 221240040 || 郑雯琪 &lt;br /&gt;
|-&lt;br /&gt;
| 221240073 || 李恒济 &lt;br /&gt;
|-&lt;br /&gt;
| 221240070 || 汪宗元 &lt;br /&gt;
|-&lt;br /&gt;
| 221900459 || 黄培耕 &lt;br /&gt;
|-&lt;br /&gt;
| 211250032 || 吕卉森 &lt;br /&gt;
|-&lt;br /&gt;
| 221850073 || 成华睿 &lt;br /&gt;
|-&lt;br /&gt;
| 221240013 || 晁宇豪 &lt;br /&gt;
|-&lt;br /&gt;
| 221240024 || 唐之尧 &lt;br /&gt;
|-&lt;br /&gt;
| 221240084 || 史嘉诚 &lt;br /&gt;
|-&lt;br /&gt;
| 221240026 || 刘俨东 &lt;br /&gt;
|-&lt;br /&gt;
| 221250191 || 吴觐宇 &lt;br /&gt;
|-&lt;br /&gt;
| 221240063 || 欧澄垚 &lt;br /&gt;
|-&lt;br /&gt;
| 221250161 || 翟志阳 &lt;br /&gt;
|-&lt;br /&gt;
| 221900133 || 赖昱萌 &lt;br /&gt;
|-&lt;br /&gt;
| 221900006 || 耿天成 &lt;br /&gt;
|}&lt;br /&gt;
&amp;lt;/center&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2024)&amp;diff=12525</id>
		<title>概率论与数理统计 (Spring 2024)</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2024)&amp;diff=12525"/>
		<updated>2024-06-12T05:52:36Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: /* Assignments */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Infobox&lt;br /&gt;
|name         = Infobox&lt;br /&gt;
|bodystyle    = &lt;br /&gt;
|title        = &amp;lt;font size=3&amp;gt;&#039;&#039;&#039;概率论与数理统计&#039;&#039;&#039;&amp;lt;br&amp;gt;&lt;br /&gt;
&#039;&#039;&#039;Probability Theory&#039;&#039;&#039; &amp;lt;br&amp;gt; &amp;amp; &#039;&#039;&#039;Mathematical Statistics&#039;&#039;&#039;&amp;lt;/font&amp;gt;&lt;br /&gt;
|titlestyle   = &lt;br /&gt;
&lt;br /&gt;
|image        = &lt;br /&gt;
|imagestyle   = &lt;br /&gt;
|caption      = &lt;br /&gt;
|captionstyle = &lt;br /&gt;
|headerstyle  = background:#ccf;&lt;br /&gt;
|labelstyle   = background:#ddf;&lt;br /&gt;
|datastyle    = &lt;br /&gt;
&lt;br /&gt;
|header1 =Instructor&lt;br /&gt;
|label1  = &lt;br /&gt;
|data1   = &lt;br /&gt;
|header2 = &lt;br /&gt;
|label2  = &lt;br /&gt;
|data2   = &#039;&#039;&#039;尹一通&#039;&#039;&#039;&lt;br /&gt;
|header3 = &lt;br /&gt;
|label3  = Email&lt;br /&gt;
|data3   = yinyt@nju.edu.cn  &lt;br /&gt;
|header4 =&lt;br /&gt;
|label4  = office&lt;br /&gt;
|data4   = 计算机系 804&lt;br /&gt;
|header5 = &lt;br /&gt;
|label5  = &lt;br /&gt;
|data5   = &#039;&#039;&#039;刘景铖&#039;&#039;&#039;&lt;br /&gt;
|header6 = &lt;br /&gt;
|label6  = Email&lt;br /&gt;
|data6   = liu@nju.edu.cn  &lt;br /&gt;
|header7 =&lt;br /&gt;
|label7  = office&lt;br /&gt;
|data7   = 计算机系 516&lt;br /&gt;
|header8 = Class&lt;br /&gt;
|label8  = &lt;br /&gt;
|data8   = &lt;br /&gt;
|header9 =&lt;br /&gt;
|label9  = Class meeting&lt;br /&gt;
|data9   = Monday, 2pm-4pm&amp;lt;br&amp;gt;&lt;br /&gt;
Wednesday (双), 10am-12pm&amp;lt;br&amp;gt;&lt;br /&gt;
仙Ⅰ-204&lt;br /&gt;
|header10=&lt;br /&gt;
|label10 = Office hour&lt;br /&gt;
|data10  = Monday, 4pm-5pm &amp;lt;br&amp;gt;计算机系 804（尹一通）&amp;lt;br&amp;gt;计算机系 516（刘景铖）&lt;br /&gt;
|header11= Textbook&lt;br /&gt;
|label11 = &lt;br /&gt;
|data11  = &lt;br /&gt;
|header12=&lt;br /&gt;
|label12 = &lt;br /&gt;
|data12  = [[File:概率导论.jpeg|border|100px]]&lt;br /&gt;
|header13=&lt;br /&gt;
|label13 = &lt;br /&gt;
|data13  = &#039;&#039;&#039;概率导论&#039;&#039;&#039;（第2版·修订版）&amp;lt;br&amp;gt; Dimitri P. Bertsekas and John N. Tsitsiklis&amp;lt;br&amp;gt; 郑忠国 童行伟 译；人民邮电出版社 (2022)&lt;br /&gt;
|header14=&lt;br /&gt;
|label14 = &lt;br /&gt;
|data14  = [[File:Grimmett_probability.jpg|border|100px]]&lt;br /&gt;
|header15=&lt;br /&gt;
|label15 = &lt;br /&gt;
|data15  = &#039;&#039;&#039;Probability and Random Processes&#039;&#039;&#039; (4E) &amp;lt;br&amp;gt; Geoffrey Grimmett and David Stirzaker &amp;lt;br&amp;gt;  Oxford University Press (2020)&lt;br /&gt;
|header16=&lt;br /&gt;
|label16 = &lt;br /&gt;
|data16  = [[File:Probability_and_Computing_2ed.jpg|border|100px]]&lt;br /&gt;
|header17=&lt;br /&gt;
|label17 = &lt;br /&gt;
|data17  = &#039;&#039;&#039;Probability and Computing&#039;&#039;&#039; (2E) &amp;lt;br&amp;gt; Michael Mitzenmacher and Eli Upfal &amp;lt;br&amp;gt;   Cambridge University Press (2017)&lt;br /&gt;
|belowstyle = background:#ddf;&lt;br /&gt;
|below = &lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
This is the webpage for the &#039;&#039;Probability Theory and Mathematical Statistics&#039;&#039; (概率论与数理统计) class of Spring 2024. Students who take this class should check this page periodically for content updates and new announcements. &lt;br /&gt;
&lt;br /&gt;
= Announcement =&lt;br /&gt;
* TBA&lt;br /&gt;
&lt;br /&gt;
= Course info =&lt;br /&gt;
* &#039;&#039;&#039;Instructor &#039;&#039;&#039;: &lt;br /&gt;
:* [http://tcs.nju.edu.cn/yinyt/ 尹一通]：[mailto:yinyt@nju.edu.cn &amp;lt;yinyt@nju.edu.cn&amp;gt;]，计算机系 804 &lt;br /&gt;
:* [https://liuexp.github.io 刘景铖]：[mailto:liu@nju.edu.cn &amp;lt;liu@nju.edu.cn&amp;gt;]，计算机系 516 &lt;br /&gt;
* &#039;&#039;&#039;Teaching assistant&#039;&#039;&#039;:&lt;br /&gt;
** [https://sites.google.com/view/xinyuanzhang 张昕渊]：[mailto:zhangxy@smail.nju.edu.cn &amp;lt;zhangxy@smail.nju.edu.cn&amp;gt;]，计算机系 410&lt;br /&gt;
** 邹宗瑞：[mailto:zou.zongrui@smail.nju.edu.cn &amp;lt;zou.zongrui@smail.nju.edu.cn&amp;gt;]，计算机系 410&lt;br /&gt;
* &#039;&#039;&#039;Class meeting&#039;&#039;&#039;:&lt;br /&gt;
** 周一：2pm-4pm，仙Ⅰ-204&lt;br /&gt;
** 周三（双）：10am-12pm，仙Ⅰ-204&lt;br /&gt;
* &#039;&#039;&#039;Office hour&#039;&#039;&#039;: &lt;br /&gt;
:* Monday, 4pm-5pm, 计算机系 804（尹一通）&lt;br /&gt;
:* Tuesday, 3pm-4pm, 计算机系 516（刘景铖）&lt;br /&gt;
:* &#039;&#039;&#039;QQ群&#039;&#039;&#039;: 629856946（申请加入需提供姓名、院系、学号）&lt;br /&gt;
&lt;br /&gt;
= Syllabus =&lt;br /&gt;
课程内容分为三大部分：&lt;br /&gt;
* &#039;&#039;&#039;经典概率论&#039;&#039;&#039;：概率空间、随机变量及其数字特征、多维与连续随机变量、极限定理等内容&lt;br /&gt;
* &#039;&#039;&#039;概率与计算&#039;&#039;&#039;：测度集中现象 (concentration of measure)、概率法 (the probabilistic method)、离散随机过程的相关专题&lt;br /&gt;
* &#039;&#039;&#039;数理统计&#039;&#039;&#039;：参数估计、假设检验、贝叶斯估计、线性回归等统计推断等概念&lt;br /&gt;
&lt;br /&gt;
对于第一和第二部分，要求清楚掌握基本概念，深刻理解关键的现象与规律以及背后的原理，并可以灵活运用所学方法求解相关问题。对于第三部分，要求熟悉数理统计的若干基本概念，以及典型的统计模型和统计推断问题。&lt;br /&gt;
&lt;br /&gt;
经过本课程的训练，力求使学生能够熟悉掌握概率的语言，并会利用概率思维来理解客观世界并对其建模，以及驾驭概率的数学工具来分析和求解专业问题。&lt;br /&gt;
&lt;br /&gt;
=== 教材与参考书 Course Materials ===&lt;br /&gt;
* &#039;&#039;&#039;[BT]&#039;&#039;&#039; 概率导论（第2版·修订版），[美]伯特瑟卡斯（Dimitri P.Bertsekas）[美]齐齐克利斯（John N.Tsitsiklis）著，郑忠国 童行伟 译，人民邮电出版社（2022）。&lt;br /&gt;
* &#039;&#039;&#039;[GS]&#039;&#039;&#039; &#039;&#039;Probability and Random Processes&#039;&#039;, by Geoffrey Grimmett and David Stirzaker; Oxford University Press; 4th edition (2020).&lt;br /&gt;
* &#039;&#039;&#039;[MU]&#039;&#039;&#039; &#039;&#039;Probability and Computing: Randomization and Probabilistic Techniques in Algorithms and Data Analysis&#039;&#039;, by Michael Mitzenmacher, Eli Upfal; Cambridge University Press; 2nd edition (2017).&lt;br /&gt;
&lt;br /&gt;
=== 成绩 Grading Policy ===&lt;br /&gt;
* 课程成绩：本课程将会有若干次作业和一次期末考试。最终成绩将由平时作业成绩和期末考试成绩综合得出。&lt;br /&gt;
* 迟交：如果有特殊的理由，无法按时完成作业，请提前联系授课老师，给出正当理由。否则迟交的作业将不被接受。&lt;br /&gt;
&lt;br /&gt;
=== &amp;lt;font color=red&amp;gt; 学术诚信 Academic Integrity &amp;lt;/font&amp;gt;===&lt;br /&gt;
学术诚信是所有从事学术活动的学生和学者最基本的职业道德底线，本课程将不遗余力的维护学术诚信规范，违反这一底线的行为将不会被容忍。&lt;br /&gt;
&lt;br /&gt;
作业完成的原则：署你名字的工作必须是你个人的贡献。在完成作业的过程中，允许讨论，前提是讨论的所有参与者均处于同等完成度。但关键想法的执行、以及作业文本的写作必须独立完成，并在作业中致谢（acknowledge）所有参与讨论的人。符合规则的讨论与致谢将不会影响得分。不允许其他任何形式的合作——尤其是与已经完成作业的同学“讨论”。&lt;br /&gt;
&lt;br /&gt;
本课程将对剽窃行为采取零容忍的态度。在完成作业过程中，对他人工作（出版物、互联网资料、其他人的作业等）直接的文本抄袭和对关键思想、关键元素的抄袭，按照 [http://www.acm.org/publications/policies/plagiarism_policy ACM Policy on Plagiarism]的解释，都将视为剽窃。剽窃者成绩将被取消。如果发现互相抄袭行为，&amp;lt;font color=red&amp;gt; 抄袭和被抄袭双方的成绩都将被取消&amp;lt;/font&amp;gt;。因此请主动防止自己的作业被他人抄袭。&lt;br /&gt;
&lt;br /&gt;
学术诚信影响学生个人的品行，也关乎整个教育系统的正常运转。为了一点分数而做出学术不端的行为，不仅使自己沦为一个欺骗者，也使他人的诚实努力失去意义。让我们一起努力维护一个诚信的环境。&lt;br /&gt;
&lt;br /&gt;
= Assignments =&lt;br /&gt;
*[[概率论与数理统计 (Spring 2024)/Problem Set 1|Problem Set 1]]  请在 &amp;lt;strike&amp;gt;2024/4/3&amp;lt;/strike&amp;gt;&amp;lt;font color=red&amp;gt;2024/4/8&amp;lt;/font&amp;gt; 上课之前(2pm UTC+8)提交到 [mailto:pr2024_nju@163.com pr2024_nju@163.com] (文件名为&#039;&amp;lt;font color=red &amp;gt;学号_姓名_A1.pdf&amp;lt;/font&amp;gt;&#039;). &lt;br /&gt;
** [[概率论与数理统计 (Spring 2024)/第一次作业提交名单|第一次作业提交名单]]&lt;br /&gt;
*[[概率论与数理统计 (Spring 2024)/Problem Set 2|Problem Set 2]]  请在 &amp;lt;strike&amp;gt;2024/4/29&amp;lt;/strike&amp;gt;&amp;lt;font color=red&amp;gt;2024/5/6&amp;lt;/font&amp;gt; 上课之前(2pm UTC+8)提交到 [mailto:pr2024_nju@163.com pr2024_nju@163.com] (文件名为&#039;&amp;lt;font color=red &amp;gt;学号_姓名_A2.pdf&amp;lt;/font&amp;gt;&#039;).&lt;br /&gt;
** [[概率论与数理统计 (Spring 2024)/第二次作业提交名单|第二次作业提交名单]]&lt;br /&gt;
*[[概率论与数理统计 (Spring 2024)/Problem Set 3|Problem Set 3]]  请在 2024/5/20 上课之前(2pm UTC+8)提交到 [mailto:pr2024_nju@163.com pr2024_nju@163.com] (文件名为&#039;&amp;lt;font color=red &amp;gt;学号_姓名_A3.pdf&amp;lt;/font&amp;gt;&#039;).&lt;br /&gt;
** [[概率论与数理统计 (Spring 2024)/第三次作业提交名单|第三次作业提交名单]]&lt;br /&gt;
*[[概率论与数理统计 (Spring 2024)/Problem Set 4|Problem Set 4]]  请在 2024/6/12 上课之前(10am UTC+8)提交到 [mailto:pr2024_nju@163.com pr2024_nju@163.com] (文件名为&#039;&amp;lt;font color=red &amp;gt;学号_姓名_A4.pdf&amp;lt;/font&amp;gt;&#039;).&lt;br /&gt;
** [[概率论与数理统计 (Spring 2024)/第四次作业提交名单|第四次作业提交名单]]&lt;br /&gt;
&lt;br /&gt;
= Lectures =&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2024/Intro.pdf 课程简介]&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2024/ProbSpace.pdf 概率空间]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 第1章&#039;&#039;&#039; 或 &#039;&#039;&#039;[GS] Chapter 1&#039;&#039;&#039;&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2024)/Karger&#039;s min-cut algorithm| Karger&#039;s min-cut algorithm]]&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2024/RandVar.pdf 随机变量]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 第2章&#039;&#039;&#039; 或 &#039;&#039;&#039;[GS] Chapter 2, Sections 3.1~3.5, 3.7&#039;&#039;&#039;&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[MU] Chapter 2&#039;&#039;&#039;&lt;br /&gt;
#* the [http://tcs.nju.edu.cn/slides/prob2024/discrete-pmf.nb &#039;&#039;&#039;Mathematica notebook file&#039;&#039;&#039;] for the PMFs of basic discrete distributions&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2024)/Average-case analysis of QuickSort|Average-case analysis of &#039;&#039;&#039;&#039;&#039;QuickSort&#039;&#039;&#039;&#039;&#039;]]&lt;br /&gt;
#* [https://www.bilibili.com/video/BV1ta411A7fp/ 高尔顿板（Galton board）视频] 和 [https://en.wikipedia.org/wiki/Galton_board 维基百科页面]&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2024/Deviation.pdf 矩与偏差]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[MU] Chapter 3&#039;&#039;&#039;&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 章节 2.4, 4.2, 4.3, 5.1&#039;&#039;&#039; 或 &#039;&#039;&#039;[GS] Sections 3.3, 3.6, 7.3&#039;&#039;&#039;&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2024)/Two-point sampling|Two-point sampling]]&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2024)/Threshold of k-clique in random graph|Threshold of &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;-clique in random graph]]&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2024)/Weierstrass Approximation Theorem|Weierstrass approximation]]&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2024/Continuous.pdf 连续分布]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 第3章, 和4.1节&#039;&#039;&#039; 或 &#039;&#039;&#039;[GS] Chapter 4&#039;&#039;&#039;&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[MU] Chapters 8, 9&#039;&#039;&#039;&lt;br /&gt;
#* [https://measure.axler.net/MIRA.pdf Measure, Integration &amp;amp; Real Analysis] by Sheldon Axler&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2024/Convergence.pdf 极限定理]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 第5章&#039;&#039;&#039; &lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[GS] Sections 5.7~5.10, 7.1~7.5&#039;&#039;&#039;&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2024/Concentration.pdf 测度集中]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[MU] Chapters 4&#039;&#039;&#039; and &#039;&#039;&#039;Sections 13.1, 13.4~13.5&#039;&#039;&#039;&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[GS] Sections 5.11, 12.1~12.3, 7.8~7.9&#039;&#039;&#039;&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2024)/Hoeffding&#039;s lemma|Hoeffding&#039;s lemma]]&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2024)/Entropy and volume of Hamming balls|Entropy and volume of Hamming balls]]&lt;br /&gt;
# [http://tcs.nju.edu.cn/slides/prob2024/Process.pdf 随机过程]&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[BT] 第6章, 第7章&#039;&#039;&#039;&lt;br /&gt;
#* 阅读：&#039;&#039;&#039;[MU]  Chapters 7, Sections 13.1~13.3&#039;&#039;&#039; or &#039;&#039;&#039;[GS] Chapters 6, Sections 12.4~12.5&#039;&#039;&#039;&lt;br /&gt;
#* [[概率论与数理统计 (Spring 2024)/OST and applications|OST and applications]]&lt;br /&gt;
#[[Media:Stat01-2024.pdf |统计、推断与点估计]]、 [[Media:Stat02-2024.pdf |假设检验]]&lt;br /&gt;
&lt;br /&gt;
= Concepts =&lt;br /&gt;
* [https://plato.stanford.edu/entries/probability-interpret/ Interpretations of probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/History_of_probability History of probability]&lt;br /&gt;
* Example problems:&lt;br /&gt;
** [https://dornsifecms.usc.edu/assets/sites/520/docs/VonNeumann-ams12p36-38.pdf von Neumann&#039;s Bernoulli factory] and other [https://peteroupc.github.io/bernoulli.html Bernoulli factory algorithms]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Boy_or_Girl_paradox Boy or Girl paradox]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Monty_Hall_problem Monty Hall problem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bertrand_paradox_(probability) Bertrand paradox]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hard_spheres Hard spheres model] and [https://en.wikipedia.org/wiki/Ising_model Ising model]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/PageRank &#039;&#039;PageRank&#039;&#039;] and stationary [https://en.wikipedia.org/wiki/Random_walk random walk]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Diffusion_process Diffusion process] and [https://en.wikipedia.org/wiki/Diffusion_model diffusion model]&lt;br /&gt;
*[https://en.wikipedia.org/wiki/Probability_space Probability space]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Sample_space Sample space]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Event_(probability_theory) Event] and [https://en.wikipedia.org/wiki/Σ-algebra &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt;-algebra]&lt;br /&gt;
** Kolmogorov&#039;s [https://en.wikipedia.org/wiki/Probability_axioms axioms of probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Discrete_uniform_distribution Classical] and [https://en.wikipedia.org/wiki/Geometric_probability goemetric probability]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Boole%27s_inequality Union bound]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Inclusion%E2%80%93exclusion_principle Inclusion-Exclusion principle]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Boole%27s_inequality#Bonferroni_inequalities Bonferroni inequalities]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Conditional_probability Conditional probability]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chain_rule_(probability) Chain rule]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_total_probability Law of total probability]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bayes%27_theorem Bayes&#039; law]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Independence_(probability_theory) Independence] &lt;br /&gt;
** [https://en.wikipedia.org/wiki/Pairwise_independence Pairwise independence]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Random_variable Random variable]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cumulative_distribution_function Cumulative distribution function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_mass_function Probability mass function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_density_function Probability density function]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Multivariate_random_variable Random vector]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Joint_probability_distribution Joint probability distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Conditional_probability_distribution Conditional probability distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Marginal_distribution Marginal distribution]&lt;br /&gt;
* Some &#039;&#039;&#039;discrete&#039;&#039;&#039; probability distributions&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Bernoulli_trial Bernoulli trial] and [https://en.wikipedia.org/wiki/Bernoulli_distribution Bernoulli distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Discrete_uniform_distribution Discrete uniform distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Binomial_distribution Binomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Geometric_distribution Geometric distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Negative_binomial_distribution Negative binomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Hypergeometric_distribution Hypergeometric distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Poisson_distribution Poisson distribution]&lt;br /&gt;
** and [https://en.wikipedia.org/wiki/List_of_probability_distributions#Discrete_distributions others]&lt;br /&gt;
* Balls into bins model&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Multinomial_distribution Multinomial distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Birthday_problem Birthday problem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Coupon_collector%27s_problem Coupon collector]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Balls_into_bins_problem Occupancy problem]&lt;br /&gt;
* Random graphs&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Erd%C5%91s%E2%80%93R%C3%A9nyi_model Erdős–Rényi random graph model]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Galton%E2%80%93Watson_process Galton–Watson branching process]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Expected_value Expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_the_unconscious_statistician Law of the unconscious statistician, &#039;&#039;LOTUS&#039;&#039;]&lt;br /&gt;
** [https://dlsun.github.io/probability/linearity.html Linearity of expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Conditional_expectation Conditional expectation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Law_of_total_expectation Law of total expectation]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Markov%27s_inequality Markov&#039;s inequality]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Chebyshev%27s_inequality Chebyshev&#039;s inequality]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Moment_(mathematics) Moment] related&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Central_moment Central moment]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Variance Variance]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Standard_deviation Standard deviation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Covariance Covariance]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Correlation Correlation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Correlation_does_not_imply_causation Correlation does not imply causation]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Skewness Skewness]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Kurtosis Kurtosis]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Moment_problem Moment problem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Measure_(mathematics) Measure]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Borel_set Borel set]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Lebesgue_integration Lebesgue integration]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Measurable_function Measurable function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cantor_set Cantor set]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Non-measurable_set Non-measurable set]&lt;br /&gt;
*[https://en.wikipedia.org/wiki/Probability_density_function Probability density function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Probability_distribution#Absolutely_continuous_probability_distribution Continuous probability distribution]&lt;br /&gt;
**[https://en.wikipedia.org/wiki/Conditional_probability_distribution#Conditional_continuous_distributions Conditional continuous distributions]&lt;br /&gt;
**[https://en.wikipedia.org/wiki/Convolution_of_probability_distributions Convolution of probability distributions] and [https://en.wikipedia.org/wiki/List_of_convolutions_of_probability_distributions List of convolutions of probability distributions]&lt;br /&gt;
* Some &#039;&#039;&#039;continuous&#039;&#039;&#039; probability distributions&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Continuous_uniform_distribution Continuous uniform distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Exponential_distribution Exponential distribution] and [https://en.wikipedia.org/wiki/Poisson_point_process Poisson point process]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Normal_distribution Normal (Gaussion) distribution]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Gaussian_function Gaussian function] and [https://en.wikipedia.org/wiki/Gaussian_integral Gaussian integral]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Standard_normal_table Standard normal table]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/68%E2%80%9395%E2%80%9399.7_rule 68–95–99.7 rule]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Sum_of_normally_distributed_random_variables Sum of normally distributed random variables]&lt;br /&gt;
*** [https://en.wikipedia.org/wiki/Multivariate_normal_distribution Multivariate normal distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Chi-squared_distribution Chi-squared distribution]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Gamma_distribution Gamma distribution] and [https://en.wikipedia.org/wiki/Gamma_function Gamma function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Beta_distribution Beta distribution] and [https://en.wikipedia.org/wiki/Beta_function Beta function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Cauchy_distribution Cauchy distribution]&lt;br /&gt;
** and [https://en.wikipedia.org/wiki/List_of_probability_distributions#Absolutely_continuous_distributions others]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Inverse_transform_sampling Inverse transform sampling]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Stochastic_dominance Stochastic dominance] and [https://en.wikipedia.org/wiki/Coupling_(probability) Coupling] &lt;br /&gt;
* [https://en.wikipedia.org/wiki/Moment-generating_function Moment-generating function]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Convergence_of_random_variables Convergence of random variables]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Pointwise_convergence Pointwise convergence]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Convergence_of_measures#Weak_convergence_of_measures Weak convergence of measures]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Convergence_in_measure Convergence in measure]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Skorokhod%27s_representation_theorem Skorokhod&#039;s representation theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Continuous_mapping_theorem Continuous mapping theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Borel%E2%80%93Cantelli_lemma Borel–Cantelli lemma] and [https://en.wikipedia.org/wiki/Zero%E2%80%93one_law zero-one laws]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Law_of_large_numbers Law of large numbers]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Central_limit_theorem Central limit theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/De_Moivre%E2%80%93Laplace_theorem De Moivre–Laplace theorem]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Berry%E2%80%93Esseen_theorem Berry–Esseen theorem]&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Characteristic_function_(probability_theory) Characteristic function]&lt;br /&gt;
** [https://en.wikipedia.org/wiki/Lévy%27s_continuity_theorem Lévy&#039;s continuity theorem]&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
	<entry>
		<id>https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2024)/Problem_Set_4&amp;diff=12443</id>
		<title>概率论与数理统计 (Spring 2024)/Problem Set 4</title>
		<link rel="alternate" type="text/html" href="https://tcs.nju.edu.cn/wiki/index.php?title=%E6%A6%82%E7%8E%87%E8%AE%BA%E4%B8%8E%E6%95%B0%E7%90%86%E7%BB%9F%E8%AE%A1_(Spring_2024)/Problem_Set_4&amp;diff=12443"/>
		<updated>2024-05-24T13:24:13Z</updated>

		<summary type="html">&lt;p&gt;Zouzongrui: /* Problem 2 (Concentration of measure, 10 points) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;*每道题目的解答都要有完整的解题过程，中英文不限。&lt;br /&gt;
&lt;br /&gt;
*我们推荐大家使用LaTeX, markdown等对作业进行排版。&lt;br /&gt;
&lt;br /&gt;
*Bonus problem为附加题（选做）。&lt;br /&gt;
&lt;br /&gt;
== Assumption throughout Problem Set 4==&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we are working on probability space &amp;lt;math&amp;gt;(\Omega,\mathcal{F},\mathbf{Pr})&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;Without further notice, we assume that the expectation of random variables are well-defined.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;p&amp;gt;The term &amp;lt;math&amp;gt;\log&amp;lt;/math&amp;gt; used in this context refers to the natural logarithm.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 1 (Continuous Random Variables, 30 points)==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Density function&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Determine the value of &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;f(x) = C\exp(-x-e^{-x}), x\in \mathbb{R}&amp;lt;/math&amp;gt; is a probability density function (PDF) for a continuous random variable.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Independence&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be independent and identically distributed continuous random variables with cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; and probability density function (PDF) &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt;. Find out the density functions of &amp;lt;math&amp;gt;V = \max\{X,Y\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;U = \min\{X,Y\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Correlation&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be uniformly distributed on &amp;lt;math&amp;gt;(-1,1)&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_k = \cos(k \pi X)&amp;lt;/math&amp;gt; for &amp;lt;math&amp;gt;k=1,2,\ldots,n&amp;lt;/math&amp;gt;. Are the random variables &amp;lt;math&amp;gt;Y_1, Y_2, \ldots, Y_n&amp;lt;/math&amp;gt; correlated? independent? You should prove your claim rigorously.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Expectation of random variables (I)&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a continuous random variable with mean &amp;lt;math&amp;gt;\mu&amp;lt;/math&amp;gt; and cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X \ge 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\int_{0}^a F(x) dx = \int_{a}^{\infty} [1-F(x)] dx&amp;lt;/math&amp;gt; if and only if &amp;lt;math&amp;gt;a = \mu&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Suppose &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; has finite variance. Show that &amp;lt;math&amp;gt;g(a) = \mathbb{E}((X-a)^2)&amp;lt;/math&amp;gt; achieves the minimum when &amp;lt;math&amp;gt;a = \mu&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Expectation of random variables (II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X, Y&amp;lt;/math&amp;gt; be two independent and identically distributed continuous random variables with cumulative distribution function (CDF) &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;. Furthermore, &amp;lt;math&amp;gt;X,Y \ge 0&amp;lt;/math&amp;gt;. Show that &amp;lt;math&amp;gt;\mathbb{E}[|X-Y|] = 2 \left(\mathbb{E}[X] - \int_{0}^{\infty} (1-F(x))^2 dx\right)&amp;lt;/math&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Conditional distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; be two random variables. The joint density of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; is given by &amp;lt;math&amp;gt;f(x,y) = c(x^2 - y^2)e^{-x}&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;0\leq x &amp;lt;\infty&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;-x\leq y \leq x&amp;lt;/math&amp;gt;. Here, &amp;lt;math&amp;gt;c\in \mathbb{R}_+&amp;lt;/math&amp;gt; is a constant. Find out the conditional distribution of &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt;, given &amp;lt;math&amp;gt;X = x&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Uniform Distribution (I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;P_i = (X_i,Y_i), 1\leq i\leq n&amp;lt;/math&amp;gt;, be independent, uniformly distributed points in the unit square &amp;lt;math&amp;gt;[0,1]^2&amp;lt;/math&amp;gt;. A point &amp;lt;math&amp;gt;P_i&amp;lt;/math&amp;gt; is called &amp;quot;peripheral&amp;quot; if, for all &amp;lt;math&amp;gt;r = 1,2,\cdots,n&amp;lt;/math&amp;gt;, either &amp;lt;math&amp;gt;X_r \leq X_i&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;Y_r \leq Y_i&amp;lt;/math&amp;gt;, or both. Find out the expected number of peripheral points.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Uniform Distribution (II)&amp;lt;/strong&amp;gt;] Derive the moment generating function of the standard uniform distribution, i.e., uniform distribution on &amp;lt;math&amp;gt;(0,1)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Exponential distribution&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; have an exponential distribution. Show that &amp;lt;math&amp;gt;\textbf{Pr}[X&amp;gt;s+x|X&amp;gt;s] = \textbf{Pr}[X&amp;gt;x]&amp;lt;/math&amp;gt;, for &amp;lt;math&amp;gt;x,s\geq 0&amp;lt;/math&amp;gt;. This is the memoryless property. Show that the exponential distribution is the only continuous distribution with this property.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Normal distribution(I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X,Y\sim N(0,1)&amp;lt;/math&amp;gt; be two independent and identically distributed normal random variables. Let &amp;lt;math&amp;gt;Z = X-Y&amp;lt;/math&amp;gt;. Find the density function of &amp;lt;math&amp;gt;Z&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;|Z|&amp;lt;/math&amp;gt; respectively.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Normal distribution(II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; have the &amp;lt;math&amp;gt;N(0,1)&amp;lt;/math&amp;gt; distribution and let &amp;lt;math&amp;gt;a&amp;gt;0&amp;lt;/math&amp;gt;. Show that the random variable &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; given by&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{equation*}&lt;br /&gt;
Y = \begin{cases}&lt;br /&gt;
X, &amp;amp; |X|&amp;lt; a \\&lt;br /&gt;
-X, &amp;amp; |X|\geq a&lt;br /&gt;
\end{cases}&lt;br /&gt;
\end{equation*}&amp;lt;/math&amp;gt;&lt;br /&gt;
has the &amp;lt;math&amp;gt;N(0,1)&amp;lt;/math&amp;gt; distribution, and find an expression for &amp;lt;math&amp;gt;\rho(a) = \textbf{Cov}(X,Y)&amp;lt;/math&amp;gt; in terms of the density function &amp;lt;math&amp;gt;\phi&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;Random process&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Given a real number &amp;lt;math&amp;gt;U&amp;lt;1&amp;lt;/math&amp;gt; as input of the following process, find out the expected returning value.&lt;br /&gt;
{{Theorem|&#039;&#039;Process&#039;&#039;|&lt;br /&gt;
:&#039;&#039;&#039;Input:&#039;&#039;&#039;  real numbers &amp;lt;math&amp;gt;U &amp;lt; 1&amp;lt;/math&amp;gt;;&lt;br /&gt;
----&lt;br /&gt;
:initialize &amp;lt;math&amp;gt;x = 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;count = 0&amp;lt;/math&amp;gt;;&lt;br /&gt;
:while &amp;lt;math&amp;gt; x &amp;lt; U &amp;lt;/math&amp;gt; do&lt;br /&gt;
:* choose &amp;lt;math&amp;gt;y \in (0,1)&amp;lt;/math&amp;gt; uniformly at random;&lt;br /&gt;
:* update &amp;lt;math&amp;gt;x = x + y&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;count = count + 1&amp;lt;/math&amp;gt;;&lt;br /&gt;
:return &amp;lt;math&amp;gt;count&amp;lt;/math&amp;gt;;&lt;br /&gt;
}}&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Random semicircle&amp;lt;/strong&amp;gt;] We sample &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; points within a circle &amp;lt;math&amp;gt;C=\{(x,y) \in \mathbb{R}^2 \mid x^2+y^2 \le 1\}&amp;lt;/math&amp;gt; independently and uniformly at random (i.e., the density function &amp;lt;math&amp;gt;f(x,y) \propto 1_{(x,y) \in C}&amp;lt;/math&amp;gt;). Find out the probability that they all lie within some semicircle of the original circle &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;. (Hint: you may apply the technique of change of variables, see [https://en.wikipedia.org/wiki/Random_variable#Functions_of_random_variables function of random variables] or Chapter 4.7 in [GS])&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Stochastic domination&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X, Y&amp;lt;/math&amp;gt; be continuous random variables. Show that &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; dominates &amp;lt;math&amp;gt;Y&amp;lt;/math&amp;gt; stochastically if and only if &amp;lt;math&amp;gt;\mathbb{E}[f(X)]\geq \mathbb{E}[f(Y)]&amp;lt;/math&amp;gt; for any non-decreasing function &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; for which the expectations exist.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 2 (Concentration of measure, 10 points) ==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Tossing coins&amp;lt;/strong&amp;gt;] We repeatedly toss a fair coin (with an equal probability of heads and tails). Let the random variable &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of throws required to obtain a total of &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; heads. Show that &amp;lt;math&amp;gt;\textbf{Pr}[X &amp;gt; 2n + 2\sqrt{n\log n}]\leq O(1/n)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Chernoff vs Chebyshev&amp;lt;/strong&amp;gt;] We have a standard six-sided die. Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be the number of times a 6 occurs in &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; throws off the die. Compare the best upper bounds on &amp;lt;math&amp;gt;\textbf{Pr}[X\geq n/4]&amp;lt;/math&amp;gt; that you can obtain using Chebyshev&#039;s inequality and Chernoff bounds.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;&amp;lt;strong&amp;gt;-th moment bound]&amp;lt;/strong&amp;gt; Let &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; be a random variable with expectation &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; such that moment generating function &amp;lt;math&amp;gt;\mathbf{E}[\exp(t|X|)]&amp;lt;/math&amp;gt; is finite for some &amp;lt;math&amp;gt; t &amp;gt; 0 &amp;lt;/math&amp;gt;. We can use the following two kinds of tail inequalities for &amp;lt;math&amp;gt; X &amp;lt;/math&amp;gt;: &lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&amp;lt;nowiki&amp;gt;     &amp;lt;/nowiki&amp;gt;&#039;&#039;&#039;&#039;&#039;Chernoff Bound&#039;&#039;&#039;&#039;&#039;&lt;br /&gt;
:&amp;lt;math&amp;gt;&lt;br /&gt;
\begin{align}&lt;br /&gt;
\mathbf{Pr}[|X| \geq \delta] \leq \min_{t \geq 0} \frac{\mathbf{E}[e^{t|X|}]}{e^{t\delta}}&lt;br /&gt;
\end{align}&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&#039;&#039;&#039;&#039;&#039;&amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;th-Moment Bound&#039;&#039;&#039;&#039;&#039;&lt;br /&gt;
:&amp;lt;math&amp;gt;&lt;br /&gt;
\begin{align}&lt;br /&gt;
\mathbf{Pr}[|X| \geq \delta] \leq \frac{\mathbf{E}[|X|^k]}{\delta^k}&lt;br /&gt;
\end{align}&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
# Show that for each &amp;lt;math&amp;gt;\delta&amp;lt;/math&amp;gt;, there exists a choice of &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; such that the &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;th-moment bound is no weaker than the Chernoff bound. (Hint: Use the probabilistic method.)&lt;br /&gt;
# Why would we still prefer the Chernoff bound to the (seemingly) stronger &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;-th moment bound?&lt;br /&gt;
&lt;br /&gt;
* [&#039;&#039;&#039;Cut size in random graph&#039;&#039;&#039;] Show that with probability at least &amp;lt;math&amp;gt;2/3&amp;lt;/math&amp;gt;, the size of the max-cut in [[wikipedia:Erdős–Rényi_model#Definition|Erdős–Rényi random graph]] &amp;lt;math&amp;gt;G(n,1/2)&amp;lt;/math&amp;gt; is at most &amp;lt;math&amp;gt;n^2/8 + O(n^{1.5})&amp;lt;/math&amp;gt;. In the &amp;lt;math&amp;gt;G(n,1/2)&amp;lt;/math&amp;gt; model, each edge is included in the graph with probability &amp;lt;math&amp;gt;1/2&amp;lt;/math&amp;gt;, independently of every other edge.&lt;br /&gt;
&lt;br /&gt;
== Problem 3 (Modes of Convergence, 15 points) (&amp;lt;strong&amp;gt;Bonus problem&amp;lt;/strong&amp;gt;)==&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Connection of convergence modes (I)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;(X_n)_{n \ge 1}, (Y_n)_{n \ge 1}, X, Y&amp;lt;/math&amp;gt; be random variables and &amp;lt;math&amp;gt;c\in\mathbb{R}&amp;lt;/math&amp;gt; be a real number. &lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt; Suppose &amp;lt;math&amp;gt;X_n \overset{D}{\to} X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_n \overset{D}{\to} c&amp;lt;/math&amp;gt;. Prove that &amp;lt;math&amp;gt;X_nY_n \overset{D}{\to} cX&amp;lt;/math&amp;gt;. &lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Construct an example such that &amp;lt;math&amp;gt;X_n \overset{D}{\to} X&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Y_n \overset{D}{\to} Y&amp;lt;/math&amp;gt; but &amp;lt;math&amp;gt;X_nY_n&amp;lt;/math&amp;gt; does not converge to &amp;lt;math&amp;gt;XY&amp;lt;/math&amp;gt; in distribution.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; [&amp;lt;strong&amp;gt;Connection of convergence modes (II)&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;(X_n)_{n \ge 1}, X&amp;lt;/math&amp;gt; be random variables. Prove that &amp;lt;math&amp;gt;X_n \overset{P}{\to} X&amp;lt;/math&amp;gt; if and only if for every subsequence &amp;lt;math&amp;gt;X_{n(m)}&amp;lt;/math&amp;gt;, there exists a further subsequence &amp;lt;math&amp;gt;Y_k = X_{n(m_k)}&amp;lt;/math&amp;gt; that converges almost surely to &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt;. (Hint: you may use the first Borel-Cantelli lemma.)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Extension of Borel-Cantelli Lemma&amp;lt;/strong&amp;gt;]&lt;br /&gt;
Let &amp;lt;math&amp;gt;(A_n)_{n \ge 1}&amp;lt;/math&amp;gt; be events. Suppose &amp;lt;math&amp;gt;\sum_{n \ge 1} \mathbf{Pr}(A_n)=+\infty&amp;lt;/math&amp;gt;. Show that &lt;br /&gt;
&amp;lt;math&amp;gt;\mathbf{Pr}(A_n \text{ i.o.}) \ge \limsup_{n \to \infty} \frac{ \left(\sum_{k=1}^n\mathbf{Pr}(A_k)\right)^2 }{\sum_{1\le j,k \le n} \mathbf{Pr}(A_j \cap A_k)}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Problem 4 (LLN and CLT, 15 points + 5 points) ==&lt;br /&gt;
&amp;lt;strong&amp;gt;In this problem, you may apply the results of Laws of Large Numbers (LLN) and the Central Limit Theorem (CLT) to solve the problems.&amp;lt;/strong&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;[&amp;lt;strong&amp;gt;St. Petersburg paradox&amp;lt;/strong&amp;gt;] Consider the well-known game involving a fair coin. In this game, if it takes &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; tosses to obtain a head, you will win &amp;lt;math&amp;gt;2^k&amp;lt;/math&amp;gt; dollars as the reward. Despite the game&#039;s expected reward being infinite, people tend to offer relatively modest amounts to participate. The following provides a mathematical explanation for this phenomenon.&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
For each &amp;lt;math&amp;gt;n \ge 1&amp;lt;/math&amp;gt;, let &amp;lt;math&amp;gt;X_{n,1}, X_{n,2},\ldots, X_{n,k}&amp;lt;/math&amp;gt; be independent random variables. Furthermore, let &amp;lt;math&amp;gt;b_n &amp;gt; 0&amp;lt;/math&amp;gt; be real numbers with &amp;lt;math&amp;gt;b_n \to \infty&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\widetilde{X}_{n,k} = X_{n,k} \mathbf{1}_{|X_{n,k}| \le b_n}&amp;lt;/math&amp;gt; for all &amp;lt;math&amp;gt;1 \le k \le n&amp;lt;/math&amp;gt;. If &amp;lt;math&amp;gt;\sum_{k=1}^n \mathbf{Pr}(|X_{n,k}| &amp;gt; b_n) \to 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;b_n^{-2} \sum_{k=1}^n \mathbf{E}[\widetilde{X}_{n,k}^2] \to 0&amp;lt;/math&amp;gt; when &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, then &amp;lt;math&amp;gt;(S_n-a_n)/b_n \overset{P}{\to} 0 &amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;S_n = \sum_{k=1}^n X_{n,k}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;a_n = \sum_{k=1}^n \mathbf{E}[\widetilde{X}_{n,k}]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
Let &amp;lt;math&amp;gt;S_n&amp;lt;/math&amp;gt; be the total winnings after playing &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds of the game. Prove that &amp;lt;math&amp;gt;\frac{S_n}{n \log_2 n} \overset{P}{\to} 1&amp;lt;/math&amp;gt;. (Therefore, a fair price to play this game &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; times is roughly &amp;lt;math&amp;gt;n \log_2 n&amp;lt;/math&amp;gt; dollars)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt; (&amp;lt;strong&amp;gt;Bonus problem, 5 points&amp;lt;/strong&amp;gt;)&lt;br /&gt;
Let &amp;lt;math&amp;gt;S_n&amp;lt;/math&amp;gt; be the total winnings after playing &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; rounds of the game. Prove that &amp;lt;math&amp;gt; \limsup_{n \to \infty} \frac{S_n}{n \log_2 n} = \infty&amp;lt;/math&amp;gt; almost surely. (Hint: You may use Borel-Cantelli lemmas)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Monte Carlo Integration&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; be a continuous function on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;U_1,U_2,\ldots,U_n \sim U(0,1)&amp;lt;/math&amp;gt; be independent random variables. Show that &amp;lt;math&amp;gt;I = \frac{1}{n}\sum_{i=1}^n f(U_i) \to \int_0^1 f(x) \mathrm{d}x&amp;lt;/math&amp;gt; in probability. (Remark: This holds so long as &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; is a measurable function on &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;\int_0^1 |f(x)| \mathrm{d}x&amp;lt;/math&amp;gt;)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;li&amp;gt;&lt;br /&gt;
[&amp;lt;strong&amp;gt;Square root&amp;lt;/strong&amp;gt;] Let &amp;lt;math&amp;gt;X_1,X_2,\ldots&amp;lt;/math&amp;gt; be i.i.d. random variables with &amp;lt;math&amp;gt;X_i \ge 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\mathbf{E}[X_1] = 1&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{Var}[X_1] = \sigma^2 \in (0,+\infty)&amp;lt;/math&amp;gt;. Show &amp;lt;math&amp;gt;\sqrt{S_n} - \sqrt{n} \overset{D}{\to} N(0,1)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \to \infty&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;S_n = \sum_{i=1}^n X_i&amp;lt;/math&amp;gt;. (Hint: You may use the statement in Problem 3.)&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;/div&gt;</summary>
		<author><name>Zouzongrui</name></author>
	</entry>
</feed>