๐Ÿ›๏ธ ISI Advanced Examination Practice

JRF (STA) โ€” Model Solutions (2026)

Subject Level

Rigorous, step-by-step descriptive proofs for the Indian Statistical Institute Junior Research Fellowship.

๐Ÿ“Œ Q1 Analysis of Recursive Sequences (12 Marks)

๐Ÿง  Approach & Key Concepts

To determine the convergence of a sequence defined by a recurrence relation, we use the Monotone Convergence Theorem, which states that any sequence that is both monotonic and bounded is convergent. The strategy involves:

โœ๏ธ Step-by-Step Proof / Derivation

Step 1: Establishing Positivity and Lower Bound

Given $x_1 = 1$ and $x_{n+1} = \frac{1}{2}\left(x_n + \frac{2}{x_n}\right)$. Since $x_1 > 0$, by induction, if $x_n > 0$, then $x_{n+1} > 0$ for all $n \in \mathbb{N}$. Applying the AM-GM Inequality for $n \geq 1$:

$$ x_{n+1} = \frac{x_n + \frac{2}{x_n}}{2} \geq \sqrt{x_n \cdot \frac{2}{x_n}} = \sqrt{2} $$

This shows that for all $n \geq 2$, the sequence is bounded below by $\sqrt{2}$.

Step 2: Proving Monotonicity

We examine the difference between successive terms $x_{n+1}$ and $x_n$ for $n \geq 2$. Since we established $x_n \geq \sqrt{2}$ for $n \geq 2$, it follows that $x_n^2 \geq 2$.

$$ x_{n+1} - x_n = \frac{1}{2}\left(x_n + \frac{2}{x_n}\right) - x_n = \frac{2 - x_n^2}{2x_n} $$

Because $x_n^2 \geq 2$ and $x_n > 0$, the numerator $2 - x_n^2 \leq 0$. Thus, $x_{n+1} \leq x_n$ for all $n \geq 2$, proving the sequence is monotonically decreasing.

Step 3: Applying the Monotone Convergence Theorem

Since the sequence $\{x_n\}$ is monotonically decreasing (for $n \geq 2$) and bounded below by $\sqrt{2}$, by the Monotone Convergence Theorem, the sequence converges to a finite limit $L$.

$$ \lim_{n \to \infty} x_n = L $$

Step 4: Finding the Limit and Limit Inferior

Taking the limit on both sides of the recurrence relation:

$$ L = \frac{1}{2}\left(L + \frac{2}{L}\right) \implies 2L^2 = L^2 + 2 \implies L^2 = 2 $$

Since $x_n > 0$ for all $n$, we must have $L = \sqrt{2}$. For a convergent sequence, the limit inferior is equal to the limit.

Final Answer / Q.E.D: The sequence $\{x_n\}$ is convergent, and $\liminf_{n \to \infty} x_n = \sqrt{2}$.

๐Ÿ“Œ Q2 Continuous Functions and Suprema in $\mathbb{R}^d$ (12 Marks)

Problem Statement: Suppose $f : \mathbb{R}^d \mapsto \mathbb{R}$ is a continuous function with a unique maxima at some $\mathbf{x}_0 \in \mathbb{R}^d$. Then, show that for any $\delta > 0$, $$ f(\mathbf{x}_0) > \sup\{f(\mathbf{x}) : \|\mathbf{x} - \mathbf{x}_0\| \geq \delta\} $$ Here, $\|\cdot\|$ denotes the Euclidean norm on $\mathbb{R}^d$.

๐Ÿง  Approach & Key Concepts

This problem tests the interaction between continuous functions, suprema, and the topological properties of $\mathbb{R}^d$. The standard intended approach relies on a Proof by Contradiction and the Bolzano-Weierstrass Theorem. However, a rigorous analysis reveals a subtle mathematical trap: the set $S = \{\mathbf{x} : \|\mathbf{x} - \mathbf{x}_0\| \geq \delta\}$ is closed but unbounded.

If a function asymptotically approaches its maximum at infinity, the strict inequality can fail. To secure full marks in an advanced exam like the ISI JRF, we must first demonstrate the logical contradiction assuming a bounded sequence (showing the examiner's implicit compact domain assumption), and then provide a rigorous Counterexample to show the mathematical boundary of the claim in $\mathbb{R}^d$.

โœ๏ธ Step-by-Step Proof / Derivation

Step 1: Setting up the Contradiction

Let $S = \{\mathbf{x} \in \mathbb{R}^d : \|\mathbf{x} - \mathbf{x}_0\| \geq \delta\}$. Because the Euclidean norm is continuous, $S$ is a closed set. Since $f$ has a unique global maximum at $\mathbf{x}_0$, we know $f(\mathbf{x}_0) > f(\mathbf{x})$ for all $\mathbf{x} \in S$. Let $M = f(\mathbf{x}_0)$.

Assume for contradiction that the strict inequality does not hold. Since $f(\mathbf{x}) \leq M$ globally, this forces the supremum over $S$ to equal the maximum:

$$ \sup_{\mathbf{x} \in S} f(\mathbf{x}) = M $$

By the definition of the supremum, there must exist a sequence $\{\mathbf{x}_n\} \subset S$ such that:

$$ \lim_{n \to \infty} f(\mathbf{x}_n) = M $$

Step 2: The Bounded Case (Applying Bolzano-Weierstrass)

Assume the sequence $\{\mathbf{x}_n\}$ is bounded. By the Bolzano-Weierstrass Theorem, it contains a convergent subsequence $\mathbf{x}_{n_k} \to \mathbf{y}$ for some limit point $\mathbf{y} \in \mathbb{R}^d$.

Because $S$ is a closed set, the limit point must lie within the set, so $\mathbf{y} \in S$. This implies $\|\mathbf{y} - \mathbf{x}_0\| \geq \delta$, which ensures $\mathbf{y} \neq \mathbf{x}_0$.

Since $f$ is continuous, the limit of the function values equals the function value of the limit:

$$ f(\mathbf{y}) = f\left(\lim_{k \to \infty} \mathbf{x}_{n_k}\right) = \lim_{k \to \infty} f(\mathbf{x}_{n_k}) = M $$

We have found a point $\mathbf{y} \neq \mathbf{x}_0$ where $f(\mathbf{y}) = f(\mathbf{x}_0) = M$. This directly contradicts the premise that $\mathbf{x}_0$ is the unique maximum. Thus, if the domain were compact, the strict inequality $f(\mathbf{x}_0) > \sup_{S} f(\mathbf{x})$ holds.

Step 3: The Critical Counterexample for the Unbounded Case

The contradiction in Step 2 fails if the sequence $\{\mathbf{x}_n\}$ is unbounded (i.e., $\|\mathbf{x}_n\| \to \infty$). In $\mathbb{R}^d$, the statement as posed in the exam is technically false without a condition that $f(\mathbf{x})$ decays at infinity (coercivity). We construct a counterexample to prove this.

Define $f: \mathbb{R}^d \mapsto \mathbb{R}$ as:

$$ f(\mathbf{x}) = 1 - \frac{\|\mathbf{x}\|^2}{1 + \|\mathbf{x}\|^4} $$

Let us evaluate the properties of this function:

Thus, $f$ has a strictly unique maximum at $\mathbf{x}_0 = \mathbf{0}$. Now, let $\delta = 1$, making the region $S = \{\mathbf{x} \in \mathbb{R}^d : \|\mathbf{x}\| \geq 1\}$. Let us find the supremum of $f$ on $S$:

$$ \sup_{\mathbf{x} \in S} f(\mathbf{x}) = \lim_{\|\mathbf{x}\| \to \infty} \left( 1 - \frac{\|\mathbf{x}\|^2}{1 + \|\mathbf{x}\|^4} \right) = 1 - 0 = 1 $$

We now test the proposed inequality $f(\mathbf{x}_0) > \sup_{\mathbf{x} \in S} f(\mathbf{x})$ against our findings:

$$ 1 > 1 $$

This is clearly a contradiction. The strict inequality fails to hold.

Final Answer / Q.E.D: The statement $f(\mathbf{x}_0) > \sup_{S} f(\mathbf{x})$ is false for general continuous functions on $\mathbb{R}^d$, as demonstrated by the counterexample $f(\mathbf{x}) = 1 - \|\mathbf{x}\|^2/(1 + \|\mathbf{x}\|^4)$ where the function asymptotically approaches its unique maximum at infinity. The theorem only holds true if we add the assumption that $\lim_{\|\mathbf{x}\| \to \infty} f(\mathbf{x}) < f(\mathbf{x}_0)$.

๐Ÿ“Œ Q3 Matrix Theory & Positive Semi-Definiteness (12 Marks)

๐Ÿง  Approach & Key Concepts

This problem tests the understanding of quadratic forms and projection matrices. We will use the definition of a Positive Semi-Definite (PSD) matrix ($v^\top A v \geq 0$) combined with the Cauchy-Schwarz Inequality. For the rank, we will utilize the property that for an idempotent matrix, the rank is equal to its trace.

โœ๏ธ Step-by-Step Proof / Derivation

Step 1: Verification of Positive Semi-Definiteness

A symmetric matrix $\mathbf{A}$ is PSD if for any non-zero vector $\mathbf{v} \in \mathbb{R}^d$, the quadratic form is non-negative. We expand $\mathbf{v}^\top \mathbf{A} \mathbf{v}$:

$$ \mathbf{v}^\top \mathbf{A} \mathbf{v} = \mathbf{v}^\top (\mathbf{I}_d - \mathbf{u}\mathbf{u}^\top) \mathbf{v} = \mathbf{v}^\top \mathbf{v} - \mathbf{v}^\top \mathbf{u}\mathbf{u}^\top \mathbf{v} $$

Using inner product notation, where $\mathbf{v}^\top \mathbf{v} = \|\mathbf{v}\|^2$ and $\mathbf{v}^\top \mathbf{u} = \mathbf{u}^\top \mathbf{v}$:

$$ \mathbf{v}^\top \mathbf{A} \mathbf{v} = \|\mathbf{v}\|^2 - (\mathbf{u}^\top \mathbf{v})^2 $$

Step 2: Applying Cauchy-Schwarz Inequality

The Cauchy-Schwarz inequality states that $(\mathbf{u}^\top \mathbf{v})^2 \leq \|\mathbf{u}\|^2 \|\mathbf{v}\|^2$. Since $\mathbf{u}$ is a unit vector, $\|\mathbf{u}\|^2 = 1$. Therefore:

$$ (\mathbf{u}^\top \mathbf{v})^2 \leq \|\mathbf{v}\|^2 \implies \|\mathbf{v}\|^2 - (\mathbf{u}^\top \mathbf{v})^2 \geq 0 $$

Thus, $\mathbf{v}^\top \mathbf{A} \mathbf{v} \geq 0$ for all $\mathbf{v}$, meaning $\mathbf{A}$ is positive semi-definite.

Step 3: Finding the Rank of A via Idempotency

We first check if $\mathbf{A}$ is idempotent ($\mathbf{A}^2 = \mathbf{A}$).

$$ \mathbf{A}^2 = (\mathbf{I} - \mathbf{u}\mathbf{u}^\top)(\mathbf{I} - \mathbf{u}\mathbf{u}^\top) = \mathbf{I} - 2\mathbf{u}\mathbf{u}^\top + \mathbf{u}(\mathbf{u}^\top \mathbf{u})\mathbf{u}^\top $$

Since $\mathbf{u}$ is a unit vector, $\mathbf{u}^\top \mathbf{u} = 1$, reducing the equation to $\mathbf{A}^2 = \mathbf{I} - \mathbf{u}\mathbf{u}^\top = \mathbf{A}$. For idempotent matrices, $\text{rank}(\mathbf{A}) = \text{tr}(\mathbf{A})$.

$$ \text{rank}(\mathbf{A}) = \text{tr}(\mathbf{I}_d) - \text{tr}(\mathbf{u}\mathbf{u}^\top) = d - \text{tr}(\mathbf{u}^\top \mathbf{u}) = d - 1 $$
Final Answer / Q.E.D: Matrix $\mathbf{A}$ is Positive Semi-Definite, and its rank is $d - 1$.

๐Ÿ“Œ Q4 Combinatorial Probability on Graphs (12 Marks)

๐Ÿง  Approach & Key Concepts

This is a conditional probability problem modeled as a graph coloring problem. The four pillars form a Cycle Graph ($C_4$). We need to find the probability of using exactly 4 colors given that it is a proper graph coloring (adjacent nodes have different colors). We will use the Chromatic Polynomial for the sample space and permutations for the favorable events.

โœ๏ธ Step-by-Step Proof / Derivation

Step 1: Define the Sample Space ($S$)

The total number of ways to color a cycle graph $C_n$ with $k$ colors such that adjacent vertices are distinct is given by the chromatic polynomial $P(C_n, k) = (k-1)^n + (-1)^n(k-1)$. For $n=4$ pillars and $k=4$ colors:

$$ |S| = (4-1)^4 + (-1)^4(4-1) = 3^4 + 3 = 81 + 3 = 84 $$

Step 2: Define the Favorable Event ($E$)

The event is that "all 4 colors are used." This implies every pillar gets a unique color, which is a simple permutation of 4 colors over 4 spots:

$$ |E| = 4! = 24 $$

We must ensure these 24 permutations satisfy the adjacency condition. Since all colors are mutually distinct, no two adjacent pillars can share a color. Thus, all 24 ways are valid.

Step 3: Calculate the Conditional Probability

Since the colors are chosen uniformly at random, the probability is the ratio of favorable outcomes to the valid sample space:

$$ P(E) = \frac{|E|}{|S|} = \frac{24}{84} = \frac{2}{7} $$
Final Answer / Q.E.D: The probability that all four colors are used, given the adjacency constraint, is $2/7$.

๐Ÿ“Œ Q5 Bounds on Variance and Expectations (12 Marks)

Problem Statement: Let $X$ be a random variable taking values in the interval $[1, 10]$.

(a) If $\mathbb{E}(X) = 2.8$, then show that the standard deviation of $X$ cannot exceed $3.6$.
(b) If $X_1$ and $X_2$ are two independent random variables having the same distribution as $X$, then show that $\mathbb{E}\left(\frac{X_1}{X_2}\right)$ cannot be smaller than 1.

๐Ÿง  Approach & Key Concepts

This problem tests your ability to establish bounds on moments of bounded random variables.

  • For part (a): We will use the algebraic property of bounded variables. Since $X \in [a, b]$, the product $(X-a)(X-b)$ is always non-positive. Taking the expectation of this quadratic provides a strict upper bound on the second moment, $\mathbb{E}(X^2)$, and consequently the variance. (This is related to the Bhatia-Davis inequality).
  • For part (b): We will exploit the independence of the variables to separate the expectations, and then apply Jensen's Inequality for the strictly convex function $g(x) = 1/x$.

โœ๏ธ Step-by-Step Proof / Derivation

Step 1: Establishing the Quadratic Bound for Part (a)

We are given that $X$ takes values in the interval $[1, 10]$. Therefore, with probability 1:

$$ 1 \leq X \leq 10 $$

This inequality implies that $(X - 1) \geq 0$ and $(X - 10) \leq 0$. Multiplying these two terms yields a non-positive result:

$$ (X - 1)(X - 10) \leq 0 $$

Expanding this expression gives:

$$ X^2 - 11X + 10 \leq 0 \implies X^2 \leq 11X - 10 $$

Step 2: Bounding the Variance and Standard Deviation

Taking the expectation on both sides of the inequality, and using the linearity of expectation, we get:

$$ \mathbb{E}(X^2) \leq 11\mathbb{E}(X) - 10 $$

We are given that $\mathbb{E}(X) = 2.8$. Substituting this value into our inequality:

$$ \mathbb{E}(X^2) \leq 11(2.8) - 10 = 30.8 - 10 = 20.8 $$

Now, we calculate the maximum possible variance of $X$ using the standard variance formula $\text{Var}(X) = \mathbb{E}(X^2) - [\mathbb{E}(X)]^2$:

$$ \text{Var}(X) \leq 20.8 - (2.8)^2 = 20.8 - 7.84 = 12.96 $$

The standard deviation $\sigma_X$ is the square root of the variance. Since variance is bounded by $12.96$:

$$ \sigma_X \leq \sqrt{12.96} = 3.6 $$

This completes the proof for part (a).


Step 3: Separating the Expectation for Part (b)

We are given two independent and identically distributed (i.i.d.) random variables, $X_1$ and $X_2$, both distributed as $X$. We need to evaluate the expectation of their ratio.

Because $X_1$ and $X_2$ are independent, any function of $X_1$ is independent of any function of $X_2$. Therefore, the expectation of their product can be factored:

$$ \mathbb{E}\left(\frac{X_1}{X_2}\right) = \mathbb{E}\left(X_1 \cdot \frac{1}{X_2}\right) = \mathbb{E}(X_1) \cdot \mathbb{E}\left(\frac{1}{X_2}\right) $$

Since $X_1$ and $X_2$ are identically distributed, $\mathbb{E}(X_1) = \mathbb{E}(X_2) = \mathbb{E}(X)$. The equation simplifies to:

$$ \mathbb{E}\left(\frac{X_1}{X_2}\right) = \mathbb{E}(X) \cdot \mathbb{E}\left(\frac{1}{X}\right) $$

Step 4: Applying Jensen's Inequality

Consider the function $g(x) = \frac{1}{x}$. We evaluate its convexity on the support of $X$, which is $[1, 10]$ (hence $x > 0$).

The second derivative of $g(x)$ is $g''(x) = \frac{2}{x^3}$. For all $x \in [1, 10]$, $g''(x) > 0$. Since the second derivative is strictly positive, $g(x) = \frac{1}{x}$ is a strictly convex function on the interval.

By Jensen's Inequality, for any convex function $g$ and random variable $X$:

$$ \mathbb{E}[g(X)] \geq g(\mathbb{E}[X]) $$

Applying this to our specific function $g(x) = 1/x$:

$$ \mathbb{E}\left(\frac{1}{X}\right) \geq \frac{1}{\mathbb{E}(X)} $$

Substitute this lower bound back into our factored expectation from Step 3:

$$ \mathbb{E}\left(\frac{X_1}{X_2}\right) = \mathbb{E}(X) \cdot \mathbb{E}\left(\frac{1}{X}\right) \geq \mathbb{E}(X) \cdot \left(\frac{1}{\mathbb{E}(X)}\right) = 1 $$
Final Answer / Q.E.D:
(a) We have successfully shown that $\text{Var}(X) \leq 12.96$, meaning the standard deviation cannot exceed $3.6$.
(b) Using Jensen's inequality and independence, we have shown that $\mathbb{E}\left(\frac{X_1}{X_2}\right) \geq 1$, meaning it cannot be smaller than $1$.

๐Ÿ“Œ Q6 Conditional Distributions and Bivariate Normality (12 Marks)

Problem Statement: Consider three random variables $X$, $Y$, and $Z$ where $Z \sim U(0,1)$, and for any given value $z$ of $Z$, the other two variables $X$ and $Y$ are independent and identically distributed as $N(z, 1)$.

(a) Verify whether $X$ and $Y$ are independent.
(b) Does $(X,Y)$ follow a bivariate normal distribution? Justify your answer.

๐Ÿง  Approach & Key Concepts

This problem explores the relationship between conditional independence and marginal independence, as well as the strict properties required for a Bivariate Normal Distribution.

  • For part (a): Two variables can be conditionally independent given $Z$, but marginally dependent if they both share the latent factor $Z$. We will use the Law of Total Covariance to show that their marginal covariance is non-zero, disproving independence.
  • For part (b): A fundamental property of the Bivariate Normal distribution is that its marginal distributions must also be strictly univariate Normal. We will compute the Moment Generating Function (MGF) of the marginal distribution of $X$ to show it is a mixture, not a pure Normal distribution.

โœ๏ธ Step-by-Step Proof / Derivation

Step 1: Setting up the Conditional Distributions for Part (a)

We are given the latent variable $Z \sim U(0,1)$. Thus, its expectation and variance are:

$$ \mathbb{E}[Z] = \frac{1}{2}, \quad \text{Var}(Z) = \frac{1}{12} $$

We are also given that conditional on $Z=z$, $X$ and $Y$ are independent and identically distributed (i.i.d.) as $N(z, 1)$. This implies:

$$ \mathbb{E}[X|Z] = Z, \quad \mathbb{E}[Y|Z] = Z $$
$$ \text{Var}(X|Z) = 1, \quad \text{Var}(Y|Z) = 1 $$

Furthermore, because $X$ and $Y$ are conditionally independent given $Z$, their conditional covariance is zero:

$$ \text{Cov}(X, Y | Z) = 0 $$

Step 2: Evaluating Marginal Independence via the Law of Total Covariance

To determine if $X$ and $Y$ are independent, we evaluate their marginal covariance. If they are independent, their covariance must be exactly $0$. We apply the Law of Total Covariance:

$$ \text{Cov}(X, Y) = \mathbb{E}[\text{Cov}(X, Y | Z)] + \text{Cov}(\mathbb{E}[X|Z], \mathbb{E}[Y|Z]) $$

Substitute the conditional expectations and covariance we established in Step 1:

$$ \text{Cov}(X, Y) = \mathbb{E}[0] + \text{Cov}(Z, Z) $$

The covariance of a variable with itself is simply its variance:

$$ \text{Cov}(X, Y) = \text{Var}(Z) = \frac{1}{12} $$

Since $\text{Cov}(X, Y) = 1/12 \neq 0$, the random variables $X$ and $Y$ are not marginally independent.


Step 3: Testing for Bivariate Normality for Part (b)

If the joint distribution $(X,Y)$ follows a bivariate normal distribution, then a necessary (though not sufficient on its own) condition is that the marginal distributions of $X$ and $Y$ must be univariate normal distributions. Let us find the marginal distribution of $X$ using its Moment Generating Function (MGF), $M_X(t)$.

By the Law of Iterated Expectations, the marginal MGF is the expected value of the conditional MGF:

$$ M_X(t) = \mathbb{E}[e^{tX}] = \mathbb{E}_Z[\mathbb{E}[e^{tX} | Z]] $$

Conditional on $Z$, $X \sim N(Z, 1)$. The MGF of a Normal distribution $N(\mu, \sigma^2)$ is $e^{\mu t + \frac{1}{2}\sigma^2 t^2}$. Substituting $\mu = Z$ and $\sigma^2 = 1$:

$$ \mathbb{E}[e^{tX} | Z] = \exp\left(Zt + \frac{1}{2}t^2\right) $$

Now, we take the expectation over the uniform distribution of $Z$:

$$ M_X(t) = \mathbb{E}_Z\left[ e^{Zt + \frac{1}{2}t^2} \right] = e^{\frac{1}{2}t^2} \mathbb{E}_Z[e^{tZ}] $$

Step 4: Evaluating the MGF of Z

Since $Z \sim U(0,1)$, its MGF is given by the integral:

$$ \mathbb{E}_Z[e^{tZ}] = \int_{0}^{1} e^{tz} \cdot 1 \, dz = \left[ \frac{e^{tz}}{t} \right]_0^1 = \frac{e^t - 1}{t} $$

Substituting this back into the marginal MGF for $X$ gives:

$$ M_X(t) = e^{\frac{1}{2}t^2} \left( \frac{e^t - 1}{t} \right) $$

The MGF of a normal distribution must be strictly of the form $\exp(\mu t + \frac{1}{2}\sigma^2 t^2)$. Because of the additional $\frac{e^t - 1}{t}$ term, $M_X(t)$ is clearly not the MGF of a normal distribution. Therefore, the marginal distribution of $X$ is not normal (it is a continuous mixture of normals).

Final Answer / Q.E.D:
(a) $X$ and $Y$ are not independent, as their marginal covariance is $1/12 \neq 0$.
(b) $(X,Y)$ does not follow a bivariate normal distribution, justified by the fact that the marginal distribution of $X$ is a mixture distribution whose MGF does not match that of a normal distribution.

๐Ÿ“Œ Q7 Bivariate Normal: Sufficiency and Ancillarity (12 Marks)

๐Ÿง  Approach & Key Concepts

This problem assesses your mastery of statistical inference, specifically data reduction techniques using Sufficient and Ancillary statistics.

  • For part (a): To find the minimal sufficient statistic, we will construct the joint likelihood function and apply the Lehmann-Scheffรฉ Theorem (also known as the ratio test for minimal sufficiency). By examining the ratio of likelihoods for two sample points, we will identify the coarsest partition of the sample space that makes the ratio parameter-free.
  • For part (b): A statistic is ancillary if its probability distribution does not depend on the parameter $\theta$. We will evaluate the marginal distribution of $Q_1$ to test for individual ancillarity. For the joint statistic $Q = (Q_1, Q_2)$, we will derive the cross-moment $\mathbb{E}[Q_1 Q_2]$ to demonstrate that their joint distribution structurally depends on the correlation parameter $\theta$.

โœ๏ธ Step-by-Step Proof / Derivation

Step 1: Constructing the Joint Likelihood Function for Part (a)

We are given a random sample $(X_1, Y_1), \dots, (X_n, Y_n)$ from a bivariate normal distribution with means $\mu_X = \mu_Y = 0$, variances $\sigma_X^2 = \sigma_Y^2 = 1$, and correlation $\rho = \theta$. The joint density (likelihood function) $L(\theta)$ is the product of the marginal densities:

$$ L(\theta | \mathbf{x}, \mathbf{y}) = \prod_{i=1}^n \frac{1}{2\pi\sqrt{1-\theta^2}} \exp\left\{ -\frac{1}{2(1-\theta^2)} (x_i^2 + y_i^2 - 2\theta x_i y_i) \right\} $$

Simplifying this by bringing the product into the exponent as a sum yields:

$$ L(\theta | \mathbf{x}, \mathbf{y}) = \left( \frac{1}{2\pi\sqrt{1-\theta^2}} \right)^n \exp\left\{ -\frac{1}{2(1-\theta^2)} \sum_{i=1}^n (x_i^2 + y_i^2) + \frac{\theta}{1-\theta^2} \sum_{i=1}^n x_i y_i \right\} $$

Step 2: Applying the Lehmann-Scheffรฉ Theorem

Let $(\mathbf{x}, \mathbf{y})$ and $(\mathbf{x}', \mathbf{y}')$ be two distinct sample points. We construct the likelihood ratio:

$$ \frac{L(\theta | \mathbf{x}, \mathbf{y})}{L(\theta | \mathbf{x}', \mathbf{y}')} = \exp\left\{ -\frac{1}{2(1-\theta^2)} \left[ \sum_{i=1}^n (x_i^2 + y_i^2) - \sum_{i=1}^n (x_i'^2 + y_i'^2) \right] + \frac{\theta}{1-\theta^2} \left[ \sum_{i=1}^n x_i y_i - \sum_{i=1}^n x_i' y_i' \right] \right\} $$

For this ratio to be independent of the parameter $\theta$ for all $\theta \in (-1, 1)$, the terms multiplied by the functionally independent parameter components, namely $-\frac{1}{2(1-\theta^2)}$ and $\frac{\theta}{1-\theta^2}$, must perfectly vanish. This requires:

$$ \sum_{i=1}^n (x_i^2 + y_i^2) = \sum_{i=1}^n (x_i'^2 + y_i'^2) \quad \text{and} \quad \sum_{i=1}^n x_i y_i = \sum_{i=1}^n x_i' y_i' $$

By the Lehmann-Scheffรฉ theorem, the minimal sufficient statistic is the vector $T$ that defines this equivalence class:

$$ T(\mathbf{X}, \mathbf{Y}) = \left( \sum_{i=1}^n (X_i^2 + Y_i^2), \sum_{i=1}^n X_i Y_i \right) $$

Step 3: Evaluating Ancillarity of $Q_1$ for Part (b)

A statistic is ancillary if its distribution does not depend on $\theta$. Let's examine $Q_1 = \sum_{i=1}^n X_i^2$.

From the joint bivariate normal density $f_\theta(x, y)$, integrating out $y$ gives the marginal distribution of $X$. Since $\mu_X = 0$ and $\sigma_X^2 = 1$, the marginal distribution is standard normal:

$$ X_i \sim N(0, 1) $$

Because the $X_i$ are independent standard normal variables, the sum of their squares follows a Chi-square distribution with $n$ degrees of freedom:

$$ Q_1 = \sum_{i=1}^n X_i^2 \sim \chi^2_{(n)} $$

Since the $\chi^2_{(n)}$ distribution has no dependence on $\theta$, we conclude that $Q_1$ is strictly ancillary for $\theta$.

Step 4: Evaluating Ancillarity of the Joint Statistic $Q = (Q_1, Q_2)$

By symmetry, the marginal distribution of $Y_i$ is also $N(0,1)$, meaning $Q_2 = \sum_{i=1}^n Y_i^2 \sim \chi^2_{(n)}$. Thus, $Q_2$ is individually ancillary. However, for the joint vector $Q = (Q_1, Q_2)$ to be ancillary, its joint distribution must be independent of $\theta$.

To check this, let us evaluate the covariance between $Q_1$ and $Q_2$:

$$ \text{Cov}(Q_1, Q_2) = \text{Cov}\left(\sum_{i=1}^n X_i^2, \sum_{j=1}^n Y_j^2\right) = \sum_{i=1}^n \text{Cov}(X_i^2, Y_i^2) $$

We compute $\text{Cov}(X_i^2, Y_i^2) = \mathbb{E}[X_i^2 Y_i^2] - \mathbb{E}[X_i^2]\mathbb{E}[Y_i^2]$. Since $X_i, Y_i \sim N(0,1)$, we know $\mathbb{E}[X_i^2] = 1$ and $\mathbb{E}[Y_i^2] = 1$.

Using the properties of conditional expectation for bivariate normals, $Y_i | X_i \sim N(\theta X_i, 1-\theta^2)$:

$$ \mathbb{E}[X_i^2 Y_i^2] = \mathbb{E}_{X}\left[X_i^2 \mathbb{E}[Y_i^2 | X_i]\right] = \mathbb{E}_{X}\left[X_i^2 (\text{Var}(Y_i|X_i) + (\mathbb{E}[Y_i|X_i])^2)\right] $$
$$ \mathbb{E}[X_i^2 Y_i^2] = \mathbb{E}_{X}\left[X_i^2 (1 - \theta^2 + \theta^2 X_i^2)\right] = (1-\theta^2)\mathbb{E}[X_i^2] + \theta^2\mathbb{E}[X_i^4] $$

Recall that the 4th moment of a standard normal is $3$. Thus:

$$ \mathbb{E}[X_i^2 Y_i^2] = (1-\theta^2)(1) + \theta^2(3) = 1 + 2\theta^2 $$

Returning to the covariance:

$$ \text{Cov}(X_i^2, Y_i^2) = (1 + 2\theta^2) - (1)(1) = 2\theta^2 \implies \text{Cov}(Q_1, Q_2) = 2n\theta^2 $$

Because the covariance of $Q_1$ and $Q_2$ is a direct function of $\theta$, their joint distribution fundamentally depends on $\theta$. Therefore, the vector $Q$ cannot be ancillary.

Final Answer / Q.E.D:
(a) The minimal sufficient statistic for $\theta$ is $T = \left( \sum_{i=1}^n (X_i^2 + Y_i^2), \sum_{i=1}^n X_i Y_i \right)$.
(b) (i) $Q_1$ is ancillary for $\theta$ because its marginal distribution is $\chi^2_{(n)}$, which is independent of $\theta$.
(b) (ii) $Q = (Q_1, Q_2)$ is not ancillary for $\theta$ because their joint distribution depends on $\theta$ (proven by $\text{Cov}(Q_1, Q_2) = 2n\theta^2$).

๐Ÿ“Œ Q8 Hypothesis Testing with Geometric Constraints (12 Marks)

Problem Statement: Let $\theta_1, \theta_2, \theta_3$ be the true angles of a huge triangular field. Suppose that each of these angles is measured by a separate person subject to an additive error. Assume that these three errors are independently distributed as $N(0, \sigma^2)$ where $\sigma^2$ is unknown. Let these measurements be $Y_1, Y_2, Y_3$, respectively. Construct a test of level $\alpha$ ($0 < \alpha < 1$) for testing $H_0 : \theta_1 = \theta_2$ against $H_1 : \theta_1 \neq \theta_2$.

๐Ÿง  Approach & Key Concepts

This problem falls under the framework of Linear Models with Constraints. The fundamental constraint is geometric: the sum of the true angles of a planar triangle is exactly $180^\circ$ (or $\pi$ radians). The strategy involves:

  • Defining a suitable unbiased estimator for the difference $\theta_1 - \theta_2$.
  • Finding an independent estimator for the unknown error variance $\sigma^2$ using the geometric constraint $\sum \theta_i = \pi$.
  • Constructing a Student's t-statistic (or equivalent F-statistic) by taking the ratio of the standard normal variable to the square root of the normalized chi-square variable.

โœ๏ธ Step-by-Step Proof / Derivation

Step 1: Formulating the Model and the Constraint

We are given that the measurements $Y_i$ are subject to independent $N(0, \sigma^2)$ errors. Thus, the measurement model is:

$$ Y_i \sim N(\theta_i, \sigma^2), \quad i = 1, 2, 3 $$

Since the angles belong to a triangle, their true sum is a known constant. Let $S = 180^\circ$ (or $\pi$). Thus, we have the deterministic constraint:

$$ \theta_1 + \theta_2 + \theta_3 = S $$

Step 2: Constructing the Test Statistic Numerator

We are testing $H_0: \theta_1 - \theta_2 = 0$. A natural, unbiased estimator for $\theta_1 - \theta_2$ is the difference in their measurements. Let $U = Y_1 - Y_2$. By the properties of normal combinations:

$$ \mathbb{E}[U] = \theta_1 - \theta_2 $$
$$ \text{Var}(U) = \text{Var}(Y_1) + \text{Var}(Y_2) = 2\sigma^2 $$

Under the null hypothesis $H_0$, $U \sim N(0, 2\sigma^2)$. Standardizing this gives us a standard normal variable:

$$ Z = \frac{Y_1 - Y_2}{\sqrt{2}\sigma} \sim N(0, 1) $$

Step 3: Estimating the Unknown Variance $\sigma^2$

Because $\sigma^2$ is unknown, we need to estimate it using the remaining degrees of freedom. Consider the sum of the measurements, $V = Y_1 + Y_2 + Y_3$.

$$ \mathbb{E}[V] = \theta_1 + \theta_2 + \theta_3 = S $$
$$ \text{Var}(V) = \text{Var}(Y_1) + \text{Var}(Y_2) + \text{Var}(Y_3) = 3\sigma^2 $$

Thus, $V \sim N(S, 3\sigma^2)$. We can construct a $\chi^2$ random variable with 1 degree of freedom:

$$ X = \frac{(V - S)^2}{3\sigma^2} = \frac{(Y_1 + Y_2 + Y_3 - S)^2}{3\sigma^2} \sim \chi^2_{(1)} $$

Step 4: Proving Independence of $U$ and $V$

For the t-statistic to be valid, the numerator ($Z$) and the denominator (function of $X$) must be independent. Since they are linear combinations of jointly normal variables, we only need to show their covariance is zero:

$$ \text{Cov}(U, V) = \text{Cov}(Y_1 - Y_2, Y_1 + Y_2 + Y_3) = \text{Var}(Y_1) - \text{Var}(Y_2) = \sigma^2 - \sigma^2 = 0 $$

Since $\text{Cov}(U, V) = 0$, $U$ and $V$ are strictly independent. Consequently, $Z$ and $X$ are independent.

Step 5: Constructing the Final t-Test

We define the test statistic $T$ as the ratio of the standard normal variable $Z$ to the square root of the chi-square variable $X$ divided by its degrees of freedom ($\nu = 1$):

$$ T = \frac{Z}{\sqrt{X / 1}} = \frac{\frac{Y_1 - Y_2}{\sqrt{2}\sigma}}{\sqrt{\frac{(Y_1 + Y_2 + Y_3 - S)^2}{3\sigma^2}}} $$

The unknown parameter $\sigma$ perfectly cancels out, leaving a completely computable statistic:

$$ T = \sqrt{\frac{3}{2}} \frac{Y_1 - Y_2}{|Y_1 + Y_2 + Y_3 - S|} $$

Under $H_0$, this statistic $T$ follows a Student's t-distribution with $1$ degree of freedom (which is equivalent to a standard Cauchy distribution).

Final Answer / Q.E.D:
The required level $\alpha$ test rejects the null hypothesis $H_0 : \theta_1 = \theta_2$ in favor of $H_1 : \theta_1 \neq \theta_2$ if the observed test statistic falls in the critical region: $$ |T| > t_{1, \alpha/2} $$ where $T = \sqrt{\frac{3}{2}} \frac{Y_1 - Y_2}{|Y_1 + Y_2 + Y_3 - 180^\circ|}$ and $t_{1, \alpha/2}$ is the upper $\alpha/2$ critical value of the Student's t-distribution with 1 degree of freedom.

๐Ÿ“Œ Q9 Most Powerful Test and Neyman-Pearson Lemma (12 Marks)

Problem Statement: Suppose $X_1, X_2, \dots, X_n$ are independent random variables with the pdf $f_\theta(x) = 2e^{-2(x-\theta)} \mathbb{I}\{x \geq \theta\}$. Let $\theta_0$ and $\theta_1$ be known real numbers with $\theta_1 > \theta_0$. For a prefixed $\alpha \in (0,1)$, derive a randomized most powerful test of size $\alpha$ for testing $H_0 : \theta = \theta_0$ against $H_1 : \theta = \theta_1$ when (a) $e^{2n(\theta_0-\theta_1)} < \alpha$ and (b) $e^{2n(\theta_0-\theta_1)} > \alpha$.

๐Ÿง  Approach & Key Concepts

This problem asks for the Most Powerful (MP) test for a simple null versus a simple alternative hypothesis. We will use the Neyman-Pearson Lemma (NPL). The key steps are to find the joint likelihood function, identify the sufficient statistic (which will be the first order statistic $X_{(1)}$ due to the indicator function), formulate the Likelihood Ratio $\Lambda$, and determine the critical region based on the size $\alpha$. Because the support of the distribution depends on the parameter $\theta$, the likelihood ratio will be piecewise, leading to different randomized test structures depending on the condition $e^{2n(\theta_0 - \theta_1)} \lessgtr \alpha$.

โœ๏ธ Step-by-Step Proof / Derivation

Step 1: Likelihood Function and Sufficient Statistic

The joint probability density function (likelihood) of the sample is:

$$ L(\theta | \mathbf{x}) = \prod_{i=1}^n 2e^{-2(x_i - \theta)} \mathbb{I}\{x_i \geq \theta\} = 2^n \exp\left(-2\sum_{i=1}^n x_i + 2n\theta\right) \mathbb{I}\{X_{(1)} \geq \theta\} $$

where $X_{(1)} = \min(X_1, \dots, X_n)$. By the Factorization Theorem, $X_{(1)}$ is a minimal sufficient statistic for $\theta$. We must find the distribution of $X_{(1)}$ under $H_0$ to evaluate probabilities. For $y \geq \theta_0$:

$$ P_{\theta_0}(X_{(1)} \geq y) = \prod_{i=1}^n P_{\theta_0}(X_i \geq y) = \left[ \int_y^\infty 2e^{-2(x-\theta_0)} dx \right]^n = \left[ e^{-2(y-\theta_0)} \right]^n = e^{-2n(y-\theta_0)} $$

Step 2: Constructing the Likelihood Ratio ($\Lambda$)

We are testing $H_0: \theta = \theta_0$ against $H_1: \theta = \theta_1$ (with $\theta_1 > \theta_0$). The likelihood ratio is:

$$ \Lambda(X_{(1)}) = \frac{L(\theta_1)}{L(\theta_0)} = \frac{e^{2n\theta_1} \mathbb{I}\{X_{(1)} \geq \theta_1\}}{e^{2n\theta_0} \mathbb{I}\{X_{(1)} \geq \theta_0\}} $$

Since $\theta_1 > \theta_0$, we analyze $\Lambda$ piecewise over the support of $H_0$ ($X_{(1)} \geq \theta_0$):

  • If $\theta_0 \leq X_{(1)} < \theta_1$, the numerator is $0$, so $\Lambda = 0$.
  • If $X_{(1)} \geq \theta_1$, both indicators are $1$, so $\Lambda = e^{2n(\theta_1 - \theta_0)} = k^*$.

By the Neyman-Pearson Lemma, the Most Powerful test function $\phi(X_{(1)})$ has the form:

$$ \phi(x) = \begin{cases} 1 & \text{if } \Lambda > c \\ \gamma & \text{if } \Lambda = c \\ 0 & \text{if } \Lambda < c \end{cases} $$

where $c$ and $\gamma$ are chosen such that $\mathbb{E}_{\theta_0}[\phi(X_{(1)})] = \alpha$. Note that $P_{\theta_0}(\Lambda = k^*) = P_{\theta_0}(X_{(1)} \geq \theta_1) = e^{-2n(\theta_1 - \theta_0)} = e^{2n(\theta_0 - \theta_1)}$.


Step 3: Deriving the Test for Case (a) $e^{2n(\theta_0 - \theta_1)} < \alpha$

In this case, the probability of obtaining the maximum likelihood ratio $k^*$ under $H_0$ is strictly less than the allowed size $\alpha$:

$$ P_{\theta_0}(\Lambda = k^*) = e^{2n(\theta_0 - \theta_1)} < \alpha $$

If we set the threshold $c = k^*$, our test size would be too small. Therefore, we must lower the threshold to $c = 0$. The test rule becomes:

  • Reject $H_0$ with probability $1$ if $\Lambda > 0$ (i.e., $X_{(1)} \geq \theta_1$).
  • Reject $H_0$ with probability $\gamma_1$ if $\Lambda = 0$ (i.e., $\theta_0 \leq X_{(1)} < \theta_1$).

We solve for $\gamma_1$ using the size constraint:

$$ \mathbb{E}_{\theta_0}[\phi] = 1 \cdot P_{\theta_0}(X_{(1)} \geq \theta_1) + \gamma_1 \cdot P_{\theta_0}(\theta_0 \leq X_{(1)} < \theta_1) = \alpha $$
$$ e^{2n(\theta_0 - \theta_1)} + \gamma_1 \left( 1 - e^{2n(\theta_0 - \theta_1)} \right) = \alpha $$
$$ \gamma_1 = \frac{\alpha - e^{2n(\theta_0 - \theta_1)}}{1 - e^{2n(\theta_0 - \theta_1)}} $$

Step 4: Deriving the Test for Case (b) $e^{2n(\theta_0 - \theta_1)} > \alpha$

In this case, the probability of obtaining the maximum likelihood ratio $k^*$ under $H_0$ exceeds $\alpha$:

$$ P_{\theta_0}(\Lambda = k^*) = e^{2n(\theta_0 - \theta_1)} > \alpha $$

If we always rejected $H_0$ when $X_{(1)} \geq \theta_1$, the Type I error would exceed $\alpha$. Thus, we must set $c = k^* = e^{2n(\theta_1 - \theta_0)}$ and introduce randomization at this boundary. The test rule becomes:

  • Reject $H_0$ with probability $\gamma_2$ if $\Lambda = k^*$ (i.e., $X_{(1)} \geq \theta_1$).
  • Reject $H_0$ with probability $0$ if $\Lambda < k^*$ (i.e., $\theta_0 \leq X_{(1)} < \theta_1$).

We solve for $\gamma_2$ using the size constraint:

$$ \mathbb{E}_{\theta_0}[\phi] = \gamma_2 \cdot P_{\theta_0}(X_{(1)} \geq \theta_1) = \alpha $$
$$ \gamma_2 \cdot e^{2n(\theta_0 - \theta_1)} = \alpha \implies \gamma_2 = \alpha e^{2n(\theta_1 - \theta_0)} $$
Final Answer / Q.E.D:
The randomized Most Powerful test function $\phi(X_{(1)})$ of size $\alpha$ is given by:

For Case (a) $e^{2n(\theta_0-\theta_1)} < \alpha$: $$ \phi(X_{(1)}) = \begin{cases} 1 & \text{if } X_{(1)} \geq \theta_1 \\ \frac{\alpha - e^{2n(\theta_0 - \theta_1)}}{1 - e^{2n(\theta_0 - \theta_1)}} & \text{if } \theta_0 \leq X_{(1)} < \theta_1 \end{cases} $$
For Case (b) $e^{2n(\theta_0-\theta_1)} > \alpha$: $$ \phi(X_{(1)}) = \begin{cases} \alpha e^{2n(\theta_1 - \theta_0)} & \text{if } X_{(1)} \geq \theta_1 \\ 0 & \text{if } \theta_0 \leq X_{(1)} < \theta_1 \end{cases} $$

๐Ÿ“Œ Q10 Multivariate Normal Distribution (12 Marks)

Problem Statement: Suppose $(X_1, X_2, X_3, X_4)$ follows a multivariate normal distribution with mean $(0,0,0,0)$ and the dispersion matrix: $$ \boldsymbol{\Sigma} = \begin{pmatrix} 25 & 0 & 0 & 1 \\ 0 & 16 & 4 & 0 \\ 0 & 4 & 9 & 0 \\ 1 & 0 & 0 & 4 \end{pmatrix} $$ (a) Find $\mathbb{P}(X_1 > X_2 X_3 X_4)$.
(b) Compute the multiple correlation coefficient of $X_1$ with $(X_2, X_3, X_4)$.
(c) Compute the partial correlation coefficient between $X_2$ and $X_3$ given $(X_1, X_4)$.

๐Ÿง  Approach & Key Concepts

This problem evaluates your structural understanding of the Multivariate Normal Distribution.

  • For part (a): Direct integration is impossible. Instead, we exploit the symmetry of the multivariate normal distribution. We look for a distribution-preserving transformation that flips the sign of the target expression, proving it is symmetric around $0$.
  • For part (b): The Multiple Correlation Coefficient $R_{1 \cdot 234}$ measures the maximum correlation between $X_1$ and a linear combination of $(X_2, X_3, X_4)$. It is derived from partitioning the covariance matrix.
  • For part (c): We compute the Partial Correlation Coefficient by finding the conditional covariance matrix. A critical property of the multivariate normal is that uncorrelated blocks are strictly independent.

โœ๏ธ Step-by-Step Proof / Derivation

Step 1: Analyzing Independence and Symmetry for Part (a)

Let $W = X_1 - X_2 X_3 X_4$. We want to find $\mathbb{P}(W > 0)$. Notice the zero entries in the covariance matrix $\boldsymbol{\Sigma}$:

  • $\text{Cov}(X_1, X_2) = 0$, $\text{Cov}(X_1, X_3) = 0$
  • $\text{Cov}(X_4, X_2) = 0$, $\text{Cov}(X_4, X_3) = 0$

Because the variables are jointly normal, zero covariance implies strict independence. Thus, the random vector $(X_1, X_4)$ is completely independent of the random vector $(X_2, X_3)$.

Furthermore, since the mean vector is $\mathbf{0}$, the distribution of $(X_1, X_4)$ is symmetric. Specifically, if we apply the transformation $(X_1, X_4) \mapsto (-X_1, -X_4)$, the mean remains $(0,0)$ and the covariance matrix remains unchanged because $\text{Cov}(-X_1, -X_4) = (-1)(-1)\text{Cov}(X_1, X_4) = 1$.

Therefore, $(-X_1, X_2, X_3, -X_4)$ has the exact same joint distribution as $(X_1, X_2, X_3, X_4)$.

Step 2: Evaluating the Probability $\mathbb{P}(W > 0)$

Let us apply this transformation to our variable $W$:

$$ W' = (-X_1) - X_2 X_3 (-X_4) = -X_1 + X_2 X_3 X_4 = -(X_1 - X_2 X_3 X_4) = -W $$

Since the joint distribution is invariant under this transformation, $W$ and $-W$ are identically distributed ($W \stackrel{d}{=} -W$). This means the distribution of $W$ is perfectly symmetric around $0$.

Because $W$ is a continuous random variable (a polynomial of continuous normals), $\mathbb{P}(W = 0) = 0$. Therefore:

$$ \mathbb{P}(W > 0) = \mathbb{P}(W < 0) = \frac{1}{2} $$

Thus, $\mathbb{P}(X_1 > X_2 X_3 X_4) = \mathbb{P}(X_1 - X_2 X_3 X_4 > 0) = 0.5$.


Step 3: Partitioning the Covariance Matrix for Part (b)

The squared multiple correlation coefficient of $X_1$ with $\mathbf{X}_{-1} = (X_2, X_3, X_4)^\top$ is given by:

$$ R_{1 \cdot 234}^2 = \frac{\boldsymbol{\Sigma}_{1, -1} \boldsymbol{\Sigma}_{-1, -1}^{-1} \boldsymbol{\Sigma}_{-1, 1}}{\boldsymbol{\Sigma}_{11}} $$

From the given $\boldsymbol{\Sigma}$, we extract the sub-matrices:

$$ \boldsymbol{\Sigma}_{11} = 25, \quad \boldsymbol{\Sigma}_{1, -1} = (0, 0, 1), \quad \boldsymbol{\Sigma}_{-1, -1} = \begin{pmatrix} 16 & 4 & 0 \\ 4 & 9 & 0 \\ 0 & 0 & 4 \end{pmatrix} $$

Step 4: Computing $R_{1 \cdot 234}$

Notice that $\boldsymbol{\Sigma}_{-1, -1}$ is block-diagonal. The vector $\boldsymbol{\Sigma}_{1, -1}$ only has a non-zero entry in the 3rd position (corresponding to $X_4$). Therefore, we only need the bottom-right entry of the inverse matrix $\boldsymbol{\Sigma}_{-1, -1}^{-1}$, which is simply $1/4$.

$$ \boldsymbol{\Sigma}_{1, -1} \boldsymbol{\Sigma}_{-1, -1}^{-1} \boldsymbol{\Sigma}_{-1, 1} = (0, 0, 1) \begin{pmatrix} * & * & 0 \\ * & * & 0 \\ 0 & 0 & 1/4 \end{pmatrix} \begin{pmatrix} 0 \\ 0 \\ 1 \end{pmatrix} = 1 \times \frac{1}{4} \times 1 = \frac{1}{4} $$

Now, divide by $\boldsymbol{\Sigma}_{11}$:

$$ R_{1 \cdot 234}^2 = \frac{1/4}{25} = \frac{1}{100} $$

Taking the positive square root (since $R \geq 0$):

$$ R_{1 \cdot 234} = \frac{1}{10} = 0.1 $$

Step 5: Computing the Partial Correlation for Part (c)

We need the partial correlation $\rho_{23 \cdot 14}$, which is derived from the conditional covariance matrix of $(X_2, X_3)$ given $(X_1, X_4)$:

$$ \boldsymbol{\Sigma}_{23 | 14} = \boldsymbol{\Sigma}_{23} - \boldsymbol{\Sigma}_{23, 14} \boldsymbol{\Sigma}_{14}^{-1} \boldsymbol{\Sigma}_{14, 23} $$

We inspect the cross-covariance matrix $\boldsymbol{\Sigma}_{23, 14}$ between the block $(X_2, X_3)$ and the block $(X_1, X_4)$. From $\boldsymbol{\Sigma}$, all these cross-terms are zero.

$$ \boldsymbol{\Sigma}_{23, 14} = \begin{pmatrix} \text{Cov}(X_2, X_1) & \text{Cov}(X_2, X_4) \\ \text{Cov}(X_3, X_1) & \text{Cov}(X_3, X_4) \end{pmatrix} = \begin{pmatrix} 0 & 0 \\ 0 & 0 \end{pmatrix} $$

Because the cross-covariance is a zero matrix (confirming our observation in Step 1 that the blocks are independent), conditioning on $(X_1, X_4)$ provides zero information about $(X_2, X_3)$. Thus, the conditional covariance is exactly equal to the marginal covariance:

$$ \boldsymbol{\Sigma}_{23 | 14} = \boldsymbol{\Sigma}_{23} = \begin{pmatrix} 16 & 4 \\ 4 & 9 \end{pmatrix} $$

The partial correlation is therefore identical to the marginal Pearson correlation between $X_2$ and $X_3$:

$$ \rho_{23 \cdot 14} = \frac{\text{Cov}(X_2, X_3 | X_1, X_4)}{\sqrt{\text{Var}(X_2 | X_1, X_4) \text{Var}(X_3 | X_1, X_4)}} = \frac{4}{\sqrt{16 \times 9}} = \frac{4}{4 \times 3} = \frac{4}{12} = \frac{1}{3} $$
Final Answer / Q.E.D:
(a) $\mathbb{P}(X_1 > X_2 X_3 X_4) = 0.5$
(b) The multiple correlation coefficient $R_{1 \cdot 234} = 0.1$
(c) The partial correlation coefficient $\rho_{23 \cdot 14} = 1/3$

๐Ÿ“š Paper Summary & Key Focus Areas

๐ŸŽฏ Core Concepts Tested in This Paper

  • Real Analysis & Topology (Q1, Q2): Mastery of the Monotone Convergence Theorem, fixed-point limits, and recognizing topological traps (e.g., knowing when to apply Bolzano-Weierstrass and when to provide counterexamples for unbounded sets).
  • Matrix Theory & Linear Algebra (Q3): Exploiting structural properties like idempotency and trace-rank equivalence for projection matrices, and applying the Cauchy-Schwarz inequality for PSD verification.
  • Combinatorics & Probability (Q4, Q5, Q6): Identifying structural isomorphisms (like mapping pillars to a $C_4$ graph) to use Chromatic Polynomials. Distinguishing between conditional and marginal independence using the Law of Total Covariance, and applying Jensen's Inequality to bound moments.
  • Advanced Statistical Inference (Q7, Q9): Deriving minimal sufficient statistics via the Lehmann-Scheffรฉ (ratio) test, proving ancillarity, and rigorously constructing randomized Most Powerful (MP) tests using the Neyman-Pearson Lemma under different boundary size ($\alpha$) conditions.
  • Multivariate Normal & Linear Models (Q8, Q10): Partitioning covariance matrices to compute multiple and partial correlation coefficients. Using zero-covariance to establish strict block independence, and constructing valid t-statistics under geometric constraints.
๐Ÿ’ก ISI Examiner Insight:
In ISI subjective papers (PSB level), getting the "final numerical answer" is only a fraction of the battle. Examiners look for:
1. Explicit Theorem Citations: Naming the theorems you use (e.g., Neyman-Pearson, Jensen's, Bolzano-Weierstrass).
2. Checking Assumptions: Proving covariance is zero *before* claiming independence (as seen in Q8 and Q10).
3. Mathematical Skepticism: Recognizing when a statement is broadly false and providing a mathematically sound counterexample (as required in Q2).

Created for ISI JRF Aspirants and Mathematics Enthusiasts.

๐Ÿ“š ISI JRF Statistics Solutions