Homework #1
EE 541: Spring 2025
Assigned: 25 August
Due: Sunday, 07 September at 23:59
BrightSpace Assignment: Homework 1
Use only Python standard library modules (https://docs.python.org/3/library/) and matplotlib
for this assignment, i.e., do not import numpy
, scikit
, or any other non-standard package.
Problem 1
An MLP has two input nodes, one hidden layer, and two outputs. The two sets of weights and biases are given by:
\[ W_1 = \begin{bmatrix} 1 & -2 \\ 3 & 4 \end{bmatrix} \quad b_1 = \begin{bmatrix} 1 \\ 0 \end{bmatrix} \]
\[ W_2 = \begin{bmatrix} 2 & 2 \\ 2 & -3 \end{bmatrix} \quad b_2 = \begin{bmatrix} 0 \\ -4 \end{bmatrix} \]
The non-linear activation for the hidden layer is ReLU (rectified linear unit) – that is \(h(x) = \max(x, 0)\). The output layer is linear (i.e., identity activation function). The output for layer \(l\) is given by \(a^{(l)} = h_l(W_l a^{(l-1)} + b_l)\).
What is the output activation for input \(x = [+1; -1]^T\)?
Problem 2
Let \(f(x, y) = 4x^2 + y^2 - xy - 13x\)
Find \(\frac{\partial f}{\partial x}\), the partial derivative of \(f\) with respect to \(x\). Find \(\frac{\partial f}{\partial y}\).
Find \((x, y) \in \mathbb{R}^2\) that minimizes \(f\).
Problem 3
A hyper-plane in \(\mathbb{R}^n\) is the set, \(\{ \mathbf{x} : \mathbf{x} \in \mathbb{R}^n, w^T x + b = 0 \}\), where \(w \in \mathbb{R}^n\) and \(b\) is a real scalar.
The solution of the following optimization problem describes the distance between a point \(x_0 \in \mathbb{R}^n\) and the hyperplane \(w^T x + b = 0\): \[ \min_x \|x_0 - x\|_2 \quad \text{s.t.} \quad w^T x + b = 0. \] Derive an analytic solution for the distance between \(x_0\) and \(w^T x + b = 0\).
What is the distance between two hyperplanes, \(w^T x + b_1 = 0\) and \(w^T x + b_2 = 0\)?
Problem 4
A function \(f(x)\) is convex if \[ f(\lambda x + (1 - \lambda)y) \leq \lambda f(x) + (1 - \lambda)f(y) \] for all \(x, y\) and \(0 < \lambda < 1\).
Use this definition to prove that \(f(x) = x^2\) is a convex function. Verify that \(f(x) = x^3\) is not a convex function.
An \(n \times n\) matrix \(A\) is a positive semi-definite matrix if \(\mathbf{x}^T A \mathbf{x} \geq 0\), for any \(\mathbf{x} \in \mathbb{R}^n\) such that \(\mathbf{x} \neq \mathbf{0}\). Prove that the function \(f(\mathbf{x}) = \mathbf{x}^T A \mathbf{x}\) is convex if \(A\) is a positive semi-definite matrix for \(\mathbf{x} \in \mathbb{R}^n\).
Problem 5
Simulate tossing a biased coin (a Bernoulli trial) where \(P[\text{HEAD}] = 0.70\).
Count the number of heads in 50 trials. Record the longest run of heads.
Repeat the 50-flip experiment 20, 100, 200, and 1000 times. Use matplotlib to generate a histogram showing the observed number of heads for each case. Comment on the limit of the histogram.
Simulate tossing the coin 500 times. Generate a histogram showing the heads run lengths.
Problem 6
Define the random variable \(N = \min \{ n: \sum_{i=1}^{n} X_i > 4\}\) as the smallest number of standard uniform random samples whose sum exceeds four. Generate a histogram using 100, 1000, and 10000 realizations of \(N\). Comment on the expected value \(E[N]\).
Problem 1: MLP
Submit calculations and show the final layer output activation. No other submission required.
Problems 2-4: Analytic problems
Write your solutions to these homework problems on separate sheets of paper. Show all work and box answers where appropriate. Do not guess.
Problem 5: Simulate tossing biased coin
Submit histograms and analysis for each case to BrightSpace. Submit your Python code as an appendix to Gradescope. A suitably annotated Jupyter notebook with inline analysis is sufficient.
Gradescope: Problem 5
Problem 6: Sum of uniform to exceed threshold
Submit histograms and your comments on \(E[N]\) to BrightSpace. Submit your Python code as an appendix to Gradescope. A suitably annotated Jupyter notebook with inline analysis is sufficient.
Gradescope: Problem 6