If some characters seem to be missing, it's because MathJax is not loaded correctly. Refreshing the page should fix it.
-
[E] Why do we say that matrices are linear transformations?
-
[E] What’s the inverse of a matrix? Do all matrices have an inverse? Is the inverse of a matrix always unique?
-
[E] What does the determinant of a matrix represent?
-
[E] What happens to the determinant of a matrix if we multiply one of its rows by a scalar
$$t \times R$$ ? -
[M] A
$$4 \times 4$$ matrix has four eigenvalues$$3, 3, 2, -1$$ . What can we say about the trace and the determinant of this matrix? -
[M] Given the following matrix:
$$ \begin{bmatrix} 1 & 4 & -2 \ -1 & 3 & 2 \ 3 & 5 & -6 \end{bmatrix} $$Without explicitly using the equation for calculating determinants, what can we say about this matrix’s determinant?
Hint: rely on a property of this matrix to determine its determinant.
-
[M] What’s the difference between the covariance matrix
$$A^TA$$ and the Gram matrix$$AA^T$$ ? -
Given
$$A \in R^{n \times m}$$ and$$b \in R^n$$ - [M] Find
$$x$$ such that:$$Ax = b$$ . - [E] When does this have a unique solution?
- [M] Why is it when A has more columns than rows,
$$Ax = b$$ has multiple solutions? - [M] Given a matrix A with no inverse. How would you solve the equation
$$Ax = b$$ ? What is the pseudoinverse and how to calculate it?
- [M] Find
-
Derivative is the backbone of gradient descent.
- [E] What does derivative represent?
- [M] What’s the difference between derivative, gradient, and Jacobian?
-
[H] Say we have the weights
$$w \in R^{d \times m}$$ and a mini-batch$$x$$ of$$n$$ elements, each element is of the shape$$1 \times d$$ so that$$x \in R^{n \times d}$$ . We have the output$$y = f(x; w) = xw$$ . What’s the dimension of the Jacobian$$\frac{\delta y}{\delta x}$$ ? -
[H] Given a very large symmetric matrix A that doesn’t fit in memory, say
$$A \in R^{1M \times 1M}$$ and a function$$f$$ that can quickly compute$$f(x) = Ax$$ for$$x \in R^{1M}$$ . Find the unit vector$$x$$ so that$$x^TAx$$ is minimal.Hint: Can you frame it as an optimization problem and use gradient descent to find an approximate solution?