# The Azimuth Project Bilinear regression (Rev #4, changes)

Showing changes from revision #3 to #4: Added | Removed | Changed

# Bilinear regression

## Idea

In statistics/machine learning the individual samples often come in the form of 2-D arrays, eg, a set of population counts of different species (one axis) at different points in time (second axis). Standard regression collapses these arrays into vectors and thus loses the structure in the regression process. Bilinear regression attempts to use the array structure by using the samples as matrices.

## Details

### Basic model

The bilinear predictor function takes the form

(1)$f(X) = tr(U^T X V) + b = \sum_{i=1:m} u^T_i X v_i + b$

When performing regularised fitting the score function is

(2)$E = \sum_{j=1:n} \left(\sum_{i=1:m} u^T_i X_j v_i + b - y_j\right)^2 + \lambda \sum_{i=1:m} R(u_i) + R(v_i)$

where $\lambda$ is the regularization strength and $R()$ is the regularization function.

The derivatives are

(3) \frac{\partial E}{\partial v_I}=2 \sum_{j=1:n} \left(\sum_{i=1:m} u^T_i X_j v_i + b - y_j\right)^2 y_j\right) (u^T_I X_j)^T + \lambda \frac{\partial R(v_I)}{\partial v_I}

and

(4) \frac{\partial E}{\partial u_I}=2 \sum_{j=1:n} \left(\sum_{i=1:m} u^T_i X_j v_i + b - y_j\right)^2 y_j\right) (X_j v_I) + \lambda \frac{\partial R(u_I)}{\partial u_I}

Finally for the constant term $b$ the derivative is

(5)$\frac{\partial E}{\partial b}=2 \sum_{j=1:n} \left(\sum_{i=1:m} u^T_i X_j v_i + b - y_j\right)$

so that given the other variables the optimal $b$ is

(6)$b = \frac{-2}{n} \sum_{j=1:n} \left(\sum_{i=1:m} u^T_i X_j v_i - y_j\right)$