Command Palette

Search for a command to run...

Linear Methods of AI

Linear Equilibrium Problem

Problem Definition

Imagine you're trying to match a puzzle with imperfect pieces. That's the situation we face in the linear equilibrium problem. We have a matrix ARm×nA \in \mathbb{R}^{m \times n} and vector bRmb \in \mathbb{R}^m, but the equation system Ax=bA \cdot x = b may not have an exact solution.

In cases like this, we search for the best solution we can obtain by minimizing the error function

F(x)=Axb22F(x) = \|A \cdot x - b\|_2^2

This function measures how far the product result AxA \cdot x is from our desired target bb.

Optimization Problem

Now let's understand what we actually want to achieve. The minimization problem

minxAxb22\min_x \|A \cdot x - b\|_2^2

is the core of the linear equilibrium problem or what is often called the linear least squares problem. Imagine looking for the best position to throw a ball so it lands as close as possible to the target, even though you can't hit it exactly.

Why do we use this approach? Because in many real situations, the data we have contains noise or disturbances that make finding an exact solution impossible.

Formula Expansion

Now let's break down this formula to see what actually happens inside it

minxAxb22=minxi=1n(Axb)i2\min_x \|A \cdot x - b\|_2^2 = \min_x \sum_{i=1}^{n} (A \cdot x - b)_i^2

This expanded form shows that we're actually summing the squares of each error component. Similar to calculating the total squared distance between several prediction points and the actual target points. By squaring each error, we give a larger penalty to big errors compared to small ones.

Alternative Norms

It turns out there are several different ways to measure error, depending on the characteristics of the problem we're dealing with.

1\ell_1-norm works by summing the absolute values of errors

minxAxb1=minxi=1n(Axb)i\min_x \|A \cdot x - b\|_1 = \min_x \sum_{i=1}^{n} |(A \cdot x - b)_i|

This approach is like measuring distance by walking in a city with grid-shaped streets. You have to move horizontally and vertically only, no diagonal movement. This method is more resistant to extreme outlier data.

\ell_\infty-norm focuses on the largest error

minxAxb=minxmaxi=1,,n(Axb)i\min_x \|A \cdot x - b\|_\infty = \min_x \max_{i=1,\ldots,n} |(A \cdot x - b)_i|

Imagine you're adjusting the height of an uneven table. This method will focus on fixing the highest or lowest table leg, ensuring no single part is too extreme.

The right choice of norm depends on the type of error we most want to avoid and the characteristics of the data we have.

Geometric Interpretation

Linear Equilibrium Problem Solution
Geometric illustration showing the relationship between vector bb, projection Ax^A\hat{x}, and error vector.

This geometric visualization helps us understand what's actually happening. The flat plane in this diagram represents the column space of matrix AA (called Image A), which is all possible results of multiplication AxA \cdot x.

Vector bb is the target we want to reach, but it may not lie within the column space of matrix AA. The optimal solution x^\hat{x} produces Ax^A\hat{x} which is the closest point to bb in that column space.

Vector Ax^bA\hat{x} - b shows the error vector connecting the best projection with the original target. Like a shadow of an object falling to the floor, Ax^A\hat{x} is the closest "shadow" of bb in the available space.

Solution Methods

Each type of norm requires a different solution approach. Problems with 1\ell_1-norm and \ell_\infty-norm can be solved using linear optimization techniques, where we transform the problem into a form that can be handled by linear programming algorithms.

Meanwhile, the least squares problem with Euclidean norm has a special advantage. When errors in the data follow a normal distribution pattern (bell-shaped), then the solution we obtain provides the best estimate in a statistical sense, called the maximum likelihood estimator.