# Nakafa Framework: LLM URL: /en/subject/university/bachelor/ai-ds/linear-methods/regularization Source: https://raw.githubusercontent.com/nakafaai/nakafa.com/refs/heads/main/packages/contents/subject/university/bachelor/ai-ds/linear-methods/regularization/en.mdx Output docs content for large language models. --- export const metadata = { title: "Regularization", description: "Master Tikhonov regularization for unstable linear systems. Learn to solve ill-conditioned problems, prevent overfitting, and stabilize parameter estimation.", authors: [{ name: "Nabil Akbarazzima Fatih" }], date: "07/15/2025", subject: "Linear Methods of AI", }; ## Problems in Linear Systems When we deal with linear equation systems where and , challenging situations often arise. If and , then the least squares system becomes unsolvable because the system is too constrained or has too many restrictions. Another equally problematic situation occurs when matrix does not have full rank, i.e., . In this condition, the equation system becomes under-constrained or has too much freedom. Imagine trying to determine the position of an object with too little or conflicting information. Regularization emerges as a solution to provide stability to these unstable problems. ## Definition of Regularization Problem To address the instability problem, we introduce a modified least squares problem where is the initial value or prior estimate for the model parameters and is the weighting factor. The additional term is called the Tikhonov regularization term. This regularization term is like giving a "preference" to the system to choose solutions that are not too far from the initial estimate . The larger the value of , the stronger this preference becomes. ## Interpretation of Regularization Through the regularization term, the least squares problem not only minimizes the difference between model and data, but also minimizes the difference between parameters and the prior estimate , weighted by . Note that the prior estimate is chosen by the researcher. The solution then not only describes the behavior of the process being investigated, but also reflects the researcher's initial assumptions. ## Matrix Formulation The regularization problem can be written in matrix form as The corresponding normal equation system becomes or in simpler form ## Properties of Regularization Solution For , the normal equation system of the regularization problem always has a unique solution. Regularization thus restores the identifiability of all parameters. The matrix has linearly independent rows in the block for , thus achieving maximum rank . The matrix becomes positive definite for , ensuring that the problem becomes well-defined and has a stable solution. ## Individual Weights for Parameters We can choose individual weighting factors for each parameter . In this case, the least squares problem becomes with diagonal matrix The weighting factors are chosen such that the matrix has full rank. ## Weight Selection Strategy For parameters that are difficult to determine well, we choose large weighting factors . Conversely, for parameters that can already be determined well, we can choose . Of course, all weighting factors can influence all parameters. If we decide to fix a parameter to a specific value or turn it into a constant, we can set the factor in principle. This also applies when we add inequality conditions to the problem, which are then satisfied in the solution with equations or . Through regularization, the solution depends not only on the data, but also on the researcher's initial assumptions. This provides flexibility in integrating domain knowledge into the parameter estimation process.