10. Approximation and Least Squares

Overdetermined Systems

Overdetermined Systems in Numerical Analysis

students, imagine trying to solve a mystery with more clues than unknowns πŸ”Ž. In many math problems, that extra information is useful, but it can also make the system impossible to satisfy exactly. That is the big idea behind overdetermined systems. In this lesson, you will learn what they are, why they appear in real life, and how they connect to least squares and best-fit approximations.

What you will learn

By the end of this lesson, students, you should be able to:

  • explain what an overdetermined system is and why it matters,
  • recognize when a system has more equations than unknowns,
  • understand why exact solutions may not exist,
  • connect overdetermined systems to least squares methods,
  • use examples from data and measurement to describe best-fit ideas πŸ“ˆ.

What is an overdetermined system?

An overdetermined system is a system of equations that has more equations than unknowns. If there are $m$ equations and $n$ unknowns, then the system is overdetermined when $m > n$.

A linear system can be written in matrix form as

$$Ax = b$$

where $A$ is an $m \times n$ matrix, $x$ is the vector of unknowns, and $b$ is the vector of known values. When $m > n$, the matrix has more rows than columns, so the system usually has more constraints than variables.

This is important because each equation asks the unknowns to satisfy one condition. With many equations and only a few variables, the conditions may conflict. For example, one measurement might suggest one value, while another measurement suggests something slightly different.

A simple example

Suppose we want to find a number $x$ that satisfies all three equations:

$$x = 2$$

$$x = 3$$

$$x = 4$$

This is clearly impossible, because one number cannot equal three different values at the same time. This system is overdetermined because there are three equations and only one unknown.

In real life, overdetermined systems often arise when data are collected from experiments, surveys, sensors, or repeated measurements. The data are not always perfectly consistent because of rounding, noise, or small errors. That is why numerical methods become so useful.

Why overdetermined systems often have no exact solution

When a system is overdetermined, it may have no solution at all. In matrix language, that means there is no vector $x$ such that $Ax = b$ exactly.

Why does this happen? The columns of $A$ represent the patterns the unknowns can create. The vector $b$ must fit perfectly inside the column space of $A$ for an exact solution to exist. But when the equations are based on real measurements, $b$ may be slightly off, so it does not lie exactly in that space.

Think of trying to draw a straight line through several points on a graph. If all the points lie exactly on one line, great. But if the points are scattered a little, no single line passes through all of them. The data may still be close to a line, just not perfect. That is where approximation comes in ✨.

Geometry idea

Each equation in a linear system can be viewed as a geometric condition. For example, in two unknowns, each equation represents a line. With more equations than unknowns, we may get several lines that do not all intersect at one point. The system is inconsistent, meaning there is no common solution.

In higher dimensions, the same idea applies. You can picture each equation as a hyperplane, and an overdetermined system asks for one point that lies on all of them. If the equations disagree even a little, such a point may not exist.

Overdetermined systems and least squares

When no exact solution exists, numerical analysis asks a different question: Which solution is closest? This leads to the least squares method.

Instead of demanding $Ax = b$, we try to make the error vector

$$r = b - Ax$$

as small as possible. This vector $r$ is called the residual. It measures the difference between the actual data $b$ and the values predicted by $Ax$.

The least squares method chooses $x$ to minimize the sum of the squared residuals:

$$\|b - Ax\|^2$$

This means we want the total error to be as small as possible. Squaring the errors is useful because it avoids cancellation between positive and negative values and gives a smooth function to optimize.

Why squares?

If one residual is $2$ and another is $-2$, their sum is $0$, but the errors are still there. Squaring gives $4$ and $4$, which correctly reflects the size of the mistakes. The total squared error is then $8$.

Least squares does not usually force every equation to be correct. Instead, it finds the best compromise among conflicting equations. This is exactly what is needed for data with measurement error.

A small numerical example

Suppose we want to estimate a number $x$ from the measurements

$$x \approx 1$$

$$x \approx 2$$

$$x \approx 4$$

There is no single number that matches all three values exactly. If we choose $x = 1$, the residuals are $0$, $1$, and $3$. If we choose $x = 2$, the residuals are $-1$, $0$, and $2$. If we choose $x = 3$, the residuals are $-2$, $-1$, and $1$.

Now compare the squared residual sums:

  • for $x = 1$, the total is $0^2 + 1^2 + 3^2 = 10$,
  • for $x = 2$, the total is $(-1)^2 + 0^2 + 2^2 = 5$,
  • for $x = 3$, the total is $(-2)^2 + (-1)^2 + 1^2 = 6$.

So $x = 2$ gives the smallest squared error among these choices. That makes it the best fit in the least squares sense.

In a more advanced setting, the best-fit solution often satisfies the normal equations:

$$A^T A x = A^T b$$

These equations come from setting the derivative of the squared error function to zero. They are central in numerical analysis, although in practice, stable algorithms such as QR decomposition are often preferred for computation.

Best-fit approximations in data problems

Overdetermined systems are everywhere in data analysis. Suppose students is tracking the growth of a plant over several days 🌱. The measurements might be:

  • Day 1: $3.1$ cm,
  • Day 2: $4.0$ cm,
  • Day 3: $5.2$ cm,
  • Day 4: $6.1$ cm.

If we want a line to model the growth, we may write

$$y = mx + c$$

and use the data points to build an overdetermined system for $m$ and $c$. Since there are more data points than parameters, an exact line through every point may not exist. Least squares gives the line that best matches the data overall.

This is useful in science and engineering because data are rarely perfect. Sensors can be slightly off. Human measurements can vary. Weather, noise, and rounding all add small differences. An overdetermined system models that reality well because it accepts that the equations may not agree exactly.

Real-world examples

  • Physics: estimating the path of an object from noisy measurements.
  • Economics: fitting a trend line to prices or sales data.
  • Computer vision: matching observed points to a model.
  • Engineering: calibrating a device using many measurements.

In each case, the goal is not exact satisfaction of every equation. The goal is a model that gives the smallest overall error.

How overdetermined systems fit into approximation and least squares

Overdetermined systems are a major reason least squares exists. They connect the idea of equations to the idea of approximation.

Here is the big picture:

  1. You start with too many equations for exact solving.
  2. The system may have no exact solution.
  3. You measure how far a trial solution is from the data using residuals.
  4. You choose the solution that minimizes the error.

So, an overdetermined system is not a failure of mathematics. It is a sign that the problem belongs to approximation. Numerical analysis studies how to find the best approximate solution efficiently and accurately.

This topic also prepares you for later ideas such as fitting polynomials, analyzing data, and solving optimization problems. The same concept appears whenever we ask, β€œWhat is the best model for these measurements?” πŸ€”

Conclusion

Overdetermined systems are systems with more equations than unknowns, usually written as $Ax = b$ with $m > n$. Because they have more constraints than variables, they often have no exact solution. In numerical analysis, this is not a dead end. Instead, it leads to least squares, where we search for the solution that minimizes $\|b - Ax\|^2$. This creates a best-fit approximation that is especially useful for real-world data, where measurements are often inconsistent or noisy. students, understanding overdetermined systems gives you a foundation for data fitting, modeling, and many practical applications in mathematics and science.

Study Notes

  • An overdetermined system has more equations than unknowns, so $m > n$.
  • It is usually written in matrix form as $Ax = b$.
  • Such a system may have no exact solution because the equations can conflict.
  • The difference between observed values and predicted values is the residual $r = b - Ax$.
  • The least squares method chooses $x$ to minimize $\|b - Ax\|^2$.
  • The best-fit solution often satisfies the normal equations $A^T A x = A^T b$.
  • Overdetermined systems are common in data fitting, measurement, science, and engineering.
  • They are a central part of approximation in numerical analysis because real data are rarely perfect.

Practice Quiz

5 questions to test your understanding