Close Menu
    Trending
    • Automating Deep Learning: A Gentle Introduction to AutoKeras and Keras Tuner
    • From Equal Weights to Smart Weights: OTPO’s Approach to Better LLM Alignment
    • The Future of AI Agent Communication with ACP
    • Vad världen har frågat ChatGPT under 2025
    • Google’s generative video model Veo 3 has a subtitles problem
    • MedGemma – Nya AI-modeller för hälso och sjukvård
    • AI text-to-speech programs could “unlearn” how to imitate certain people
    • AI’s giants want to take over the classroom
    ProfitlyAI
    • Home
    • Latest News
    • AI Technology
    • Latest AI Innovations
    • AI Tools & Technologies
    • Artificial Intelligence
    ProfitlyAI
    Home » A Bird’s-Eye View of Linear Algebra: Measure of a Map — Determinants
    Artificial Intelligence

    A Bird’s-Eye View of Linear Algebra: Measure of a Map — Determinants

    ProfitlyAIBy ProfitlyAIJune 10, 2025No Comments19 Mins Read
    Share Facebook Twitter Pinterest LinkedIn Tumblr Reddit Telegram Email
    Share
    Facebook Twitter LinkedIn Pinterest Email


    chapter within the in-progress ebook on linear algebra. The desk of contents to date:

    1. Chapter-1: The basics
    2. Chapter-2: Measure of a map (present)

    Keep tuned for future chapters.

    Linear algebra is the device of many dimensions. It doesn’t matter what you may be doing, as quickly as you scale to ( n ) dimensions, linear algebra comes into the image.

    Within the previous chapter, we described summary linear maps. On this one, we roll up our sleeves and begin to take care of matrices. Sensible issues like numerical stability, environment friendly algorithms, and so forth. will now begin to be explored.

    Be aware: all photographs on this article, until in any other case acknowledged are by the creator.

    I) Easy methods to quantify a linear map

    Determinants are one of the vital historic ideas in linear algebra. The roots of the topic lay in fixing methods of linear equations. And determinants would “decide” if there even was an answer price in search of. However in a lot of the instances, the place the system does have an answer, it supplies additional helpful data. Within the fashionable framework of linear maps, determinants present a single quantification of linear maps.

    We mentioned within the previous chapter the idea of vector areas (principally n-dimensional collections of numbers — and extra usually collections of fields) and linear maps that function on two of these vector areas, taking objects in a single to the opposite.

    For example of those sorts of maps, one vector area could possibly be the floor of the planet you’re sitting on and the opposite could possibly be the floor of the desk you may be sitting at. Literal maps of the world are additionally maps on this sense since they “map” each level on the floor of the Earth to a degree on a paper or floor of a desk, though they aren’t linear maps since they don’t protect relative areas (Greenland seems a lot bigger than it’s for instance in a number of the projections).

    An precise map of the floor of the Earth can be a map within the sense of linear algebra, however it isn’t a linear map. Picture by midjourney.

    As soon as we decide a basis for the vector area (a group of n “unbiased” vectors within the area; there could possibly be infinite selections basically), all linear maps on that vector area get distinctive matrices assigned to them.

    In the interim, let’s prohibit our consideration to maps that take vectors from an 𝑛-dimensional area again to the 𝑛-dimensional area (we’ll generalize later). The matrices corresponding to those linear maps are 𝑛×𝑛 (see part III of chapter 1). It may be helpful to “quantify” such a linear map, specific its impact on the vector area, ℝⁿ in a single quantity. The sort of map we’re coping with, successfully takes vectors from ℝⁿ and “distorts” them into another vectors in the identical area. Each the unique vector 𝑣 and the vector 𝑢 that the map transformed it into have some lengths (say |𝑣| and |𝑢|). We are able to take into consideration how a lot the size of the vector is modified by the map, |𝑢|∕|𝑣|. Perhaps that may quantify the affect of the map? How a lot it “stretches” vectors?

    This strategy has a deadly flaw. The ratio relies upon not simply on the linear map, but in addition on the vector 𝑣 it acts on. It’s subsequently not strictly a property of the linear map itself.

    What if we take two vectors as an alternative now, 𝑣₁ and 𝑣₂ that are transformed by the linear map into the vectors 𝑢₁ and 𝑢₂. Simply because the measure of the one vector, 𝑣 was its size, the measure of two vectors is the world of the parallelogram contained between them.

    The world of the parallelogram shaped by two vectors. Picture by midjourney.

    Simply as we thought-about the quantity by which the size of 𝑣 modified, we are able to now discuss by way of the quantity by which the world between 𝑣₁ and 𝑣₂ modifications as soon as they cross by the linear map and grow to be 𝑢₁, 𝑢₂. And alas, this once more relies upon not simply on the linear map, but in addition the vectors chosen.

    Subsequent, we are able to go to a few vectors and take into account the change in quantity of the parallelepiped between them and run into the identical downside of the preliminary vectors having a say.

    A 3 dimensional area in three dimensional area. If a linear map acts on these three vectors, the amount modifications by the identical quantity it doesn’t matter what the preliminary selection of the vectors. Picture by midjourney.

    However now take into account an n-dimensional area within the authentic vector area. This area could have some “n-dimensional measure”. To know this, a two dimensional measure is an space (measured in sq. kilometers). A 3 dimensional measure is the amount used for measuring water (in liters). A 4 dimensional measure has no counterpart within the bodily world we’re used to, however is simply as mathematically sound, a measure of the quantity of 4 dimensional area enclosed inside a parallelepiped shaped of 4 4- d vectors and so forth.

    The measure in a 2-d area is space and that in three-D area is quantity. These notions may be prolonged to 4-d area and up. Picture by midjourney

    The 𝑛 authentic vectors (𝑣₁, 𝑣₂, …, 𝑣ₙ) type a parallelepiped which is reworked by the linear map into 𝑛 new vectors, 𝑢₁, 𝑢₂, …, 𝑢ₙ which type their very own parallelepiped. We are able to then ask concerning the 𝑛-dimensional measure of the brand new area in relation to the unique one. And this ratio, it seems, is certainly a operate solely of the linear map. No matter what the unique area seemed like, the place it was positioned and so forth, the ratio of its measure as soon as the linear map acted on it to its measure earlier than would be the identical — a operate purely of the linear map. This ratio of 𝑛-dimensional measures (after to earlier than) then is what we’ve been in search of: an unique property of the linear map that quantifies its impact in a single quantity.

    This ratio by which the measure of any 𝑛-dimensional patch of area is modified by the linear map is an effective approach to quantify the impact it has on the area it acts on. It’s known as the determinant of the linear map (the rationale for that identify will grow to be obvious in part V).

    For now, we merely acknowledged the truth that the quantity by which a linear map from ℝⁿ to ℝⁿ “stretches” any patch of 𝑛-dimensional area relies upon solely on the map with out providing a proof for the reason that objective right here was motivation. We’ll cowl a proof later (part VI), as soon as we arm ourselves with some weapons.

    II) Calculating determinants

    Now, how do we discover this determinant given a linear map from the vector area ℝⁿ again to ℝⁿ? We are able to take any 𝑛 vectors, discover the measure of the parallelepiped between them and the measure of the brand new parallelepiped as soon as the linear map has acted on all of them. Lastly, divide the latter by the previous.

    We have to make these steps extra concrete. First, let’s begin taking part in round on this ℝⁿ vector area.

    The ℝⁿ vector area is only a assortment of 𝑛 actual numbers. The only vector is simply 𝑛 zeros — [0, 0, …, 0]. That is the zero vector. If we multiply a scalar with it, we simply get the zero vector again. Not attention-grabbing. For the following easiest vector, we are able to change the primary 0 with a 1. This results in the vector: 𝑒₁ = [1, 0, 0, …, 0]. Now, multiplying by a scalar, 𝑐 offers us a distinct vector.

    $$c.[1, 0, 0,.., 0] = [c, 0, 0, …, 0]$$

    We are able to “span” an infinite variety of vectors with 𝑒₁ relying on the scalar 𝑐 we select.

    If 𝑒₁ is the vector with simply the primary factor being 1 and the remaining being 0, then what’s 𝑒₂? The second factor being 1 and the remaining being 0 looks as if a logical selection.

    $$e_2 = [0,1,0,0,dots 0]$$

    Taking this to its logical conclusion, we get a group of n vectors:

    These vectors type a foundation of the vector area that’s ℝⁿ. What does this imply? Any vector 𝑣 in ℝⁿ may be expressed as a linear mixture of those 𝑛 vectors. Which implies that for some scalars 𝑐₁, 𝑐₂, …, 𝑐ₙ:

    $$v = c_1.e_1+c_2.e_2+dots +c_n.e_n$$

    All vectors, 𝑣 are “spanned” by the set of vectors 𝑒₁, 𝑒₂, …, 𝑒ₙ.

    This specific assortment of vectors isn’t the one foundation. Any set of 𝑛 vectors works. The one caveat is that not one of the 𝑛 vectors ought to be “spanned” by the remaining. In different phrases, all of the 𝑛 vectors ought to be linearly unbiased. If we select 𝑛 random numbers from most steady distributions and repeat the method 𝑛 occasions to create the 𝑛 vectors, you’re going to get a set of linearly unbiased vectors with 100% likelihood (“nearly absolutely” in likelihood phrases). It’s simply very, most unlikely {that a} random vector occurs to be “spanned” by another 𝑘 < 𝑛 random vectors.

    Going again to our recipe initially of this part to search out the determinant of a linear map, we now have a foundation to precise our vectors in. Fixing the premise additionally means our linear map may be expressed as a matrix (see part III of chapter 1). Since this linear map is taking vectors from ℝⁿ again to ℝⁿ, the corresponding matrix is 𝑛 × 𝑛.

    Subsequent, we wanted 𝑛 vectors to type our parallelepiped. Why not take the 𝑒₁, 𝑒₂, …, 𝑒ₙ commonplace foundation we outlined earlier than? The measure of the patch of area contained between these vectors occurs to be 1, by very definition. The image under for ℝ³ will hopefully make this clear.

    The usual patch of area contained between vectors e1, e2, e3, …, en. On this case, now we have three vectors for the reason that area is three-dimensional. Picture created with midjourney

    If we acquire these vectors from the usual foundation right into a matrix (rows or columns), we get the id matrix (1’s on the principle diagonal, 0’s all over the place else):

    Eq (0): The id matrix

    Once we mentioned we might apply our linear remodel to any n-dimensional patch of area, we would as nicely apply it to this “commonplace” patch.

    However, it’s simple to point out that multiplying any matrix with the id matrix ends in the identical matrix. So, the ensuing vectors after the linear map is utilized are the columns of the matrix representing the linear map itself. So, the quantity by which the linear map modified the amount of the “commonplace patch” is identical because the n-dimensional measure of the parallelepiped between the column vectors of the matrix representing the map itself.

    To recap, we began by motivating the determinant because the ratio by which a linear map modifications the measure of an n-dimensional patch of area. And now, we confirmed that this ratio itself is an n-dimensional measure. Specifically, the measure contained between the column vectors of any matrix representing the linear map.

    III) Motivating the fundamental properties

    We described within the earlier part how a determinant of a linear map ought to merely be the measure contained between the vectors of any of its matrix representations. On this part, we use two dimensional area (the place measures are areas) to encourage some elementary properties a determinant will need to have.

    The primary property is multi-linearity. A determinant is a operate that takes a bunch of vectors (collected in a matrix) and maps them to a single scalar. Since we’re limiting to two-dimensional area, we’ll take into account two vectors, each two dimensional. Our determinant (since we’ve motivated it to be the world of the parallelogram between the vectors) may be expressed as:

    $$det = A(v_1, v_2)$$

    How ought to this operate behave if we add a vector to one of many two vectors? The multi-linearity property requires:

    $$A(v_1+v_3, v_2) = A(v_1,v_2)+A(v_3,v_2)tag{1}$$
    That is obvious from the transferring image under (observe the brand new space getting added).

    The additive property of a determinant. Picture by creator

    And this visualization may also be used to see (by scaling one of many vectors as an alternative of including one other vector to it):
    $$A(c.v_1, v_2) = c.A(v_1, v_2) tag{2}$$
    This second property has an necessary implication. What if we plug a unfavourable c into the equation?

    The world, 𝐴(𝑣₁, 𝑣₂) ought to then be the other signal to 𝐴(𝑐·𝑣₁, 𝑣₂).

    Which suggests we have to introduce the notion of unfavourable space and a unfavourable determinant.

    This makes quite a lot of sense if we’re okay with the idea of unfavourable lengths. If lengths — measures in 1-D area — may be constructive or unfavourable, then it stands to cause that areas — measures in 2-D area — must also be allowed to be unfavourable. And so, measures in area of any dimensionality ought to as nicely.

    Collectively, equations (1) and (2) are the multi-linearity property.

    One other necessary property that has to do with the signal of the determinant is the alternating property. It requires:

    $$A(v_1, v_2) = -A(v_2, v_1)$$

    Swapping the order of two vectors negates the signal of the determinant (or measure between them). Should you discovered concerning the cross product of 3-D vectors, this property might be very pure. To encourage it, let’s assume first of the one-dimensional distance between two place vectors, 𝑑(𝑣₁, 𝑣₂). It’s clear that 𝑑(𝑣₁, 𝑣₂) = −𝑑(𝑣₂, 𝑣₁) since once we go from 𝑣₂ to 𝑣₁, we’re touring in the other way to once we go from 𝑣₁ to 𝑣₂. Equally, if the world spanned between vectors 𝑣₁ and 𝑣₂ is constructive, then that between 𝑣₂ and 𝑣₁ should be unfavourable. This property holds in 𝑛-dimensional area as nicely. If in 𝐴(𝑣₁, 𝑣₂, …, 𝑣ₙ) we swap two of the vectors, it causes the signal to modify.

    The alternating property additionally implies that if one of many vectors is solely a scalar a number of of the opposite, the determinant should be 0. It’s because swapping the 2 vectors ought to negate the determinant:

    $$start{align}A(v_1, v_1) = -A(v_1, v_1)
    => 2 A(v_1, v_1) = 0
    => A(v_1, v_1) = 0end{align}$$

    We even have by multi-linearity (equation 2):
    $$A(v_1, c.v_1) = c A(v_1, v_1) = 0$$
    This is smart geometrically since if two vectors are parallel to one another, the world between them is ( 0 ).

    The video [6] covers the geometric motivation of those properties with actually good visualizations and video [4] visualizes the alternating property fairly nicely.

    IV) Getting algebraic: Deriving the Leibniz formulation

    On this part, we transfer away from geometric instinct and strategy the subject of determinants from an alternate route — that of chilly, algebraic calculations.

    See, the multi-linearity and alternating properties which we motivated within the final part with geometry are (remarkably) sufficient to present us a really particular algebraic formulation for the determinant, known as the Leibniz formulation.

    That formulation helps us see properties of the determinant that will be actually, actually arduous to look at from the geometric strategy or with different algebraic formulation.

    The Leibniz formulation can then be lowered to the Laplace growth, involving going alongside a row or column and calculating cofactors — which many individuals see in highschool.

    Let’s derive the Leibniz formulation. We want a operate that takes the 𝑛 column vectors, 𝛼₁, 𝛼₂, …, 𝛼ₙ of the matrix as enter and converts them right into a scalar, 𝑐.

    $$c=f(vec{a_1}, vec{a_2}, dots vec{a_n})$$

    We are able to specific every column vector by way of the usual foundation of the area.

    Now, we are able to apply the property of multi-linearity. For now, to the primary column, 𝛼₁.

    We are able to do the identical for the second column. Let’s take simply the primary time period from the summation above and check out the ensuing phrases.

    Be aware that within the first time period, we get the vector 𝑒₁ showing twice. And by the alternating property, the operate 𝑓 for that time period turns into 0.

    To ensure that two 𝑒₁’s to seem, the second indices of the 2 𝑎’s within the product should every grow to be 1.

    So, as soon as we do that for all of the columns, the phrases that received’t grow to be zero by the alternating property would be the ones the place the second indices of the 𝑎’s don’t have any repetition — so all distinct numbers from 1 to 𝑛. In different phrases, we’re in search of permutations of 1 to 𝑛 to seem within the second indices of the 𝑎’s.

    What concerning the first indices of the 𝑎’s? These are merely the numbers 1 to 𝑛 so as since we pull out the 𝑎₁ₓ’s first, then the 𝑎₂ₓ’s, and so forth. In additional compact algebraic notation,

    Within the expression on the suitable, the areas 𝑓(𝑒_{𝑗₁}, 𝑒_{𝑗₂}, …, 𝑒_{𝑗ₙ}) can both be +1, −1, or 0 for the reason that 𝑒ⱼ’s are all unit vectors orthogonal to one another. We already established that any time period that has any repeated 𝑒ⱼ’s will grow to be 0, leaving us with simply permutations (no repetition). Amongst these permutations, we are going to typically get +1 and typically −1.

    The idea of permutations carries with it signs. The indicators of the areas are equal to the indicators of the permutations. If we denote by 𝑆ₙ the set of all permutations of [1, 2, …, 𝑛], then we get the Leibniz formulation of the determinant:

    $$det([vec{a_1}, vec{a_2}, dots vec{a_n}]) = |A| = sumlimits_{sigma in S_n} sgn(sigma) prod limits_{i=1}^n a_{i,sigma(i)} tag{3}$$

    This formulation can be described intimately in mathexchange post, [3]. And to make issues concrete, right here is a few easy Python code that implements it (together with a take a look at case).

    One shouldn’t really use this formulation to calculate the determinant of a matrix (until it’s only for enjoyable or exposition). It really works, however is comically inefficient given the sum over all permutations (which is 𝑛!, which is super-exponential).

    Nonetheless, many theoretical properties of the determinant grow to be trivial to see with the Leibniz formulation after they could be very arduous to decipher or show if we began from one other of its types. For instance:

    1. Proposition-1: With this formulation it turns into obvious {that a} matrix and its transpose have the identical determinant: |𝐴| = |𝐴ᵀ|. It’s a easy consequence of the symmetry of the formulation.
    2. Proposition-2: A really related derivation to the above can be utilized to point out that for 2 matrices 𝐴 and 𝐵, |𝐴𝐵| = |𝐴| ⋅ |𝐵|. See this answer within the mathexchange post, [8]. This can be a very handy property since matrix multiplication comes up on a regular basis in varied decompositions of matrices, and reasoning concerning the determinants of these decompositions generally is a highly effective device.
    3. Proposition-3: With the Leibniz formulation, we are able to simply see that if the matrix is higher triangular or decrease triangular (decrease triangular means each factor of the matrix above the diagonal is zero), the determinant is solely the product of the entries on the diagonal. It’s because all permutations bar one: (𝑎₁₁ ⋅ 𝑎₂₂ ⋯ 𝑎ₙₙ) (the principle diagonal) get some zero time period or the opposite and make their phrases within the summation 0.
    An higher triangular matrix. All entries under the principle diagonal are 0.

    The third reality really results in essentially the most environment friendly algorithm for calculating a determinant that the majority linear algebra libraries use. A matrix may be decomposed effectively into decrease and higher triangular matrices (known as the LU decomposition which we’ll cowl within the subsequent chapter). After doing this decomposition, the third reality is used to multiply the diagonals of these decrease and higher matrices to get their determinants. And eventually, the second reality is used to multiply these two determinants and get the determinant of the unique matrix.

    Lots of people in highschool or college when first uncovered to the determinant, study concerning the Laplace growth, which includes increasing a couple of row or column, discovering co-factors for every factor and summing. That may be derived from the above Leibniz growth by accumulating related phrases. See this answer to the mathexchange post, [2].

    V) Historic motivation

    The determinant was first found within the context of linear methods of equations. Say now we have 𝑛 equations in 𝑛 variables (𝑥₀, 𝑥₁, …, 𝑥ₙ).

    This technique may be expressed in matrix type:

    And extra compactly:

    $$A.x = b$$

    An necessary query is whether or not or not the system above has a novel resolution, x. And the determinant is a operate that “determines” this. There’s a distinctive resolution if and provided that the determinant of A is non-zero.

    This traditionally impressed strategy motivates the determinant as a polynomial that arises once we attempt to remedy a linear system of equations related to the linear map. We are going to cowl this in additional depth in chapter 5.

    For extra on this, see the wonderful reply within the mathexchange post, [8].

    VI) Proof of the property we motivated with

    We began this chapter by motivating the determinant as the quantity by which the ℝⁿ → ℝⁿ linear map modifications the measure of an n-dimensional patch of area. We additionally mentioned that this doesn’t work for 1, 2, … n − 1 dimensional measures. Beneath is a proof of this the place we use a number of the properties we encountered in the remainder of the sections.

    Outline (𝑉, 𝑈) as 𝑛 × 𝑘 matrices, the place

    $$ V = (v_1, v_2, dots, v_k) $$

    By definition,

    $$|v_1, v_2, dots, v_k| = sqrt{det(V^t V)} $$ and

    $$ |u_1, u_2, dots, u_k| = sqrt{det(U^t U)} = sqrt{det((AV)^t (AV))} = sqrt{det(V^t A^t A V)} $$

    Solely when n = ok is V is a sq. matrix, so

    $$|v_1, v_2, dots, v_k| = sqrt{det(V^t A^t A V)}$$

    $$= sqrt{det(V^t) det(A^t) det(A) det(V)} $$
    $$= det(A) sqrt{det(V^t V)} = det(A) |v_1, v_2, dots, v_k| $$

    References

    [1] Mathexchange put up: Determinant of a linear map doesn’t rely upon the bases: https://math.stackexchange.com/questions/962382/determinant-of-linear-transformation

    [2] Mathexchange put up: Determinant of a matrix Laplace growth (highschool formulation) https://math.stackexchange.com/a/4225580/155881

    [3] Mathexchange put up: Understanding Leibniz formulation for determinants https://math.stackexchange.com/questions/319321/understanding-the-leibniz-formula-for-determinants#:~:text=The%20formula%20says%20that%20det,permutation%20get%20a%20minus%20sign.&text=where%20the%20minus%20signs%20correspond%20to%20the%20odd%20permutations%20from%20above.

    [4] Youtube video: 3B1B on determinants https://www.youtube.com/watch?v=Ip3X9LOh2dk&t=295s

    [5] Connecting Leibniz formulation with geometry https://math.stackexchange.com/questions/593222/leibniz-formula-and-determinants

    [6] Youtube video: Leibniz formulation is space: https://www.youtube.com/watch?v=9IswLDsEWFk

    [7] Mathexchange put up: product of determinants is determinant of product https://math.stackexchange.com/questions/60284/how-to-show-that-detab-deta-detb

    [8] Historic context for motivating determinant: https://math.stackexchange.com/a/4782557/155881



    Source link

    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Previous ArticleMastering SQL Window Functions | Towards Data Science
    Next Article The Hidden Dangers of Open-Source Data: Rethinking Your AI Training Strategy
    ProfitlyAI
    • Website

    Related Posts

    Artificial Intelligence

    Automating Deep Learning: A Gentle Introduction to AutoKeras and Keras Tuner

    July 15, 2025
    Artificial Intelligence

    From Equal Weights to Smart Weights: OTPO’s Approach to Better LLM Alignment

    July 15, 2025
    Artificial Intelligence

    The Future of AI Agent Communication with ACP

    July 15, 2025
    Add A Comment
    Leave A Reply Cancel Reply

    Top Posts

    Amazon Just Jumped Into the AI Agent Race

    April 10, 2025

    Agentic GraphRAG for Commercial Contracts

    April 3, 2025

    Why humanoid robots need their own safety rules

    June 11, 2025

    Scaling Human-in-the-Loop: Overcoming AI Evaluation Challenges

    April 9, 2025

    Make Your Data Move: Creating Animations in Python for Science and Machine Learning

    May 6, 2025
    Categories
    • AI Technology
    • AI Tools & Technologies
    • Artificial Intelligence
    • Latest AI Innovations
    • Latest News
    Most Popular

    Vad världen har frågat ChatGPT under 2025

    July 15, 2025

    Nyfiken på GPT-4.1 -Så här testar du den på Poe och Polychat

    April 16, 2025

    The Art of the Phillips Curve

    May 12, 2025
    Our Picks

    Automating Deep Learning: A Gentle Introduction to AutoKeras and Keras Tuner

    July 15, 2025

    From Equal Weights to Smart Weights: OTPO’s Approach to Better LLM Alignment

    July 15, 2025

    The Future of AI Agent Communication with ACP

    July 15, 2025
    Categories
    • AI Technology
    • AI Tools & Technologies
    • Artificial Intelligence
    • Latest AI Innovations
    • Latest News
    • Privacy Policy
    • Disclaimer
    • Terms and Conditions
    • About us
    • Contact us
    Copyright © 2025 ProfitlyAI All Rights Reserved.

    Type above and press Enter to search. Press Esc to cancel.