I know how basic operations are performed on matrices, I can do transformations, find inverses, etc. But now that I think about it, I actually don't "understand" or know what I've been doing all this time. Our teacher made us memorise some rules and I've been following it like a machine.

  1. So what exactly is a matrix? And what is a determinant?

  2. What do they represent?

  3. Is there a geometrical interpretation?

  4. How are they used? Or, rather, what for are they used?

  5. How do I understand the "properties" of matrix?

I just don't wanna mindlessly cram all those properties, I want to understand them better.

Any links, which would improve my understanding towards determinants and matrices? Please use simpler words. Thanks :)

  • 14,399
  • 4
  • 29
  • 70
  • 4,395
  • 14
  • 37
  • Are you familiar with vectors? – Michael Hoppe May 15 '18 at 19:18
  • 2
    Matrices represent linear maps. – Angina Seng May 15 '18 at 19:20
  • @Michael Hoppe Yes! Infact I understand other 3D space chapters also. – William May 15 '18 at 19:21
  • $m\times n$ matrices can be thought of as representations of linear transformations from $\Bbb R^n$ to $\Bbb R^m$. Multiplication of matrices corresponds to compositions of those linear transformations. Determinants are useful identifiers of square matrices and have various practical applications, for example in determining whether a matrix is invertible or not. – JMoravitz May 15 '18 at 19:22
  • 24
    I'd recommend 3Blue1Browns [essence of linear algebra](https://www.youtube.com/playlist?list=PLZHQObOWTQDPD3MizzM2xVFitgF8hE_ab) playlist. It is really good and is very intuitive. It also answers all the questions you have. It is binge watchable and quite interesting – The Integrator May 15 '18 at 19:24
  • 3
    That is actually a lot of questions that might need thorough research for complete answers but as Lord Shark noticed, many things are explained by the *linear algebra* point of view (especially the multiplication). If you see a matrix as a set of vectors (its columns) then the **determinant is the volume** that is "defined" by those vectors. They are used to solve linear algebra problems (among many, many other things) because it is just more convenient some times. – Bill O'Haran May 15 '18 at 19:28
  • The best way to understand the matrices is to relate them to the linear transformations and the determinant to the linear independence of the vectors. – user296113 May 15 '18 at 19:35
  • Have you looked at these popular posts: https://math.stackexchange.com/q/668/321264, https://math.stackexchange.com/q/250534/321264 ? – StubbornAtom May 15 '18 at 19:55
  • 3
    Professors who teach matrices as grids of numbers on which we define an arbitrary operation which for some reason is called "multiplication", rather than as representations of functions, should be tried in the Hague. I've never understood why so many (read: more than zero) books and lecture notes exist explaining the topic from that point of view. – Jack M May 15 '18 at 23:17
  • 30
    Unfortunately no one can be told what a matrix is. You have to see it for yourself. – HostileFork says dont trust SE May 16 '18 at 02:17
  • 1
    It's already been mentioned, but OP, I highly recommend [3B1B's series](https://www.youtube.com/playlist?list=PLZHQObOWTQDPD3MizzM2xVFitgF8hE_ab). In fact, I should finish watching it later today... – Mateen Ulhaq May 16 '18 at 04:25
  • 1
    In general, mathematics isn't concerned with what things _are_, but with what they _do_. For example, it doesn't matter if real numbers "are" Dedekind cuts or equivalence classes of Cauchy sequences. What matters is that they are an ordered field, have the least upper bound property and so on. As for matrices, the most important thing to know is that finite-dimensional linear transformations can be represented as matrices, and composition of transformations corresponds to matrix multiplication. – Meni Rosenfeld May 16 '18 at 10:26
  • I like to think of the determinant as a quantity that describes both the magnitude of a matrix's elements and the degree of linear independence between its rows (or columns). A matrix with rows that are "almost" dependent will have a small determinant, and a matrix with rows that are truly dependent will have a 0 determinant. – Meni Rosenfeld May 16 '18 at 10:28
  • Related:https://math.stackexchange.com/questions/1811886/what-is-the-most-rigorous-definition-of-a-matrix. – StubbornAtom May 17 '18 at 19:13

4 Answers4


A matrix is a compact but general way to represent any linear transform. (Linearity means that the image of a sum is the sum of the images.) Examples of linear transforms are rotations, scalings, projections. They map points/lines/planes to point/lines /planes.

So a linear transform can be represented by an array of coefficients. The size of the matrix tells you the number of dimension of the domain and the image spaces. The composition of two linear transforms corresponds to the product of their matrices. The inverse of a linear transform corresponds to the matrix inverse.

A determinant measures the volume of the image of a unit cube by the transformation; it is a single number. (When the number of dimensions of the domain and image differ, this volume is zero, so that such "determinants" are never considered.) For instance, a rotation preserves the volumes, so that the determinant of a rotation matrix is always 1. When a determinant is zero, the linear transform is "singular", which means that it loses some dimensions (the transformed volume is flat), and cannot be inverted.

The determinants are a fundamental tool in the resolution of systems of linear equations.

As you will later learn, a linear transformation can be decomposed in a pure rotation, a pure (anisotropic) scaling and another pure rotation. Only the scaling deforms the volumes, and the determinant of the transform is the product of the scaling coefficients.

  • 2
    Comments are not for extended discussion; this conversation has been [moved to chat](https://chat.stackexchange.com/rooms/77645/discussion-on-answer-by-yves-daoust-what-exactly-is-a-matrix). – Jyrki Lahtonen May 17 '18 at 19:26

1. Definition of a matrix.

The question of what a matrix is, precisely, is one I had for a long time as a high school student. It took many tries to get a straight answer, because people tend to conflate "matrix" with "linear transformation". The two are closely related, but NOT the same thing. So let me start with the fully rigorous definition of a matrix:

An $m$ by $n$ matrix is a function of two variables, the first of which has domain $\{1,2,\dots,m\}$ and the second of which has domain $\{1,2,\dots,n\}$.

This is the formal definition of matrices, but it's not how we usually think about them. We have a special notation for matrices--the "box of numbers" you are familiar with, where the value of the function at $(1,1)$ is put in the top left corner, the value at $(2,1)$ is put just below it, etc. We usually think of the matrix as just this box, and forget that it is a function. However, sometimes you need to remember that a matrix has a more formal definition, like when implementing matrices on a computer (most programming languages have matrices built into them).

2. What matrices represent.

Matrices can represent different things in different contexts, but there is one application that is most common. The most common application is linear transformations (a.k.a. linear maps), but before I get into that, let me briefly mention some other applications:

  • Matrices can be used to store data. For example, images on a computer are often stored as a matrix, where the matrix's value at $(i,j)$ is the intensity of light on the camera pixel that is $i^{th}$ from the top and $j^{th}$ from the left.
  • Matrices can be used as computational tools. For example, one way to compute the Fibonacci numbers is from powers of the matrix $$M = \begin{bmatrix} 1 & 1 \\ 1 & 0 \\ \end{bmatrix}$$ It turns out that $(M^k)_{11}$ is the $k^{th}$ Fibonacci number.
  • Matrices can be used to encode some mathematical structure. I'm going to be sort of hand-wavy about this, but an example of what I have in mind is an adjacency matrix for a graph or network, which tells you which nodes are connected to which.

So the point is that a matrix can be used for lots of things. However, one usage prevails as most common, and that is representing linear transformations. The prevalence of this usage is why people often conflate the two concepts. A linear transformation is a function $f$ of vectors which has the following properties:

  • $f(x+y) = f(x) + f(y)$ for any vectors $x$ and $y$.
  • $f(ax) = af(x)$ for any vector $x$ and any scalar $a$.

These properties are what it takes to ensure that the function $f$ has "no curvature". So it's like a straight line, but possibly in higher dimensions.

The relationship between matrices and linear transformations comes from the fact that a linear transformation is completely specified by the values it takes on a basis for its domain. (I presume you know what a basis is.) To see how this works, suppose we have a linear transformation $f$ which has domain $V$ and range $W$, where $V$ is a vector space with basis $v_1,v_2,\dots, v_n$ and $W$ is a vector space with basis $w_1,w_2,\dots,w_m$. Then there is a matrix $M$ representing $f$ with respect to these bases, which has as element $(i,j)$ the coefficient of $w_i$ when you express $f(v_j)$ as a sum of basis elements in $W$.

The reason that this is a good idea is that if you have some miscellaneous vector $x = a_1 v_1 + a_2 v_2 + \cdots + a_n v_n \in V$, then if you represent $x$ as a column vector $[a_1,a_2,\dots,a_n]^T$ and $f$ as its matrix $M$, then the value $f(x)$ is given by the matrix product of $M$ and $[a_1,a_2,\dots,a_n]^T$. So the matrix $M$ completely encodes the linear transformation $f$, and matrix multiplication tells you how to decode it, i.e. how to use the matrix to get values of $f$.

3. Geometrical intuition.

In my opinion, the most important theorem for getting intuition for matrices and linear transformations is the singular value decomposition theorem. This says that any linear transformation can be written as a sequence of three simple transformations: a rotation, a stretching, and another rotation. Note that the stretching operation can stretch by different amounts in different orthogonal directions. This tells you that all linear transformations are some combination of rotation and stretching.

Other properties of matrices often have direct geometric interpretation, too. For example, the determinant tells you how a linear transformation changes volumes. By the singular value decomposition, a linear transformation turns a cube into some sort of stretched and rotated parallelogram. The determinant is the ratio of the volume of the resulting parallelogram to that of the cube you started with.

Not all properties of a matrix can be easily associated with familiar geometric concepts, though. I don't know of a good geometric picture for the trace, for instance. That doesn't mean that the trace is any less useful or easy to work with, though!

4. Other properties.

Almost all of the "properties" and "operations" for matrices come from properties of linear maps and theorems about them. For example, the standard multiplication of matrices is designed specifically to give the values of linear maps as explained above. This is NOT the only type of multiplication that can be defined on matrices, and in fact there are other types of multiplication for matrices (for example, the Hadamard product and the Kronecker product). These other types of multiplication are sometimes useful, but generally not as useful as regular matrix multiplication, so people often don't know (or care) about them.

5. TL;DR

The moral of the story is that you can use matrices for whatever you want (and they are indeed used in many different ways), but the way that most people use them most of the time is to represent linear maps, and the standard definitions and "properties" of matrices reflect this bias. The study of linear maps goes by the name "linear algebra", and a textbook on this subject is a good place to start if you want to learn more about matrices. (Depending on your background, you may find some good reference suggestions here: link.)

  • 14,399
  • 4
  • 29
  • 70
  • 2
    A similar way to look at the essential characteristic of what a matrix itself is, is to define it as an ordered $n\times m$-tuple, that is usually written out in a grid format rather than in a list format, like a vector. – Todd Wilcox May 16 '18 at 14:18
  • Good answer, but I would put the summary at the beginning — especially if you call it _Too Long; Did not Read_! – PJTraill May 17 '18 at 08:59
  • Although it's techincally correct that matrices and maps are not the same thing, for the (frequent) case of $f: \mathbb K^n \rightarrow \mathbb K^m$ with respect to the standard basis, we have a one-to-one connection between those concepts (even more: we have an isomorphism between categories), and so it's useful to just identify both. The same thing is done e.g. in differential geometry, where we have three definitions for the tangent bundle (equivalent curves, vectors in each chart, and abstract derivations) and don't explicitly write an isomorphism each time. – Lukas Juhrich May 20 '18 at 14:06

Too long for a comment ... Just to start:

One essential way to understand matrices is to consider them as a collection of column vectors.

  • Now the multiplication of a matrix with a vector is a linear combination of those column vectors, that is, an element of the span of the column vectors.

For example: $$\begin{pmatrix}1&2\\3&4\\5&6\end{pmatrix}\begin{pmatrix}-3\\4\end{pmatrix}=-3\begin{pmatrix}1\\3\\5\end{pmatrix}+4\begin{pmatrix} 2\\4\\6\end{pmatrix}. $$

From here some properties of matrices are understable:

—If the linear system $Ax=b$ is solvable, the vector $b$ is contained in the span of the column vectors of $A$, which is a geometric interpretation of the solvability.

—The solution set of $Ax=0$ is non-empty. Iff it has more than one solution, the zero vector is a nontrivial linear combination of the column vector, ie, they are linearly dependent.

— Convince yourself that the rank of a matrix is the dimension of the span of its column vectors. From here it’s clear why a linear system is solvable iff the rank of the coefficient matrix equal the rank of the augmented matrix and so on.

For me — some decades ago — it was very helpful to connect properties of matrices with linear systems.

Michael Hoppe
  • 16,684
  • 3
  • 28
  • 47
  • 1
    Why necessarily _column_ vectors? If you multiply a matrix by a vector on the left, the matrix will behave as a collection of _row_ vectors. – Ruslan May 15 '18 at 21:45
  • @Ruslan *One* essential way. Michael Artin's *Algebra* makes good use of hypervectors for talking about bases and coordinates; from this point of view a matrix is a hypervector of $\mathbb{R}^n$ vectors, where you erase the inner parentheses because they are redundant. – Kyle Miller May 15 '18 at 21:57
  1. Matrix is an infinite or finite collection of some entities arranged in rows and columns. The entities are typically composed of numbers, symbols or expressions. The determinant is one of the basic operation that can be performed on a square matrix. It is, therefore, a more specialized concept.

  2. Matrices can be thought of as a generalization of the number concept. Therefore, they can represent many things:

    Group operations 
    Symmetry transformations 
    Complex numbers, quaternions 
    First derivatives of multivariate functions (the Jacobian matrix) 
    Second derivatives of multivariate functions (the Hessian matrix)
  3. Some matrices have geometric interpretation, for instance:

        describing rotations, linear transforms, etc.
  4. Matrices are used whenever one needs to represent in a short form an infinite or finite collection of some entities arranged in rows and columns.

  5. Properties of a matrix are first of all the properties of its entities (for instance $M(n, R)$ are square $n$-by-$n$ matrices over the $R$-ring). Second, properties can refer to the structure of a matrix itself. For instance it can be symmetric, orthogonal, etc.

  • 777
  • 5
  • 16