Is a linear tranformation onto or one-to-one?
The definition of one-to-one was pretty straight forward. If the vectors are lin.indep the transformation would be one-to-one. But could it also be onto?
The definition of onto was a little more abstract. Would a zero-row in reduced echelon form have any effect on this? I just assumed that because it has a couple of free variables it would be onto, but that zero-row set me off a bit. Say if the matrix is 4 by 5. With two free variables. And the zero-row in echelon form.
(Sorry for not posting the given matrix, but that is to specific. And I don't want to get a ban from uni for asking online. The task is determine the onto/one-to-one of to matrices)
I'll check back after class and update the question if more information is desirable.
Update
The definitions in the book is this;
Onto:
$T: mathbb R^n to mathbb R^m $ is said to be onto $mathbb R^m $ if each b in
$mathbb R^m $ is the image of at least one x in $mathbb R^n $
One-to-one:
$T: mathbb R^n to mathbb R^m $ is said to be one-to-one $mathbb R^m $ if each b in
$mathbb R^m $ is the image of at most one x in $mathbb R^n $
And then, there is another theorem that states that a linear transformation is one-to-one iff the equation T(x) = 0 has only the trivial solution.
That doesn't say anything about onto.
Then there is this bit that confused be about onto:
Let $T: mathbb R^n to mathbb R^m $ be a linear transformation and let A be the standard matrix for T. Then:
T maps $T: mathbb R^n$ onto $mathbb R^m $ iff the columns of A span $mathbb R^m $.
linear-algebra
add a comment |
The definition of one-to-one was pretty straight forward. If the vectors are lin.indep the transformation would be one-to-one. But could it also be onto?
The definition of onto was a little more abstract. Would a zero-row in reduced echelon form have any effect on this? I just assumed that because it has a couple of free variables it would be onto, but that zero-row set me off a bit. Say if the matrix is 4 by 5. With two free variables. And the zero-row in echelon form.
(Sorry for not posting the given matrix, but that is to specific. And I don't want to get a ban from uni for asking online. The task is determine the onto/one-to-one of to matrices)
I'll check back after class and update the question if more information is desirable.
Update
The definitions in the book is this;
Onto:
$T: mathbb R^n to mathbb R^m $ is said to be onto $mathbb R^m $ if each b in
$mathbb R^m $ is the image of at least one x in $mathbb R^n $
One-to-one:
$T: mathbb R^n to mathbb R^m $ is said to be one-to-one $mathbb R^m $ if each b in
$mathbb R^m $ is the image of at most one x in $mathbb R^n $
And then, there is another theorem that states that a linear transformation is one-to-one iff the equation T(x) = 0 has only the trivial solution.
That doesn't say anything about onto.
Then there is this bit that confused be about onto:
Let $T: mathbb R^n to mathbb R^m $ be a linear transformation and let A be the standard matrix for T. Then:
T maps $T: mathbb R^n$ onto $mathbb R^m $ iff the columns of A span $mathbb R^m $.
linear-algebra
What is your definition of an onto matrix? Full (row) rank?
– wildildildlife
Mar 11 '11 at 11:00
I'll update when I get home :)
– Algific
Mar 11 '11 at 11:19
add a comment |
The definition of one-to-one was pretty straight forward. If the vectors are lin.indep the transformation would be one-to-one. But could it also be onto?
The definition of onto was a little more abstract. Would a zero-row in reduced echelon form have any effect on this? I just assumed that because it has a couple of free variables it would be onto, but that zero-row set me off a bit. Say if the matrix is 4 by 5. With two free variables. And the zero-row in echelon form.
(Sorry for not posting the given matrix, but that is to specific. And I don't want to get a ban from uni for asking online. The task is determine the onto/one-to-one of to matrices)
I'll check back after class and update the question if more information is desirable.
Update
The definitions in the book is this;
Onto:
$T: mathbb R^n to mathbb R^m $ is said to be onto $mathbb R^m $ if each b in
$mathbb R^m $ is the image of at least one x in $mathbb R^n $
One-to-one:
$T: mathbb R^n to mathbb R^m $ is said to be one-to-one $mathbb R^m $ if each b in
$mathbb R^m $ is the image of at most one x in $mathbb R^n $
And then, there is another theorem that states that a linear transformation is one-to-one iff the equation T(x) = 0 has only the trivial solution.
That doesn't say anything about onto.
Then there is this bit that confused be about onto:
Let $T: mathbb R^n to mathbb R^m $ be a linear transformation and let A be the standard matrix for T. Then:
T maps $T: mathbb R^n$ onto $mathbb R^m $ iff the columns of A span $mathbb R^m $.
linear-algebra
The definition of one-to-one was pretty straight forward. If the vectors are lin.indep the transformation would be one-to-one. But could it also be onto?
The definition of onto was a little more abstract. Would a zero-row in reduced echelon form have any effect on this? I just assumed that because it has a couple of free variables it would be onto, but that zero-row set me off a bit. Say if the matrix is 4 by 5. With two free variables. And the zero-row in echelon form.
(Sorry for not posting the given matrix, but that is to specific. And I don't want to get a ban from uni for asking online. The task is determine the onto/one-to-one of to matrices)
I'll check back after class and update the question if more information is desirable.
Update
The definitions in the book is this;
Onto:
$T: mathbb R^n to mathbb R^m $ is said to be onto $mathbb R^m $ if each b in
$mathbb R^m $ is the image of at least one x in $mathbb R^n $
One-to-one:
$T: mathbb R^n to mathbb R^m $ is said to be one-to-one $mathbb R^m $ if each b in
$mathbb R^m $ is the image of at most one x in $mathbb R^n $
And then, there is another theorem that states that a linear transformation is one-to-one iff the equation T(x) = 0 has only the trivial solution.
That doesn't say anything about onto.
Then there is this bit that confused be about onto:
Let $T: mathbb R^n to mathbb R^m $ be a linear transformation and let A be the standard matrix for T. Then:
T maps $T: mathbb R^n$ onto $mathbb R^m $ iff the columns of A span $mathbb R^m $.
linear-algebra
linear-algebra
edited Mar 11 '11 at 16:19
Arturo Magidin
261k32584904
261k32584904
asked Mar 11 '11 at 10:52
Algific
89931325
89931325
What is your definition of an onto matrix? Full (row) rank?
– wildildildlife
Mar 11 '11 at 11:00
I'll update when I get home :)
– Algific
Mar 11 '11 at 11:19
add a comment |
What is your definition of an onto matrix? Full (row) rank?
– wildildildlife
Mar 11 '11 at 11:00
I'll update when I get home :)
– Algific
Mar 11 '11 at 11:19
What is your definition of an onto matrix? Full (row) rank?
– wildildildlife
Mar 11 '11 at 11:00
What is your definition of an onto matrix? Full (row) rank?
– wildildildlife
Mar 11 '11 at 11:00
I'll update when I get home :)
– Algific
Mar 11 '11 at 11:19
I'll update when I get home :)
– Algific
Mar 11 '11 at 11:19
add a comment |
3 Answers
3
active
oldest
votes
"One-to-one" and "onto" are properties of functions in general, not just linear transformations.
Definition. Let $fcolon Xto Y$ be a function.
$f$ is one-to-one if and only if for every $yin Y$ there is at most one $xin X$ such that $f(x)=y$; equivalently, if and only if $f(x_1)=f(x_2)$ implies $x_1=x_2$.
$f$ is onto (or onto $Y$, if the codomain is not clear from context) if and only if for every $yin Y$ there at least one $xin X$ such that $f(x)=y$.
This definition applies to linear transformations as well, and in particular for linear transformations $Tcolon mathbb{R}^ntomathbb{R}^m$, and by extension to matrices, since an $mtimes n$ matrix $A$ can be identified with the linear transformation $L_Acolonmathbb{R}^ntomathbb{R}^m$ given by $L_A(mathbf{x}) = Amathbf{x}$.
So, the definitions are for any functions. But when our sets $X$ and $Y$ have more structure to them, and the functions are not arbitrary, but special kinds of functions, we can often obtain other ways of characterizing a function as one-to-one or onto which is easier/better/more useful/more conceptual/has interesting applications. This is indeed the case when we have such a rich structure as linear transformations and vector spaces.
One-to-one is probably the easiest; this is because whether a function is one-to-one depends only on its domain, and not on its codomain. By contrast, whether a function is onto depends on both on the domain and the codomain (so, for instance, $f(x)=x^2$ is onto if we think of it as a function $fcolonmathbb{R}to[0,infty)$, but not if we think of it as a function $fcolonmathbb{R}tomathbb{R}$, or $fcolon[2,infty)to[0,infty)$).
Theorem. Let $Tcolonmathbb{R}^ntomathbb{R}^m$ be a linear transformation. The following are equivalent:
$T$ is one-to-one.
$T(mathbf{x})=mathbf{0}$ has only the trivial solution $mathbf{x}=mathbf{0}$.
If $A$ is the standard matrix of $T$, then the columns of $A$ are linearly independent.
Proof. The equivalence of (1) and (2) is basic in linear algebra, so let's deal with that:
(1)$Rightarrow$(2): If $T$ is one-to-one, then for all $mathbf{x}$, since $T(mathbf{0})=mathbf{0}$ (being linear), then $T(mathbf{x})=mathbf{0}=T(mathbf{0})$ implies $mathbf{x}=mathbf{0}$; this proves (2).
(2)$Rightarrow$(1): Suppose $T(mathbf{x}_1)=T(mathbf{x}_2)$. Then
$$mathbf{0} = T(mathbf{x}_1) - T(mathbf{x}_2) = T(mathbf{x}_1-mathbf{x}_2),$$
since $T$ is linear; because we are assuming (2), $T(mathbf{x}_1-mathbf{x}_2)=mathbf{0}$ implies that $mathbf{x}_1-mathbf{x}_2 = mathbf{0}$, so $mathbf{x}_1=mathbf{x}_2$, which proves that $T$ is indeed one-to-one.
The key to the connection with (3) (and eventually to your confusion) is that multiplying a matrix by a vector can be seen as an operation on columns. If
$$A=left(begin{array}{cccc}
a_{11} & a_{12} & ldots & a_{1n}\
a_{21} & a_{22} & ldots & a_{2n}\
vdots & vdots & ddots & vdots\
a_{m1} & a_{m2} & ldots & a_{mn}
end{array}right),$$
then let the columns of $A$, $A_1$, $A_2,ldots,A_n$ be:
$$A_1 = left(begin{array}{c}a_{11}\a_{21}\ vdots\a_{m1}end{array}right),quad A_2 = left(begin{array}{c}a_{12}\a_{22}\ vdots\ a_{m2}end{array}right),quadldots,A_n = left(begin{array}{c}a_{1n}\a_{2n}\ vdots\ a_{mn}end{array}right).$$
Then we have the following:
$$Aleft(begin{array}{c}x_1\x_2\ vdots\x_nend{array}right) = x_1 A_1 + a_2 A_2 + cdots + a_nA_n.$$
That is, multiplying $A$ by $mathbf{x}$ gives a linear combination of the columns of $A$. This gives the direct connection we need between conditions (1) and (2), and condition (3).
(2)$Rightarrow$(3): To show that the columns of $A$ are linearly independent, we need to show that if $alpha_1A_1 + cdots + alpha_nA_n = mathbf{0}$, then $alpha_1=cdots=alpha_n=0$. So suppose $alpha_1A_1+cdots+alpha_nA_n = mathbf{0}$. Then
$$T(mathbf{alpha}) = Amathbf{alpha} = alpha_1A_1+cdots+alpha_nA_n = mathbf{0},qquadtext{where }mathbf{alpha}=left(begin{array}{c}alpha_1\ alpha_2\ vdots\ alpha_nend{array}right).$$
Because we are assuming (2), that means that from $T(mathbf{alpha}) = mathbf{0}$ we can conclude that $alpha=mathbf{0}$; therefore, $alpha_1=cdots=alpha_n=0$. This proves that $A_1,ldots,A_n$ are linearly independent.
(3)$Rightarrow$(2): Suppose the columns of $A$ are linearly independent, and
$$mathbf{0} = T(mathbf{x}) = Amathbf{x}quadtext{where }mathbf{x}=left(begin{array}{c}a_1\a_2\ vdots\ a_nend{array}right).$$
This means that $a_1A_1 + cdots a_nA_n = mathbf{0}$; since the columns of $A$ are assumed to be linearly independent, we conclude that $a_1=cdots=a_n=0$, so $mathbf{x}=mathbf{0}$, proving (2). QED
What about onto? There are two things here. One is a theorem similar to the one above; the other is the Rank-Nullity Theorem.
Theorem. Let $Tcolonmathbb{R}^ntomathbb{R}^m$ be a linear transformation. The following are equivalent:
$T$ is onto.
The equation $T(mathbf{x})=mathbf{b}$ has solutions for every $mathbf{b}inmathbb{R}^m$.
If $A$ is the standard matrix of $T$, then the columns of $A$ span $mathbb{R}^m$. That is: every $mathbf{b}inmathbb{R}^m$ is a linear combination of the columns of $A$.
Proof. (1)$Leftrightarrow$(2) is essentially the definition, only cast in terms of equations for the sake of similarity to the previous theorem.
(2)$Rightarrow$(3) Let $mathbf{b}inmathbb{R}^m$. Then by (2) there exists an $mathbf{a}inmathbb{R}^n$ such that $T(mathbf{a})=mathbf{b}$. We have:
$$mathbf{b} = T(mathbf{a}) = Amathbf{a} = Aleft(begin{array}{c}a_1\a_2\ vdots\a_nend{array}right) = a_1A_1 + a_2A_2 + cdots + a_nA_n.$$
That is, we can express $mathbf{b}$ as a linear combination of the columns of $A$. Since $mathbf{b}$ is arbitrary, every vector in $mathbb{R}^m$ can be expressed as a linear combination of the columns of $A$, so the columns of $A$ span $mathbb{R}^m$; this proves (3).
(3)$Rightarrow$(2) Suppose the columns of $A$ span $mathbb{R}^m$ and let $mathbf{b}inmathbb{R}^m$. We want to show that $T(mathbf{x})=mathbf{b}$ has at least one solution.
Since the columns of $A$ span $mathbb{R}^m$, there exist scalars $alpha_1,ldots,alpha_n$ such that
$$mathbf{b} = alpha_1 A_1 + cdots + alpha_n A_n = Aleft(begin{array}{c}alpha_1\ alpha_2\ vdots\ alpha_nend{array}right) = T(mathbf{alpha}).$$
So $mathbf{alpha}$, where
$$mathbf{alpha} = left(begin{array}{c}alpha_1\ alpha_2\ vdots\ alpha_nend{array}right),$$
is a solution to $T(mathbf{x})=mathbf{b}$. This establishes (2). QED
So: "one-to-one"-ness is related to linear independence; "onto"-ness is related to spanning properties. Note that linear independence is an intrinsic property (it depends only on the set of vectors), whereas spanning is an extrinsic property (it depends also on the space we are considering; it is contextual). This matches the fact that whether a function is one-to-one or not depends only on the domain, but whether it is onto depends on both the domain and the codomain of the function.
But there is a deep connection between the two. Remember the following:
Definition. Let $A$ be an $mtimes n$ matrix. The nullity of $A$, $mathrm{nullity}(A)$, is the dimension of the kernel of $A$, that is, of the subspace of $mathbb{R}^n$ given by
$$mathrm{ker}(A) = Bigl{ mathbf{x}inmathbb{R}^nBigm| Amathbf{x}=mathbf{0}Bigr}.$$
The rank of $A$, $mathrm{rank}(A)$ is the dimension of the image of $A$; that is, of the subspace of $mathbb{R}^m$ given by
begin{align*}
mathrm{Im}(A) &= Bigl{ mathbf{b}inmathbb{R}^mBigm| Amathbf{x}=mathbf{b}text{ has at least one solution}Bigr}\
&= Bigl{ A(mathbf{x})Bigm|mathbf{x}inmathbb{R}^nBigr}.
end{align*}
The deep connection between them is given by the Rank-Nullity Theorem:
Rank-Nullity Theorem. Let $A$ be an $mtimes n$ matrix. Then
$$mathrm{rank}(A) + mathrm{nullity}(A) = n.$$
Now we get two more equivalences for one-to-one and onto:
Theorem. Let $Tcolonmathbb{R}^ntomathbb{R^m}$ be a linear transformation. The following are equivalent:
$T$ is one-to-one.
The equation $T(mathbf{x})=mathbf{0}$ has only the trivial solution $mathbf{x}=mathbf{0}$.
If $A$ is the standard matrix of $T$, then the columns of $A$ are linearly independent.
$mathrm{ker}(A) = {mathbf{0}}$.
$mathrm{nullity}(A) = 0$.
$mathrm{rank}(A) = n$.
Proof. The equivalence of (4) and (5) follows because only the trivial subspace has dimension $0$; the equivalence of (4) and (2) follows by definition of the kernel. The equivalence of (5) and (6) follows from the Rank-Nullity Theorem, since $n = mathrm{nullity}(A)+mathrm{rank}(A)$, so $mathrm{nullity}(A) = 0$ if and only if $mathrm{rank}(A) = n$. Since we already know (1), (2), and (3) are equivalent, the result follows. QED
Theorem. Let $Tcolonmathbb{R}^ntomathbb{R}^m$ be a linear transformation. The following are equivalent:
$T$ is onto.
The equation $T(mathbf{x})=mathbf{b}$ has solutions for every $mathbf{b}inmathbb{R}^m$.
If $A$ is the standard matrix of $T$, then the columns of $A$ span $mathbb{R}^m$. That is: every $mathbf{b}inmathbb{R}^m$ is a linear combination of the columns of $A$.
$mathrm{Im}(A) = mathbb{R}^m$.
$mathrm{rank}(A) = m$.
$mathrm{nullity}(A) = n-m$.
Proof. We already know that (1), (2), and (3) are equivalent. The equivalence of (4) and (2) follows by the definition of the image. The equivalence of (4) and (5) follows because the only subspace of $mathbb{R}^m$ that has dimension $m$ is the whole space. Finally, the equivalence of (5) and (6) follows from the rank nullity theorem: since $n = mathrm{rank}(A)+mathrm{nullity}(A)$, then $mathrm{nullity}(A) = n - mathrm{rank}(A)$. So the rank equals $m$ if and only if the nullity equals $n-m$. QED
So now you have a whole bunch of ways of checking if a matrix is one-to-one, and of checking if a matrix is onto. None of them is "better" than the others: for some matrices, one will be easier to check, for other matrices, it may be a different one which is easy to check. Also, the rank of a matrix is closely related to its row-echelon form, so that might help as well.
Note a few things: generally, "onto" and "one-to-one" are independent of one another. You can have a matrix be onto but not one-to-one; or be one-to-one but not onto; or be both; or be neither. The Rank-Nullity Theorem does place some restrictions: if $A$ is $mtimes n$ and $mlt n$, then the matrix cannot be onto (because $1leqmathrm{rank}(A)leq m$, so if $mathrm{rank}(A)+mathrm{nullity}(A) = n$, we must have $mathrm{nullity}(A)gt 0$); dually, if $mgt n$ then $A$ cannot be onto. In particular, the only matrices that can be both one-to-one and onto are square matrices. On the other hand, you can have an $mtimes n$ matrix with $mlt n$ that is onto, or one that is not onto. And you can have $mtimes n$ matrices with $mgt n$ that are one-to-one, and matrices that are not one-to-one.
Sorry if this is a stupid question, but is it always the case that whether a function is one-to-one depends only on the domain? For instance with the example $f(x) = x^2$ if the codomain were $0$, wouldn't the function still be one-to-one?
– ghshtalt
Mar 11 '11 at 18:13
1
@ghshtalt: For $f(x)=x^2$ to have codomain $0$, you would need the domain to be just $0$ (otherwise, you wouldn't have a function). But of course, if the domain of a function is a single point, then the function is one-to-one, regardless of what the codomain (or what the function) is. So in fact, it's because the domain is a single point that such a function would be one-to-one.
– Arturo Magidin
Mar 11 '11 at 18:19
oh ok, so it wouldn't have been a function in the first place... Thanks for clearing that up for me! And thanks for another great answer of course!
– ghshtalt
Mar 11 '11 at 18:22
I believe I have an example that is not one-to-one nor onto. The matrix is lin.dep (free variables), and for a random value in the codomain the reduced augmented matrix is inconsistent. Hence there is not one value in the domain that corresponds to that value in the codomain. Right, right?
– Algific
Mar 13 '11 at 17:26
3
@Algific: Matrices by themselves are nor "linearly independent" or "linearly dependent". Sets of vectors are linearly independent or linearly dependent. If you mean that you have a matrix whose columns are linearly dependent (and somehow relating that to "free variables", yet another concept that is not directly applicable to matrices, but rather to other things to which you can associate matrices), then it is indeed true that any matrix whose columns are linearly dependent gives a linear transformation that is not one-to-one. If not all systems Ax=b have solutions, then not onto.
– Arturo Magidin
Mar 13 '11 at 20:28
|
show 3 more comments
let $T(x)=Ax$ be a linear transformation.
$T(x)$ is one-to-one if the columns of $A$ are linearly independent.
$T(x)$ is onto if every row of $A$ has a pivot.
Is this truly how easy it is? I know the first one, one-to-one if linearly independent, but didn't know the second one. So does it matter if the system has free variables? As long as every row has a pivot it's onto?
– joe_04_04
Mar 23 '17 at 9:19
This still didn't answer the parent question of "Would a zero-row in reduced echelon form have any effect on (ONTO)? " That is to ask "T(x) is onto if every non-zero row of A has a pivot."?
– Ron Jensen
Sep 26 '17 at 20:56
add a comment |
A function is onto if its codomain is equal to its image, that is, $f: X to Y$ is onto if $f(X) = Y$.
Assuming you are working with real matrices, a matrix $A$ is usually seen as mapping vectors from $mathbb R^n$ to $mathbb R^m$. You want to show that for all $vec y in mathbb R^m$ there exists a $vec x in mathbb R^n$ such that $A vec x = vec y$. In some cases this is not possible depending on $m$ and $n$. How does $m$ and $n$ relate to the size of the matrix?
One way to check that a matrix is onto is to check if you can find vectors $vec v_1, vec v_2, dots, vec v_m$ such that $Avec v_i = vec e_i$ for $i = 1, 2, dots, m$, where $vec e_1, vec e_2, dots, vec e_m$ is the standard basis in $mathbb R^m$ (why?). Can you use the echelon form for this? Do you actually need to get the standard basis or is some other set of vectors ok?
The question is sort of an open one, it says discuss the concepts of one-to-one and onto with regard to the two following matrices.
– Algific
Mar 11 '11 at 12:11
Well, to start of, check if they are one-to-one and onto. Does a matrix have to be one-to-one to be onto? Does it have to be onto to be one-to-one?
– Calle
Mar 11 '11 at 12:17
The problem is that the theorems doesn't mention the other state. Funny enough.
– Algific
Mar 11 '11 at 16:06
add a comment |
Your Answer
StackExchange.ifUsing("editor", function () {
return StackExchange.using("mathjaxEditing", function () {
StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
});
});
}, "mathjax-editing");
StackExchange.ready(function() {
var channelOptions = {
tags: "".split(" "),
id: "69"
};
initTagRenderer("".split(" "), "".split(" "), channelOptions);
StackExchange.using("externalEditor", function() {
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled) {
StackExchange.using("snippets", function() {
createEditor();
});
}
else {
createEditor();
}
});
function createEditor() {
StackExchange.prepareEditor({
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: true,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
imageUploader: {
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
},
noCode: true, onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
});
}
});
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f26371%2fis-a-linear-tranformation-onto-or-one-to-one%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
3 Answers
3
active
oldest
votes
3 Answers
3
active
oldest
votes
active
oldest
votes
active
oldest
votes
"One-to-one" and "onto" are properties of functions in general, not just linear transformations.
Definition. Let $fcolon Xto Y$ be a function.
$f$ is one-to-one if and only if for every $yin Y$ there is at most one $xin X$ such that $f(x)=y$; equivalently, if and only if $f(x_1)=f(x_2)$ implies $x_1=x_2$.
$f$ is onto (or onto $Y$, if the codomain is not clear from context) if and only if for every $yin Y$ there at least one $xin X$ such that $f(x)=y$.
This definition applies to linear transformations as well, and in particular for linear transformations $Tcolon mathbb{R}^ntomathbb{R}^m$, and by extension to matrices, since an $mtimes n$ matrix $A$ can be identified with the linear transformation $L_Acolonmathbb{R}^ntomathbb{R}^m$ given by $L_A(mathbf{x}) = Amathbf{x}$.
So, the definitions are for any functions. But when our sets $X$ and $Y$ have more structure to them, and the functions are not arbitrary, but special kinds of functions, we can often obtain other ways of characterizing a function as one-to-one or onto which is easier/better/more useful/more conceptual/has interesting applications. This is indeed the case when we have such a rich structure as linear transformations and vector spaces.
One-to-one is probably the easiest; this is because whether a function is one-to-one depends only on its domain, and not on its codomain. By contrast, whether a function is onto depends on both on the domain and the codomain (so, for instance, $f(x)=x^2$ is onto if we think of it as a function $fcolonmathbb{R}to[0,infty)$, but not if we think of it as a function $fcolonmathbb{R}tomathbb{R}$, or $fcolon[2,infty)to[0,infty)$).
Theorem. Let $Tcolonmathbb{R}^ntomathbb{R}^m$ be a linear transformation. The following are equivalent:
$T$ is one-to-one.
$T(mathbf{x})=mathbf{0}$ has only the trivial solution $mathbf{x}=mathbf{0}$.
If $A$ is the standard matrix of $T$, then the columns of $A$ are linearly independent.
Proof. The equivalence of (1) and (2) is basic in linear algebra, so let's deal with that:
(1)$Rightarrow$(2): If $T$ is one-to-one, then for all $mathbf{x}$, since $T(mathbf{0})=mathbf{0}$ (being linear), then $T(mathbf{x})=mathbf{0}=T(mathbf{0})$ implies $mathbf{x}=mathbf{0}$; this proves (2).
(2)$Rightarrow$(1): Suppose $T(mathbf{x}_1)=T(mathbf{x}_2)$. Then
$$mathbf{0} = T(mathbf{x}_1) - T(mathbf{x}_2) = T(mathbf{x}_1-mathbf{x}_2),$$
since $T$ is linear; because we are assuming (2), $T(mathbf{x}_1-mathbf{x}_2)=mathbf{0}$ implies that $mathbf{x}_1-mathbf{x}_2 = mathbf{0}$, so $mathbf{x}_1=mathbf{x}_2$, which proves that $T$ is indeed one-to-one.
The key to the connection with (3) (and eventually to your confusion) is that multiplying a matrix by a vector can be seen as an operation on columns. If
$$A=left(begin{array}{cccc}
a_{11} & a_{12} & ldots & a_{1n}\
a_{21} & a_{22} & ldots & a_{2n}\
vdots & vdots & ddots & vdots\
a_{m1} & a_{m2} & ldots & a_{mn}
end{array}right),$$
then let the columns of $A$, $A_1$, $A_2,ldots,A_n$ be:
$$A_1 = left(begin{array}{c}a_{11}\a_{21}\ vdots\a_{m1}end{array}right),quad A_2 = left(begin{array}{c}a_{12}\a_{22}\ vdots\ a_{m2}end{array}right),quadldots,A_n = left(begin{array}{c}a_{1n}\a_{2n}\ vdots\ a_{mn}end{array}right).$$
Then we have the following:
$$Aleft(begin{array}{c}x_1\x_2\ vdots\x_nend{array}right) = x_1 A_1 + a_2 A_2 + cdots + a_nA_n.$$
That is, multiplying $A$ by $mathbf{x}$ gives a linear combination of the columns of $A$. This gives the direct connection we need between conditions (1) and (2), and condition (3).
(2)$Rightarrow$(3): To show that the columns of $A$ are linearly independent, we need to show that if $alpha_1A_1 + cdots + alpha_nA_n = mathbf{0}$, then $alpha_1=cdots=alpha_n=0$. So suppose $alpha_1A_1+cdots+alpha_nA_n = mathbf{0}$. Then
$$T(mathbf{alpha}) = Amathbf{alpha} = alpha_1A_1+cdots+alpha_nA_n = mathbf{0},qquadtext{where }mathbf{alpha}=left(begin{array}{c}alpha_1\ alpha_2\ vdots\ alpha_nend{array}right).$$
Because we are assuming (2), that means that from $T(mathbf{alpha}) = mathbf{0}$ we can conclude that $alpha=mathbf{0}$; therefore, $alpha_1=cdots=alpha_n=0$. This proves that $A_1,ldots,A_n$ are linearly independent.
(3)$Rightarrow$(2): Suppose the columns of $A$ are linearly independent, and
$$mathbf{0} = T(mathbf{x}) = Amathbf{x}quadtext{where }mathbf{x}=left(begin{array}{c}a_1\a_2\ vdots\ a_nend{array}right).$$
This means that $a_1A_1 + cdots a_nA_n = mathbf{0}$; since the columns of $A$ are assumed to be linearly independent, we conclude that $a_1=cdots=a_n=0$, so $mathbf{x}=mathbf{0}$, proving (2). QED
What about onto? There are two things here. One is a theorem similar to the one above; the other is the Rank-Nullity Theorem.
Theorem. Let $Tcolonmathbb{R}^ntomathbb{R}^m$ be a linear transformation. The following are equivalent:
$T$ is onto.
The equation $T(mathbf{x})=mathbf{b}$ has solutions for every $mathbf{b}inmathbb{R}^m$.
If $A$ is the standard matrix of $T$, then the columns of $A$ span $mathbb{R}^m$. That is: every $mathbf{b}inmathbb{R}^m$ is a linear combination of the columns of $A$.
Proof. (1)$Leftrightarrow$(2) is essentially the definition, only cast in terms of equations for the sake of similarity to the previous theorem.
(2)$Rightarrow$(3) Let $mathbf{b}inmathbb{R}^m$. Then by (2) there exists an $mathbf{a}inmathbb{R}^n$ such that $T(mathbf{a})=mathbf{b}$. We have:
$$mathbf{b} = T(mathbf{a}) = Amathbf{a} = Aleft(begin{array}{c}a_1\a_2\ vdots\a_nend{array}right) = a_1A_1 + a_2A_2 + cdots + a_nA_n.$$
That is, we can express $mathbf{b}$ as a linear combination of the columns of $A$. Since $mathbf{b}$ is arbitrary, every vector in $mathbb{R}^m$ can be expressed as a linear combination of the columns of $A$, so the columns of $A$ span $mathbb{R}^m$; this proves (3).
(3)$Rightarrow$(2) Suppose the columns of $A$ span $mathbb{R}^m$ and let $mathbf{b}inmathbb{R}^m$. We want to show that $T(mathbf{x})=mathbf{b}$ has at least one solution.
Since the columns of $A$ span $mathbb{R}^m$, there exist scalars $alpha_1,ldots,alpha_n$ such that
$$mathbf{b} = alpha_1 A_1 + cdots + alpha_n A_n = Aleft(begin{array}{c}alpha_1\ alpha_2\ vdots\ alpha_nend{array}right) = T(mathbf{alpha}).$$
So $mathbf{alpha}$, where
$$mathbf{alpha} = left(begin{array}{c}alpha_1\ alpha_2\ vdots\ alpha_nend{array}right),$$
is a solution to $T(mathbf{x})=mathbf{b}$. This establishes (2). QED
So: "one-to-one"-ness is related to linear independence; "onto"-ness is related to spanning properties. Note that linear independence is an intrinsic property (it depends only on the set of vectors), whereas spanning is an extrinsic property (it depends also on the space we are considering; it is contextual). This matches the fact that whether a function is one-to-one or not depends only on the domain, but whether it is onto depends on both the domain and the codomain of the function.
But there is a deep connection between the two. Remember the following:
Definition. Let $A$ be an $mtimes n$ matrix. The nullity of $A$, $mathrm{nullity}(A)$, is the dimension of the kernel of $A$, that is, of the subspace of $mathbb{R}^n$ given by
$$mathrm{ker}(A) = Bigl{ mathbf{x}inmathbb{R}^nBigm| Amathbf{x}=mathbf{0}Bigr}.$$
The rank of $A$, $mathrm{rank}(A)$ is the dimension of the image of $A$; that is, of the subspace of $mathbb{R}^m$ given by
begin{align*}
mathrm{Im}(A) &= Bigl{ mathbf{b}inmathbb{R}^mBigm| Amathbf{x}=mathbf{b}text{ has at least one solution}Bigr}\
&= Bigl{ A(mathbf{x})Bigm|mathbf{x}inmathbb{R}^nBigr}.
end{align*}
The deep connection between them is given by the Rank-Nullity Theorem:
Rank-Nullity Theorem. Let $A$ be an $mtimes n$ matrix. Then
$$mathrm{rank}(A) + mathrm{nullity}(A) = n.$$
Now we get two more equivalences for one-to-one and onto:
Theorem. Let $Tcolonmathbb{R}^ntomathbb{R^m}$ be a linear transformation. The following are equivalent:
$T$ is one-to-one.
The equation $T(mathbf{x})=mathbf{0}$ has only the trivial solution $mathbf{x}=mathbf{0}$.
If $A$ is the standard matrix of $T$, then the columns of $A$ are linearly independent.
$mathrm{ker}(A) = {mathbf{0}}$.
$mathrm{nullity}(A) = 0$.
$mathrm{rank}(A) = n$.
Proof. The equivalence of (4) and (5) follows because only the trivial subspace has dimension $0$; the equivalence of (4) and (2) follows by definition of the kernel. The equivalence of (5) and (6) follows from the Rank-Nullity Theorem, since $n = mathrm{nullity}(A)+mathrm{rank}(A)$, so $mathrm{nullity}(A) = 0$ if and only if $mathrm{rank}(A) = n$. Since we already know (1), (2), and (3) are equivalent, the result follows. QED
Theorem. Let $Tcolonmathbb{R}^ntomathbb{R}^m$ be a linear transformation. The following are equivalent:
$T$ is onto.
The equation $T(mathbf{x})=mathbf{b}$ has solutions for every $mathbf{b}inmathbb{R}^m$.
If $A$ is the standard matrix of $T$, then the columns of $A$ span $mathbb{R}^m$. That is: every $mathbf{b}inmathbb{R}^m$ is a linear combination of the columns of $A$.
$mathrm{Im}(A) = mathbb{R}^m$.
$mathrm{rank}(A) = m$.
$mathrm{nullity}(A) = n-m$.
Proof. We already know that (1), (2), and (3) are equivalent. The equivalence of (4) and (2) follows by the definition of the image. The equivalence of (4) and (5) follows because the only subspace of $mathbb{R}^m$ that has dimension $m$ is the whole space. Finally, the equivalence of (5) and (6) follows from the rank nullity theorem: since $n = mathrm{rank}(A)+mathrm{nullity}(A)$, then $mathrm{nullity}(A) = n - mathrm{rank}(A)$. So the rank equals $m$ if and only if the nullity equals $n-m$. QED
So now you have a whole bunch of ways of checking if a matrix is one-to-one, and of checking if a matrix is onto. None of them is "better" than the others: for some matrices, one will be easier to check, for other matrices, it may be a different one which is easy to check. Also, the rank of a matrix is closely related to its row-echelon form, so that might help as well.
Note a few things: generally, "onto" and "one-to-one" are independent of one another. You can have a matrix be onto but not one-to-one; or be one-to-one but not onto; or be both; or be neither. The Rank-Nullity Theorem does place some restrictions: if $A$ is $mtimes n$ and $mlt n$, then the matrix cannot be onto (because $1leqmathrm{rank}(A)leq m$, so if $mathrm{rank}(A)+mathrm{nullity}(A) = n$, we must have $mathrm{nullity}(A)gt 0$); dually, if $mgt n$ then $A$ cannot be onto. In particular, the only matrices that can be both one-to-one and onto are square matrices. On the other hand, you can have an $mtimes n$ matrix with $mlt n$ that is onto, or one that is not onto. And you can have $mtimes n$ matrices with $mgt n$ that are one-to-one, and matrices that are not one-to-one.
Sorry if this is a stupid question, but is it always the case that whether a function is one-to-one depends only on the domain? For instance with the example $f(x) = x^2$ if the codomain were $0$, wouldn't the function still be one-to-one?
– ghshtalt
Mar 11 '11 at 18:13
1
@ghshtalt: For $f(x)=x^2$ to have codomain $0$, you would need the domain to be just $0$ (otherwise, you wouldn't have a function). But of course, if the domain of a function is a single point, then the function is one-to-one, regardless of what the codomain (or what the function) is. So in fact, it's because the domain is a single point that such a function would be one-to-one.
– Arturo Magidin
Mar 11 '11 at 18:19
oh ok, so it wouldn't have been a function in the first place... Thanks for clearing that up for me! And thanks for another great answer of course!
– ghshtalt
Mar 11 '11 at 18:22
I believe I have an example that is not one-to-one nor onto. The matrix is lin.dep (free variables), and for a random value in the codomain the reduced augmented matrix is inconsistent. Hence there is not one value in the domain that corresponds to that value in the codomain. Right, right?
– Algific
Mar 13 '11 at 17:26
3
@Algific: Matrices by themselves are nor "linearly independent" or "linearly dependent". Sets of vectors are linearly independent or linearly dependent. If you mean that you have a matrix whose columns are linearly dependent (and somehow relating that to "free variables", yet another concept that is not directly applicable to matrices, but rather to other things to which you can associate matrices), then it is indeed true that any matrix whose columns are linearly dependent gives a linear transformation that is not one-to-one. If not all systems Ax=b have solutions, then not onto.
– Arturo Magidin
Mar 13 '11 at 20:28
|
show 3 more comments
"One-to-one" and "onto" are properties of functions in general, not just linear transformations.
Definition. Let $fcolon Xto Y$ be a function.
$f$ is one-to-one if and only if for every $yin Y$ there is at most one $xin X$ such that $f(x)=y$; equivalently, if and only if $f(x_1)=f(x_2)$ implies $x_1=x_2$.
$f$ is onto (or onto $Y$, if the codomain is not clear from context) if and only if for every $yin Y$ there at least one $xin X$ such that $f(x)=y$.
This definition applies to linear transformations as well, and in particular for linear transformations $Tcolon mathbb{R}^ntomathbb{R}^m$, and by extension to matrices, since an $mtimes n$ matrix $A$ can be identified with the linear transformation $L_Acolonmathbb{R}^ntomathbb{R}^m$ given by $L_A(mathbf{x}) = Amathbf{x}$.
So, the definitions are for any functions. But when our sets $X$ and $Y$ have more structure to them, and the functions are not arbitrary, but special kinds of functions, we can often obtain other ways of characterizing a function as one-to-one or onto which is easier/better/more useful/more conceptual/has interesting applications. This is indeed the case when we have such a rich structure as linear transformations and vector spaces.
One-to-one is probably the easiest; this is because whether a function is one-to-one depends only on its domain, and not on its codomain. By contrast, whether a function is onto depends on both on the domain and the codomain (so, for instance, $f(x)=x^2$ is onto if we think of it as a function $fcolonmathbb{R}to[0,infty)$, but not if we think of it as a function $fcolonmathbb{R}tomathbb{R}$, or $fcolon[2,infty)to[0,infty)$).
Theorem. Let $Tcolonmathbb{R}^ntomathbb{R}^m$ be a linear transformation. The following are equivalent:
$T$ is one-to-one.
$T(mathbf{x})=mathbf{0}$ has only the trivial solution $mathbf{x}=mathbf{0}$.
If $A$ is the standard matrix of $T$, then the columns of $A$ are linearly independent.
Proof. The equivalence of (1) and (2) is basic in linear algebra, so let's deal with that:
(1)$Rightarrow$(2): If $T$ is one-to-one, then for all $mathbf{x}$, since $T(mathbf{0})=mathbf{0}$ (being linear), then $T(mathbf{x})=mathbf{0}=T(mathbf{0})$ implies $mathbf{x}=mathbf{0}$; this proves (2).
(2)$Rightarrow$(1): Suppose $T(mathbf{x}_1)=T(mathbf{x}_2)$. Then
$$mathbf{0} = T(mathbf{x}_1) - T(mathbf{x}_2) = T(mathbf{x}_1-mathbf{x}_2),$$
since $T$ is linear; because we are assuming (2), $T(mathbf{x}_1-mathbf{x}_2)=mathbf{0}$ implies that $mathbf{x}_1-mathbf{x}_2 = mathbf{0}$, so $mathbf{x}_1=mathbf{x}_2$, which proves that $T$ is indeed one-to-one.
The key to the connection with (3) (and eventually to your confusion) is that multiplying a matrix by a vector can be seen as an operation on columns. If
$$A=left(begin{array}{cccc}
a_{11} & a_{12} & ldots & a_{1n}\
a_{21} & a_{22} & ldots & a_{2n}\
vdots & vdots & ddots & vdots\
a_{m1} & a_{m2} & ldots & a_{mn}
end{array}right),$$
then let the columns of $A$, $A_1$, $A_2,ldots,A_n$ be:
$$A_1 = left(begin{array}{c}a_{11}\a_{21}\ vdots\a_{m1}end{array}right),quad A_2 = left(begin{array}{c}a_{12}\a_{22}\ vdots\ a_{m2}end{array}right),quadldots,A_n = left(begin{array}{c}a_{1n}\a_{2n}\ vdots\ a_{mn}end{array}right).$$
Then we have the following:
$$Aleft(begin{array}{c}x_1\x_2\ vdots\x_nend{array}right) = x_1 A_1 + a_2 A_2 + cdots + a_nA_n.$$
That is, multiplying $A$ by $mathbf{x}$ gives a linear combination of the columns of $A$. This gives the direct connection we need between conditions (1) and (2), and condition (3).
(2)$Rightarrow$(3): To show that the columns of $A$ are linearly independent, we need to show that if $alpha_1A_1 + cdots + alpha_nA_n = mathbf{0}$, then $alpha_1=cdots=alpha_n=0$. So suppose $alpha_1A_1+cdots+alpha_nA_n = mathbf{0}$. Then
$$T(mathbf{alpha}) = Amathbf{alpha} = alpha_1A_1+cdots+alpha_nA_n = mathbf{0},qquadtext{where }mathbf{alpha}=left(begin{array}{c}alpha_1\ alpha_2\ vdots\ alpha_nend{array}right).$$
Because we are assuming (2), that means that from $T(mathbf{alpha}) = mathbf{0}$ we can conclude that $alpha=mathbf{0}$; therefore, $alpha_1=cdots=alpha_n=0$. This proves that $A_1,ldots,A_n$ are linearly independent.
(3)$Rightarrow$(2): Suppose the columns of $A$ are linearly independent, and
$$mathbf{0} = T(mathbf{x}) = Amathbf{x}quadtext{where }mathbf{x}=left(begin{array}{c}a_1\a_2\ vdots\ a_nend{array}right).$$
This means that $a_1A_1 + cdots a_nA_n = mathbf{0}$; since the columns of $A$ are assumed to be linearly independent, we conclude that $a_1=cdots=a_n=0$, so $mathbf{x}=mathbf{0}$, proving (2). QED
What about onto? There are two things here. One is a theorem similar to the one above; the other is the Rank-Nullity Theorem.
Theorem. Let $Tcolonmathbb{R}^ntomathbb{R}^m$ be a linear transformation. The following are equivalent:
$T$ is onto.
The equation $T(mathbf{x})=mathbf{b}$ has solutions for every $mathbf{b}inmathbb{R}^m$.
If $A$ is the standard matrix of $T$, then the columns of $A$ span $mathbb{R}^m$. That is: every $mathbf{b}inmathbb{R}^m$ is a linear combination of the columns of $A$.
Proof. (1)$Leftrightarrow$(2) is essentially the definition, only cast in terms of equations for the sake of similarity to the previous theorem.
(2)$Rightarrow$(3) Let $mathbf{b}inmathbb{R}^m$. Then by (2) there exists an $mathbf{a}inmathbb{R}^n$ such that $T(mathbf{a})=mathbf{b}$. We have:
$$mathbf{b} = T(mathbf{a}) = Amathbf{a} = Aleft(begin{array}{c}a_1\a_2\ vdots\a_nend{array}right) = a_1A_1 + a_2A_2 + cdots + a_nA_n.$$
That is, we can express $mathbf{b}$ as a linear combination of the columns of $A$. Since $mathbf{b}$ is arbitrary, every vector in $mathbb{R}^m$ can be expressed as a linear combination of the columns of $A$, so the columns of $A$ span $mathbb{R}^m$; this proves (3).
(3)$Rightarrow$(2) Suppose the columns of $A$ span $mathbb{R}^m$ and let $mathbf{b}inmathbb{R}^m$. We want to show that $T(mathbf{x})=mathbf{b}$ has at least one solution.
Since the columns of $A$ span $mathbb{R}^m$, there exist scalars $alpha_1,ldots,alpha_n$ such that
$$mathbf{b} = alpha_1 A_1 + cdots + alpha_n A_n = Aleft(begin{array}{c}alpha_1\ alpha_2\ vdots\ alpha_nend{array}right) = T(mathbf{alpha}).$$
So $mathbf{alpha}$, where
$$mathbf{alpha} = left(begin{array}{c}alpha_1\ alpha_2\ vdots\ alpha_nend{array}right),$$
is a solution to $T(mathbf{x})=mathbf{b}$. This establishes (2). QED
So: "one-to-one"-ness is related to linear independence; "onto"-ness is related to spanning properties. Note that linear independence is an intrinsic property (it depends only on the set of vectors), whereas spanning is an extrinsic property (it depends also on the space we are considering; it is contextual). This matches the fact that whether a function is one-to-one or not depends only on the domain, but whether it is onto depends on both the domain and the codomain of the function.
But there is a deep connection between the two. Remember the following:
Definition. Let $A$ be an $mtimes n$ matrix. The nullity of $A$, $mathrm{nullity}(A)$, is the dimension of the kernel of $A$, that is, of the subspace of $mathbb{R}^n$ given by
$$mathrm{ker}(A) = Bigl{ mathbf{x}inmathbb{R}^nBigm| Amathbf{x}=mathbf{0}Bigr}.$$
The rank of $A$, $mathrm{rank}(A)$ is the dimension of the image of $A$; that is, of the subspace of $mathbb{R}^m$ given by
begin{align*}
mathrm{Im}(A) &= Bigl{ mathbf{b}inmathbb{R}^mBigm| Amathbf{x}=mathbf{b}text{ has at least one solution}Bigr}\
&= Bigl{ A(mathbf{x})Bigm|mathbf{x}inmathbb{R}^nBigr}.
end{align*}
The deep connection between them is given by the Rank-Nullity Theorem:
Rank-Nullity Theorem. Let $A$ be an $mtimes n$ matrix. Then
$$mathrm{rank}(A) + mathrm{nullity}(A) = n.$$
Now we get two more equivalences for one-to-one and onto:
Theorem. Let $Tcolonmathbb{R}^ntomathbb{R^m}$ be a linear transformation. The following are equivalent:
$T$ is one-to-one.
The equation $T(mathbf{x})=mathbf{0}$ has only the trivial solution $mathbf{x}=mathbf{0}$.
If $A$ is the standard matrix of $T$, then the columns of $A$ are linearly independent.
$mathrm{ker}(A) = {mathbf{0}}$.
$mathrm{nullity}(A) = 0$.
$mathrm{rank}(A) = n$.
Proof. The equivalence of (4) and (5) follows because only the trivial subspace has dimension $0$; the equivalence of (4) and (2) follows by definition of the kernel. The equivalence of (5) and (6) follows from the Rank-Nullity Theorem, since $n = mathrm{nullity}(A)+mathrm{rank}(A)$, so $mathrm{nullity}(A) = 0$ if and only if $mathrm{rank}(A) = n$. Since we already know (1), (2), and (3) are equivalent, the result follows. QED
Theorem. Let $Tcolonmathbb{R}^ntomathbb{R}^m$ be a linear transformation. The following are equivalent:
$T$ is onto.
The equation $T(mathbf{x})=mathbf{b}$ has solutions for every $mathbf{b}inmathbb{R}^m$.
If $A$ is the standard matrix of $T$, then the columns of $A$ span $mathbb{R}^m$. That is: every $mathbf{b}inmathbb{R}^m$ is a linear combination of the columns of $A$.
$mathrm{Im}(A) = mathbb{R}^m$.
$mathrm{rank}(A) = m$.
$mathrm{nullity}(A) = n-m$.
Proof. We already know that (1), (2), and (3) are equivalent. The equivalence of (4) and (2) follows by the definition of the image. The equivalence of (4) and (5) follows because the only subspace of $mathbb{R}^m$ that has dimension $m$ is the whole space. Finally, the equivalence of (5) and (6) follows from the rank nullity theorem: since $n = mathrm{rank}(A)+mathrm{nullity}(A)$, then $mathrm{nullity}(A) = n - mathrm{rank}(A)$. So the rank equals $m$ if and only if the nullity equals $n-m$. QED
So now you have a whole bunch of ways of checking if a matrix is one-to-one, and of checking if a matrix is onto. None of them is "better" than the others: for some matrices, one will be easier to check, for other matrices, it may be a different one which is easy to check. Also, the rank of a matrix is closely related to its row-echelon form, so that might help as well.
Note a few things: generally, "onto" and "one-to-one" are independent of one another. You can have a matrix be onto but not one-to-one; or be one-to-one but not onto; or be both; or be neither. The Rank-Nullity Theorem does place some restrictions: if $A$ is $mtimes n$ and $mlt n$, then the matrix cannot be onto (because $1leqmathrm{rank}(A)leq m$, so if $mathrm{rank}(A)+mathrm{nullity}(A) = n$, we must have $mathrm{nullity}(A)gt 0$); dually, if $mgt n$ then $A$ cannot be onto. In particular, the only matrices that can be both one-to-one and onto are square matrices. On the other hand, you can have an $mtimes n$ matrix with $mlt n$ that is onto, or one that is not onto. And you can have $mtimes n$ matrices with $mgt n$ that are one-to-one, and matrices that are not one-to-one.
Sorry if this is a stupid question, but is it always the case that whether a function is one-to-one depends only on the domain? For instance with the example $f(x) = x^2$ if the codomain were $0$, wouldn't the function still be one-to-one?
– ghshtalt
Mar 11 '11 at 18:13
1
@ghshtalt: For $f(x)=x^2$ to have codomain $0$, you would need the domain to be just $0$ (otherwise, you wouldn't have a function). But of course, if the domain of a function is a single point, then the function is one-to-one, regardless of what the codomain (or what the function) is. So in fact, it's because the domain is a single point that such a function would be one-to-one.
– Arturo Magidin
Mar 11 '11 at 18:19
oh ok, so it wouldn't have been a function in the first place... Thanks for clearing that up for me! And thanks for another great answer of course!
– ghshtalt
Mar 11 '11 at 18:22
I believe I have an example that is not one-to-one nor onto. The matrix is lin.dep (free variables), and for a random value in the codomain the reduced augmented matrix is inconsistent. Hence there is not one value in the domain that corresponds to that value in the codomain. Right, right?
– Algific
Mar 13 '11 at 17:26
3
@Algific: Matrices by themselves are nor "linearly independent" or "linearly dependent". Sets of vectors are linearly independent or linearly dependent. If you mean that you have a matrix whose columns are linearly dependent (and somehow relating that to "free variables", yet another concept that is not directly applicable to matrices, but rather to other things to which you can associate matrices), then it is indeed true that any matrix whose columns are linearly dependent gives a linear transformation that is not one-to-one. If not all systems Ax=b have solutions, then not onto.
– Arturo Magidin
Mar 13 '11 at 20:28
|
show 3 more comments
"One-to-one" and "onto" are properties of functions in general, not just linear transformations.
Definition. Let $fcolon Xto Y$ be a function.
$f$ is one-to-one if and only if for every $yin Y$ there is at most one $xin X$ such that $f(x)=y$; equivalently, if and only if $f(x_1)=f(x_2)$ implies $x_1=x_2$.
$f$ is onto (or onto $Y$, if the codomain is not clear from context) if and only if for every $yin Y$ there at least one $xin X$ such that $f(x)=y$.
This definition applies to linear transformations as well, and in particular for linear transformations $Tcolon mathbb{R}^ntomathbb{R}^m$, and by extension to matrices, since an $mtimes n$ matrix $A$ can be identified with the linear transformation $L_Acolonmathbb{R}^ntomathbb{R}^m$ given by $L_A(mathbf{x}) = Amathbf{x}$.
So, the definitions are for any functions. But when our sets $X$ and $Y$ have more structure to them, and the functions are not arbitrary, but special kinds of functions, we can often obtain other ways of characterizing a function as one-to-one or onto which is easier/better/more useful/more conceptual/has interesting applications. This is indeed the case when we have such a rich structure as linear transformations and vector spaces.
One-to-one is probably the easiest; this is because whether a function is one-to-one depends only on its domain, and not on its codomain. By contrast, whether a function is onto depends on both on the domain and the codomain (so, for instance, $f(x)=x^2$ is onto if we think of it as a function $fcolonmathbb{R}to[0,infty)$, but not if we think of it as a function $fcolonmathbb{R}tomathbb{R}$, or $fcolon[2,infty)to[0,infty)$).
Theorem. Let $Tcolonmathbb{R}^ntomathbb{R}^m$ be a linear transformation. The following are equivalent:
$T$ is one-to-one.
$T(mathbf{x})=mathbf{0}$ has only the trivial solution $mathbf{x}=mathbf{0}$.
If $A$ is the standard matrix of $T$, then the columns of $A$ are linearly independent.
Proof. The equivalence of (1) and (2) is basic in linear algebra, so let's deal with that:
(1)$Rightarrow$(2): If $T$ is one-to-one, then for all $mathbf{x}$, since $T(mathbf{0})=mathbf{0}$ (being linear), then $T(mathbf{x})=mathbf{0}=T(mathbf{0})$ implies $mathbf{x}=mathbf{0}$; this proves (2).
(2)$Rightarrow$(1): Suppose $T(mathbf{x}_1)=T(mathbf{x}_2)$. Then
$$mathbf{0} = T(mathbf{x}_1) - T(mathbf{x}_2) = T(mathbf{x}_1-mathbf{x}_2),$$
since $T$ is linear; because we are assuming (2), $T(mathbf{x}_1-mathbf{x}_2)=mathbf{0}$ implies that $mathbf{x}_1-mathbf{x}_2 = mathbf{0}$, so $mathbf{x}_1=mathbf{x}_2$, which proves that $T$ is indeed one-to-one.
The key to the connection with (3) (and eventually to your confusion) is that multiplying a matrix by a vector can be seen as an operation on columns. If
$$A=left(begin{array}{cccc}
a_{11} & a_{12} & ldots & a_{1n}\
a_{21} & a_{22} & ldots & a_{2n}\
vdots & vdots & ddots & vdots\
a_{m1} & a_{m2} & ldots & a_{mn}
end{array}right),$$
then let the columns of $A$, $A_1$, $A_2,ldots,A_n$ be:
$$A_1 = left(begin{array}{c}a_{11}\a_{21}\ vdots\a_{m1}end{array}right),quad A_2 = left(begin{array}{c}a_{12}\a_{22}\ vdots\ a_{m2}end{array}right),quadldots,A_n = left(begin{array}{c}a_{1n}\a_{2n}\ vdots\ a_{mn}end{array}right).$$
Then we have the following:
$$Aleft(begin{array}{c}x_1\x_2\ vdots\x_nend{array}right) = x_1 A_1 + a_2 A_2 + cdots + a_nA_n.$$
That is, multiplying $A$ by $mathbf{x}$ gives a linear combination of the columns of $A$. This gives the direct connection we need between conditions (1) and (2), and condition (3).
(2)$Rightarrow$(3): To show that the columns of $A$ are linearly independent, we need to show that if $alpha_1A_1 + cdots + alpha_nA_n = mathbf{0}$, then $alpha_1=cdots=alpha_n=0$. So suppose $alpha_1A_1+cdots+alpha_nA_n = mathbf{0}$. Then
$$T(mathbf{alpha}) = Amathbf{alpha} = alpha_1A_1+cdots+alpha_nA_n = mathbf{0},qquadtext{where }mathbf{alpha}=left(begin{array}{c}alpha_1\ alpha_2\ vdots\ alpha_nend{array}right).$$
Because we are assuming (2), that means that from $T(mathbf{alpha}) = mathbf{0}$ we can conclude that $alpha=mathbf{0}$; therefore, $alpha_1=cdots=alpha_n=0$. This proves that $A_1,ldots,A_n$ are linearly independent.
(3)$Rightarrow$(2): Suppose the columns of $A$ are linearly independent, and
$$mathbf{0} = T(mathbf{x}) = Amathbf{x}quadtext{where }mathbf{x}=left(begin{array}{c}a_1\a_2\ vdots\ a_nend{array}right).$$
This means that $a_1A_1 + cdots a_nA_n = mathbf{0}$; since the columns of $A$ are assumed to be linearly independent, we conclude that $a_1=cdots=a_n=0$, so $mathbf{x}=mathbf{0}$, proving (2). QED
What about onto? There are two things here. One is a theorem similar to the one above; the other is the Rank-Nullity Theorem.
Theorem. Let $Tcolonmathbb{R}^ntomathbb{R}^m$ be a linear transformation. The following are equivalent:
$T$ is onto.
The equation $T(mathbf{x})=mathbf{b}$ has solutions for every $mathbf{b}inmathbb{R}^m$.
If $A$ is the standard matrix of $T$, then the columns of $A$ span $mathbb{R}^m$. That is: every $mathbf{b}inmathbb{R}^m$ is a linear combination of the columns of $A$.
Proof. (1)$Leftrightarrow$(2) is essentially the definition, only cast in terms of equations for the sake of similarity to the previous theorem.
(2)$Rightarrow$(3) Let $mathbf{b}inmathbb{R}^m$. Then by (2) there exists an $mathbf{a}inmathbb{R}^n$ such that $T(mathbf{a})=mathbf{b}$. We have:
$$mathbf{b} = T(mathbf{a}) = Amathbf{a} = Aleft(begin{array}{c}a_1\a_2\ vdots\a_nend{array}right) = a_1A_1 + a_2A_2 + cdots + a_nA_n.$$
That is, we can express $mathbf{b}$ as a linear combination of the columns of $A$. Since $mathbf{b}$ is arbitrary, every vector in $mathbb{R}^m$ can be expressed as a linear combination of the columns of $A$, so the columns of $A$ span $mathbb{R}^m$; this proves (3).
(3)$Rightarrow$(2) Suppose the columns of $A$ span $mathbb{R}^m$ and let $mathbf{b}inmathbb{R}^m$. We want to show that $T(mathbf{x})=mathbf{b}$ has at least one solution.
Since the columns of $A$ span $mathbb{R}^m$, there exist scalars $alpha_1,ldots,alpha_n$ such that
$$mathbf{b} = alpha_1 A_1 + cdots + alpha_n A_n = Aleft(begin{array}{c}alpha_1\ alpha_2\ vdots\ alpha_nend{array}right) = T(mathbf{alpha}).$$
So $mathbf{alpha}$, where
$$mathbf{alpha} = left(begin{array}{c}alpha_1\ alpha_2\ vdots\ alpha_nend{array}right),$$
is a solution to $T(mathbf{x})=mathbf{b}$. This establishes (2). QED
So: "one-to-one"-ness is related to linear independence; "onto"-ness is related to spanning properties. Note that linear independence is an intrinsic property (it depends only on the set of vectors), whereas spanning is an extrinsic property (it depends also on the space we are considering; it is contextual). This matches the fact that whether a function is one-to-one or not depends only on the domain, but whether it is onto depends on both the domain and the codomain of the function.
But there is a deep connection between the two. Remember the following:
Definition. Let $A$ be an $mtimes n$ matrix. The nullity of $A$, $mathrm{nullity}(A)$, is the dimension of the kernel of $A$, that is, of the subspace of $mathbb{R}^n$ given by
$$mathrm{ker}(A) = Bigl{ mathbf{x}inmathbb{R}^nBigm| Amathbf{x}=mathbf{0}Bigr}.$$
The rank of $A$, $mathrm{rank}(A)$ is the dimension of the image of $A$; that is, of the subspace of $mathbb{R}^m$ given by
begin{align*}
mathrm{Im}(A) &= Bigl{ mathbf{b}inmathbb{R}^mBigm| Amathbf{x}=mathbf{b}text{ has at least one solution}Bigr}\
&= Bigl{ A(mathbf{x})Bigm|mathbf{x}inmathbb{R}^nBigr}.
end{align*}
The deep connection between them is given by the Rank-Nullity Theorem:
Rank-Nullity Theorem. Let $A$ be an $mtimes n$ matrix. Then
$$mathrm{rank}(A) + mathrm{nullity}(A) = n.$$
Now we get two more equivalences for one-to-one and onto:
Theorem. Let $Tcolonmathbb{R}^ntomathbb{R^m}$ be a linear transformation. The following are equivalent:
$T$ is one-to-one.
The equation $T(mathbf{x})=mathbf{0}$ has only the trivial solution $mathbf{x}=mathbf{0}$.
If $A$ is the standard matrix of $T$, then the columns of $A$ are linearly independent.
$mathrm{ker}(A) = {mathbf{0}}$.
$mathrm{nullity}(A) = 0$.
$mathrm{rank}(A) = n$.
Proof. The equivalence of (4) and (5) follows because only the trivial subspace has dimension $0$; the equivalence of (4) and (2) follows by definition of the kernel. The equivalence of (5) and (6) follows from the Rank-Nullity Theorem, since $n = mathrm{nullity}(A)+mathrm{rank}(A)$, so $mathrm{nullity}(A) = 0$ if and only if $mathrm{rank}(A) = n$. Since we already know (1), (2), and (3) are equivalent, the result follows. QED
Theorem. Let $Tcolonmathbb{R}^ntomathbb{R}^m$ be a linear transformation. The following are equivalent:
$T$ is onto.
The equation $T(mathbf{x})=mathbf{b}$ has solutions for every $mathbf{b}inmathbb{R}^m$.
If $A$ is the standard matrix of $T$, then the columns of $A$ span $mathbb{R}^m$. That is: every $mathbf{b}inmathbb{R}^m$ is a linear combination of the columns of $A$.
$mathrm{Im}(A) = mathbb{R}^m$.
$mathrm{rank}(A) = m$.
$mathrm{nullity}(A) = n-m$.
Proof. We already know that (1), (2), and (3) are equivalent. The equivalence of (4) and (2) follows by the definition of the image. The equivalence of (4) and (5) follows because the only subspace of $mathbb{R}^m$ that has dimension $m$ is the whole space. Finally, the equivalence of (5) and (6) follows from the rank nullity theorem: since $n = mathrm{rank}(A)+mathrm{nullity}(A)$, then $mathrm{nullity}(A) = n - mathrm{rank}(A)$. So the rank equals $m$ if and only if the nullity equals $n-m$. QED
So now you have a whole bunch of ways of checking if a matrix is one-to-one, and of checking if a matrix is onto. None of them is "better" than the others: for some matrices, one will be easier to check, for other matrices, it may be a different one which is easy to check. Also, the rank of a matrix is closely related to its row-echelon form, so that might help as well.
Note a few things: generally, "onto" and "one-to-one" are independent of one another. You can have a matrix be onto but not one-to-one; or be one-to-one but not onto; or be both; or be neither. The Rank-Nullity Theorem does place some restrictions: if $A$ is $mtimes n$ and $mlt n$, then the matrix cannot be onto (because $1leqmathrm{rank}(A)leq m$, so if $mathrm{rank}(A)+mathrm{nullity}(A) = n$, we must have $mathrm{nullity}(A)gt 0$); dually, if $mgt n$ then $A$ cannot be onto. In particular, the only matrices that can be both one-to-one and onto are square matrices. On the other hand, you can have an $mtimes n$ matrix with $mlt n$ that is onto, or one that is not onto. And you can have $mtimes n$ matrices with $mgt n$ that are one-to-one, and matrices that are not one-to-one.
"One-to-one" and "onto" are properties of functions in general, not just linear transformations.
Definition. Let $fcolon Xto Y$ be a function.
$f$ is one-to-one if and only if for every $yin Y$ there is at most one $xin X$ such that $f(x)=y$; equivalently, if and only if $f(x_1)=f(x_2)$ implies $x_1=x_2$.
$f$ is onto (or onto $Y$, if the codomain is not clear from context) if and only if for every $yin Y$ there at least one $xin X$ such that $f(x)=y$.
This definition applies to linear transformations as well, and in particular for linear transformations $Tcolon mathbb{R}^ntomathbb{R}^m$, and by extension to matrices, since an $mtimes n$ matrix $A$ can be identified with the linear transformation $L_Acolonmathbb{R}^ntomathbb{R}^m$ given by $L_A(mathbf{x}) = Amathbf{x}$.
So, the definitions are for any functions. But when our sets $X$ and $Y$ have more structure to them, and the functions are not arbitrary, but special kinds of functions, we can often obtain other ways of characterizing a function as one-to-one or onto which is easier/better/more useful/more conceptual/has interesting applications. This is indeed the case when we have such a rich structure as linear transformations and vector spaces.
One-to-one is probably the easiest; this is because whether a function is one-to-one depends only on its domain, and not on its codomain. By contrast, whether a function is onto depends on both on the domain and the codomain (so, for instance, $f(x)=x^2$ is onto if we think of it as a function $fcolonmathbb{R}to[0,infty)$, but not if we think of it as a function $fcolonmathbb{R}tomathbb{R}$, or $fcolon[2,infty)to[0,infty)$).
Theorem. Let $Tcolonmathbb{R}^ntomathbb{R}^m$ be a linear transformation. The following are equivalent:
$T$ is one-to-one.
$T(mathbf{x})=mathbf{0}$ has only the trivial solution $mathbf{x}=mathbf{0}$.
If $A$ is the standard matrix of $T$, then the columns of $A$ are linearly independent.
Proof. The equivalence of (1) and (2) is basic in linear algebra, so let's deal with that:
(1)$Rightarrow$(2): If $T$ is one-to-one, then for all $mathbf{x}$, since $T(mathbf{0})=mathbf{0}$ (being linear), then $T(mathbf{x})=mathbf{0}=T(mathbf{0})$ implies $mathbf{x}=mathbf{0}$; this proves (2).
(2)$Rightarrow$(1): Suppose $T(mathbf{x}_1)=T(mathbf{x}_2)$. Then
$$mathbf{0} = T(mathbf{x}_1) - T(mathbf{x}_2) = T(mathbf{x}_1-mathbf{x}_2),$$
since $T$ is linear; because we are assuming (2), $T(mathbf{x}_1-mathbf{x}_2)=mathbf{0}$ implies that $mathbf{x}_1-mathbf{x}_2 = mathbf{0}$, so $mathbf{x}_1=mathbf{x}_2$, which proves that $T$ is indeed one-to-one.
The key to the connection with (3) (and eventually to your confusion) is that multiplying a matrix by a vector can be seen as an operation on columns. If
$$A=left(begin{array}{cccc}
a_{11} & a_{12} & ldots & a_{1n}\
a_{21} & a_{22} & ldots & a_{2n}\
vdots & vdots & ddots & vdots\
a_{m1} & a_{m2} & ldots & a_{mn}
end{array}right),$$
then let the columns of $A$, $A_1$, $A_2,ldots,A_n$ be:
$$A_1 = left(begin{array}{c}a_{11}\a_{21}\ vdots\a_{m1}end{array}right),quad A_2 = left(begin{array}{c}a_{12}\a_{22}\ vdots\ a_{m2}end{array}right),quadldots,A_n = left(begin{array}{c}a_{1n}\a_{2n}\ vdots\ a_{mn}end{array}right).$$
Then we have the following:
$$Aleft(begin{array}{c}x_1\x_2\ vdots\x_nend{array}right) = x_1 A_1 + a_2 A_2 + cdots + a_nA_n.$$
That is, multiplying $A$ by $mathbf{x}$ gives a linear combination of the columns of $A$. This gives the direct connection we need between conditions (1) and (2), and condition (3).
(2)$Rightarrow$(3): To show that the columns of $A$ are linearly independent, we need to show that if $alpha_1A_1 + cdots + alpha_nA_n = mathbf{0}$, then $alpha_1=cdots=alpha_n=0$. So suppose $alpha_1A_1+cdots+alpha_nA_n = mathbf{0}$. Then
$$T(mathbf{alpha}) = Amathbf{alpha} = alpha_1A_1+cdots+alpha_nA_n = mathbf{0},qquadtext{where }mathbf{alpha}=left(begin{array}{c}alpha_1\ alpha_2\ vdots\ alpha_nend{array}right).$$
Because we are assuming (2), that means that from $T(mathbf{alpha}) = mathbf{0}$ we can conclude that $alpha=mathbf{0}$; therefore, $alpha_1=cdots=alpha_n=0$. This proves that $A_1,ldots,A_n$ are linearly independent.
(3)$Rightarrow$(2): Suppose the columns of $A$ are linearly independent, and
$$mathbf{0} = T(mathbf{x}) = Amathbf{x}quadtext{where }mathbf{x}=left(begin{array}{c}a_1\a_2\ vdots\ a_nend{array}right).$$
This means that $a_1A_1 + cdots a_nA_n = mathbf{0}$; since the columns of $A$ are assumed to be linearly independent, we conclude that $a_1=cdots=a_n=0$, so $mathbf{x}=mathbf{0}$, proving (2). QED
What about onto? There are two things here. One is a theorem similar to the one above; the other is the Rank-Nullity Theorem.
Theorem. Let $Tcolonmathbb{R}^ntomathbb{R}^m$ be a linear transformation. The following are equivalent:
$T$ is onto.
The equation $T(mathbf{x})=mathbf{b}$ has solutions for every $mathbf{b}inmathbb{R}^m$.
If $A$ is the standard matrix of $T$, then the columns of $A$ span $mathbb{R}^m$. That is: every $mathbf{b}inmathbb{R}^m$ is a linear combination of the columns of $A$.
Proof. (1)$Leftrightarrow$(2) is essentially the definition, only cast in terms of equations for the sake of similarity to the previous theorem.
(2)$Rightarrow$(3) Let $mathbf{b}inmathbb{R}^m$. Then by (2) there exists an $mathbf{a}inmathbb{R}^n$ such that $T(mathbf{a})=mathbf{b}$. We have:
$$mathbf{b} = T(mathbf{a}) = Amathbf{a} = Aleft(begin{array}{c}a_1\a_2\ vdots\a_nend{array}right) = a_1A_1 + a_2A_2 + cdots + a_nA_n.$$
That is, we can express $mathbf{b}$ as a linear combination of the columns of $A$. Since $mathbf{b}$ is arbitrary, every vector in $mathbb{R}^m$ can be expressed as a linear combination of the columns of $A$, so the columns of $A$ span $mathbb{R}^m$; this proves (3).
(3)$Rightarrow$(2) Suppose the columns of $A$ span $mathbb{R}^m$ and let $mathbf{b}inmathbb{R}^m$. We want to show that $T(mathbf{x})=mathbf{b}$ has at least one solution.
Since the columns of $A$ span $mathbb{R}^m$, there exist scalars $alpha_1,ldots,alpha_n$ such that
$$mathbf{b} = alpha_1 A_1 + cdots + alpha_n A_n = Aleft(begin{array}{c}alpha_1\ alpha_2\ vdots\ alpha_nend{array}right) = T(mathbf{alpha}).$$
So $mathbf{alpha}$, where
$$mathbf{alpha} = left(begin{array}{c}alpha_1\ alpha_2\ vdots\ alpha_nend{array}right),$$
is a solution to $T(mathbf{x})=mathbf{b}$. This establishes (2). QED
So: "one-to-one"-ness is related to linear independence; "onto"-ness is related to spanning properties. Note that linear independence is an intrinsic property (it depends only on the set of vectors), whereas spanning is an extrinsic property (it depends also on the space we are considering; it is contextual). This matches the fact that whether a function is one-to-one or not depends only on the domain, but whether it is onto depends on both the domain and the codomain of the function.
But there is a deep connection between the two. Remember the following:
Definition. Let $A$ be an $mtimes n$ matrix. The nullity of $A$, $mathrm{nullity}(A)$, is the dimension of the kernel of $A$, that is, of the subspace of $mathbb{R}^n$ given by
$$mathrm{ker}(A) = Bigl{ mathbf{x}inmathbb{R}^nBigm| Amathbf{x}=mathbf{0}Bigr}.$$
The rank of $A$, $mathrm{rank}(A)$ is the dimension of the image of $A$; that is, of the subspace of $mathbb{R}^m$ given by
begin{align*}
mathrm{Im}(A) &= Bigl{ mathbf{b}inmathbb{R}^mBigm| Amathbf{x}=mathbf{b}text{ has at least one solution}Bigr}\
&= Bigl{ A(mathbf{x})Bigm|mathbf{x}inmathbb{R}^nBigr}.
end{align*}
The deep connection between them is given by the Rank-Nullity Theorem:
Rank-Nullity Theorem. Let $A$ be an $mtimes n$ matrix. Then
$$mathrm{rank}(A) + mathrm{nullity}(A) = n.$$
Now we get two more equivalences for one-to-one and onto:
Theorem. Let $Tcolonmathbb{R}^ntomathbb{R^m}$ be a linear transformation. The following are equivalent:
$T$ is one-to-one.
The equation $T(mathbf{x})=mathbf{0}$ has only the trivial solution $mathbf{x}=mathbf{0}$.
If $A$ is the standard matrix of $T$, then the columns of $A$ are linearly independent.
$mathrm{ker}(A) = {mathbf{0}}$.
$mathrm{nullity}(A) = 0$.
$mathrm{rank}(A) = n$.
Proof. The equivalence of (4) and (5) follows because only the trivial subspace has dimension $0$; the equivalence of (4) and (2) follows by definition of the kernel. The equivalence of (5) and (6) follows from the Rank-Nullity Theorem, since $n = mathrm{nullity}(A)+mathrm{rank}(A)$, so $mathrm{nullity}(A) = 0$ if and only if $mathrm{rank}(A) = n$. Since we already know (1), (2), and (3) are equivalent, the result follows. QED
Theorem. Let $Tcolonmathbb{R}^ntomathbb{R}^m$ be a linear transformation. The following are equivalent:
$T$ is onto.
The equation $T(mathbf{x})=mathbf{b}$ has solutions for every $mathbf{b}inmathbb{R}^m$.
If $A$ is the standard matrix of $T$, then the columns of $A$ span $mathbb{R}^m$. That is: every $mathbf{b}inmathbb{R}^m$ is a linear combination of the columns of $A$.
$mathrm{Im}(A) = mathbb{R}^m$.
$mathrm{rank}(A) = m$.
$mathrm{nullity}(A) = n-m$.
Proof. We already know that (1), (2), and (3) are equivalent. The equivalence of (4) and (2) follows by the definition of the image. The equivalence of (4) and (5) follows because the only subspace of $mathbb{R}^m$ that has dimension $m$ is the whole space. Finally, the equivalence of (5) and (6) follows from the rank nullity theorem: since $n = mathrm{rank}(A)+mathrm{nullity}(A)$, then $mathrm{nullity}(A) = n - mathrm{rank}(A)$. So the rank equals $m$ if and only if the nullity equals $n-m$. QED
So now you have a whole bunch of ways of checking if a matrix is one-to-one, and of checking if a matrix is onto. None of them is "better" than the others: for some matrices, one will be easier to check, for other matrices, it may be a different one which is easy to check. Also, the rank of a matrix is closely related to its row-echelon form, so that might help as well.
Note a few things: generally, "onto" and "one-to-one" are independent of one another. You can have a matrix be onto but not one-to-one; or be one-to-one but not onto; or be both; or be neither. The Rank-Nullity Theorem does place some restrictions: if $A$ is $mtimes n$ and $mlt n$, then the matrix cannot be onto (because $1leqmathrm{rank}(A)leq m$, so if $mathrm{rank}(A)+mathrm{nullity}(A) = n$, we must have $mathrm{nullity}(A)gt 0$); dually, if $mgt n$ then $A$ cannot be onto. In particular, the only matrices that can be both one-to-one and onto are square matrices. On the other hand, you can have an $mtimes n$ matrix with $mlt n$ that is onto, or one that is not onto. And you can have $mtimes n$ matrices with $mgt n$ that are one-to-one, and matrices that are not one-to-one.
edited Dec 10 '18 at 0:52
answered Mar 11 '11 at 17:46
Arturo Magidin
261k32584904
261k32584904
Sorry if this is a stupid question, but is it always the case that whether a function is one-to-one depends only on the domain? For instance with the example $f(x) = x^2$ if the codomain were $0$, wouldn't the function still be one-to-one?
– ghshtalt
Mar 11 '11 at 18:13
1
@ghshtalt: For $f(x)=x^2$ to have codomain $0$, you would need the domain to be just $0$ (otherwise, you wouldn't have a function). But of course, if the domain of a function is a single point, then the function is one-to-one, regardless of what the codomain (or what the function) is. So in fact, it's because the domain is a single point that such a function would be one-to-one.
– Arturo Magidin
Mar 11 '11 at 18:19
oh ok, so it wouldn't have been a function in the first place... Thanks for clearing that up for me! And thanks for another great answer of course!
– ghshtalt
Mar 11 '11 at 18:22
I believe I have an example that is not one-to-one nor onto. The matrix is lin.dep (free variables), and for a random value in the codomain the reduced augmented matrix is inconsistent. Hence there is not one value in the domain that corresponds to that value in the codomain. Right, right?
– Algific
Mar 13 '11 at 17:26
3
@Algific: Matrices by themselves are nor "linearly independent" or "linearly dependent". Sets of vectors are linearly independent or linearly dependent. If you mean that you have a matrix whose columns are linearly dependent (and somehow relating that to "free variables", yet another concept that is not directly applicable to matrices, but rather to other things to which you can associate matrices), then it is indeed true that any matrix whose columns are linearly dependent gives a linear transformation that is not one-to-one. If not all systems Ax=b have solutions, then not onto.
– Arturo Magidin
Mar 13 '11 at 20:28
|
show 3 more comments
Sorry if this is a stupid question, but is it always the case that whether a function is one-to-one depends only on the domain? For instance with the example $f(x) = x^2$ if the codomain were $0$, wouldn't the function still be one-to-one?
– ghshtalt
Mar 11 '11 at 18:13
1
@ghshtalt: For $f(x)=x^2$ to have codomain $0$, you would need the domain to be just $0$ (otherwise, you wouldn't have a function). But of course, if the domain of a function is a single point, then the function is one-to-one, regardless of what the codomain (or what the function) is. So in fact, it's because the domain is a single point that such a function would be one-to-one.
– Arturo Magidin
Mar 11 '11 at 18:19
oh ok, so it wouldn't have been a function in the first place... Thanks for clearing that up for me! And thanks for another great answer of course!
– ghshtalt
Mar 11 '11 at 18:22
I believe I have an example that is not one-to-one nor onto. The matrix is lin.dep (free variables), and for a random value in the codomain the reduced augmented matrix is inconsistent. Hence there is not one value in the domain that corresponds to that value in the codomain. Right, right?
– Algific
Mar 13 '11 at 17:26
3
@Algific: Matrices by themselves are nor "linearly independent" or "linearly dependent". Sets of vectors are linearly independent or linearly dependent. If you mean that you have a matrix whose columns are linearly dependent (and somehow relating that to "free variables", yet another concept that is not directly applicable to matrices, but rather to other things to which you can associate matrices), then it is indeed true that any matrix whose columns are linearly dependent gives a linear transformation that is not one-to-one. If not all systems Ax=b have solutions, then not onto.
– Arturo Magidin
Mar 13 '11 at 20:28
Sorry if this is a stupid question, but is it always the case that whether a function is one-to-one depends only on the domain? For instance with the example $f(x) = x^2$ if the codomain were $0$, wouldn't the function still be one-to-one?
– ghshtalt
Mar 11 '11 at 18:13
Sorry if this is a stupid question, but is it always the case that whether a function is one-to-one depends only on the domain? For instance with the example $f(x) = x^2$ if the codomain were $0$, wouldn't the function still be one-to-one?
– ghshtalt
Mar 11 '11 at 18:13
1
1
@ghshtalt: For $f(x)=x^2$ to have codomain $0$, you would need the domain to be just $0$ (otherwise, you wouldn't have a function). But of course, if the domain of a function is a single point, then the function is one-to-one, regardless of what the codomain (or what the function) is. So in fact, it's because the domain is a single point that such a function would be one-to-one.
– Arturo Magidin
Mar 11 '11 at 18:19
@ghshtalt: For $f(x)=x^2$ to have codomain $0$, you would need the domain to be just $0$ (otherwise, you wouldn't have a function). But of course, if the domain of a function is a single point, then the function is one-to-one, regardless of what the codomain (or what the function) is. So in fact, it's because the domain is a single point that such a function would be one-to-one.
– Arturo Magidin
Mar 11 '11 at 18:19
oh ok, so it wouldn't have been a function in the first place... Thanks for clearing that up for me! And thanks for another great answer of course!
– ghshtalt
Mar 11 '11 at 18:22
oh ok, so it wouldn't have been a function in the first place... Thanks for clearing that up for me! And thanks for another great answer of course!
– ghshtalt
Mar 11 '11 at 18:22
I believe I have an example that is not one-to-one nor onto. The matrix is lin.dep (free variables), and for a random value in the codomain the reduced augmented matrix is inconsistent. Hence there is not one value in the domain that corresponds to that value in the codomain. Right, right?
– Algific
Mar 13 '11 at 17:26
I believe I have an example that is not one-to-one nor onto. The matrix is lin.dep (free variables), and for a random value in the codomain the reduced augmented matrix is inconsistent. Hence there is not one value in the domain that corresponds to that value in the codomain. Right, right?
– Algific
Mar 13 '11 at 17:26
3
3
@Algific: Matrices by themselves are nor "linearly independent" or "linearly dependent". Sets of vectors are linearly independent or linearly dependent. If you mean that you have a matrix whose columns are linearly dependent (and somehow relating that to "free variables", yet another concept that is not directly applicable to matrices, but rather to other things to which you can associate matrices), then it is indeed true that any matrix whose columns are linearly dependent gives a linear transformation that is not one-to-one. If not all systems Ax=b have solutions, then not onto.
– Arturo Magidin
Mar 13 '11 at 20:28
@Algific: Matrices by themselves are nor "linearly independent" or "linearly dependent". Sets of vectors are linearly independent or linearly dependent. If you mean that you have a matrix whose columns are linearly dependent (and somehow relating that to "free variables", yet another concept that is not directly applicable to matrices, but rather to other things to which you can associate matrices), then it is indeed true that any matrix whose columns are linearly dependent gives a linear transformation that is not one-to-one. If not all systems Ax=b have solutions, then not onto.
– Arturo Magidin
Mar 13 '11 at 20:28
|
show 3 more comments
let $T(x)=Ax$ be a linear transformation.
$T(x)$ is one-to-one if the columns of $A$ are linearly independent.
$T(x)$ is onto if every row of $A$ has a pivot.
Is this truly how easy it is? I know the first one, one-to-one if linearly independent, but didn't know the second one. So does it matter if the system has free variables? As long as every row has a pivot it's onto?
– joe_04_04
Mar 23 '17 at 9:19
This still didn't answer the parent question of "Would a zero-row in reduced echelon form have any effect on (ONTO)? " That is to ask "T(x) is onto if every non-zero row of A has a pivot."?
– Ron Jensen
Sep 26 '17 at 20:56
add a comment |
let $T(x)=Ax$ be a linear transformation.
$T(x)$ is one-to-one if the columns of $A$ are linearly independent.
$T(x)$ is onto if every row of $A$ has a pivot.
Is this truly how easy it is? I know the first one, one-to-one if linearly independent, but didn't know the second one. So does it matter if the system has free variables? As long as every row has a pivot it's onto?
– joe_04_04
Mar 23 '17 at 9:19
This still didn't answer the parent question of "Would a zero-row in reduced echelon form have any effect on (ONTO)? " That is to ask "T(x) is onto if every non-zero row of A has a pivot."?
– Ron Jensen
Sep 26 '17 at 20:56
add a comment |
let $T(x)=Ax$ be a linear transformation.
$T(x)$ is one-to-one if the columns of $A$ are linearly independent.
$T(x)$ is onto if every row of $A$ has a pivot.
let $T(x)=Ax$ be a linear transformation.
$T(x)$ is one-to-one if the columns of $A$ are linearly independent.
$T(x)$ is onto if every row of $A$ has a pivot.
edited Oct 11 '13 at 23:03
answered Oct 11 '13 at 16:48
vajra78
309213
309213
Is this truly how easy it is? I know the first one, one-to-one if linearly independent, but didn't know the second one. So does it matter if the system has free variables? As long as every row has a pivot it's onto?
– joe_04_04
Mar 23 '17 at 9:19
This still didn't answer the parent question of "Would a zero-row in reduced echelon form have any effect on (ONTO)? " That is to ask "T(x) is onto if every non-zero row of A has a pivot."?
– Ron Jensen
Sep 26 '17 at 20:56
add a comment |
Is this truly how easy it is? I know the first one, one-to-one if linearly independent, but didn't know the second one. So does it matter if the system has free variables? As long as every row has a pivot it's onto?
– joe_04_04
Mar 23 '17 at 9:19
This still didn't answer the parent question of "Would a zero-row in reduced echelon form have any effect on (ONTO)? " That is to ask "T(x) is onto if every non-zero row of A has a pivot."?
– Ron Jensen
Sep 26 '17 at 20:56
Is this truly how easy it is? I know the first one, one-to-one if linearly independent, but didn't know the second one. So does it matter if the system has free variables? As long as every row has a pivot it's onto?
– joe_04_04
Mar 23 '17 at 9:19
Is this truly how easy it is? I know the first one, one-to-one if linearly independent, but didn't know the second one. So does it matter if the system has free variables? As long as every row has a pivot it's onto?
– joe_04_04
Mar 23 '17 at 9:19
This still didn't answer the parent question of "Would a zero-row in reduced echelon form have any effect on (ONTO)? " That is to ask "T(x) is onto if every non-zero row of A has a pivot."?
– Ron Jensen
Sep 26 '17 at 20:56
This still didn't answer the parent question of "Would a zero-row in reduced echelon form have any effect on (ONTO)? " That is to ask "T(x) is onto if every non-zero row of A has a pivot."?
– Ron Jensen
Sep 26 '17 at 20:56
add a comment |
A function is onto if its codomain is equal to its image, that is, $f: X to Y$ is onto if $f(X) = Y$.
Assuming you are working with real matrices, a matrix $A$ is usually seen as mapping vectors from $mathbb R^n$ to $mathbb R^m$. You want to show that for all $vec y in mathbb R^m$ there exists a $vec x in mathbb R^n$ such that $A vec x = vec y$. In some cases this is not possible depending on $m$ and $n$. How does $m$ and $n$ relate to the size of the matrix?
One way to check that a matrix is onto is to check if you can find vectors $vec v_1, vec v_2, dots, vec v_m$ such that $Avec v_i = vec e_i$ for $i = 1, 2, dots, m$, where $vec e_1, vec e_2, dots, vec e_m$ is the standard basis in $mathbb R^m$ (why?). Can you use the echelon form for this? Do you actually need to get the standard basis or is some other set of vectors ok?
The question is sort of an open one, it says discuss the concepts of one-to-one and onto with regard to the two following matrices.
– Algific
Mar 11 '11 at 12:11
Well, to start of, check if they are one-to-one and onto. Does a matrix have to be one-to-one to be onto? Does it have to be onto to be one-to-one?
– Calle
Mar 11 '11 at 12:17
The problem is that the theorems doesn't mention the other state. Funny enough.
– Algific
Mar 11 '11 at 16:06
add a comment |
A function is onto if its codomain is equal to its image, that is, $f: X to Y$ is onto if $f(X) = Y$.
Assuming you are working with real matrices, a matrix $A$ is usually seen as mapping vectors from $mathbb R^n$ to $mathbb R^m$. You want to show that for all $vec y in mathbb R^m$ there exists a $vec x in mathbb R^n$ such that $A vec x = vec y$. In some cases this is not possible depending on $m$ and $n$. How does $m$ and $n$ relate to the size of the matrix?
One way to check that a matrix is onto is to check if you can find vectors $vec v_1, vec v_2, dots, vec v_m$ such that $Avec v_i = vec e_i$ for $i = 1, 2, dots, m$, where $vec e_1, vec e_2, dots, vec e_m$ is the standard basis in $mathbb R^m$ (why?). Can you use the echelon form for this? Do you actually need to get the standard basis or is some other set of vectors ok?
The question is sort of an open one, it says discuss the concepts of one-to-one and onto with regard to the two following matrices.
– Algific
Mar 11 '11 at 12:11
Well, to start of, check if they are one-to-one and onto. Does a matrix have to be one-to-one to be onto? Does it have to be onto to be one-to-one?
– Calle
Mar 11 '11 at 12:17
The problem is that the theorems doesn't mention the other state. Funny enough.
– Algific
Mar 11 '11 at 16:06
add a comment |
A function is onto if its codomain is equal to its image, that is, $f: X to Y$ is onto if $f(X) = Y$.
Assuming you are working with real matrices, a matrix $A$ is usually seen as mapping vectors from $mathbb R^n$ to $mathbb R^m$. You want to show that for all $vec y in mathbb R^m$ there exists a $vec x in mathbb R^n$ such that $A vec x = vec y$. In some cases this is not possible depending on $m$ and $n$. How does $m$ and $n$ relate to the size of the matrix?
One way to check that a matrix is onto is to check if you can find vectors $vec v_1, vec v_2, dots, vec v_m$ such that $Avec v_i = vec e_i$ for $i = 1, 2, dots, m$, where $vec e_1, vec e_2, dots, vec e_m$ is the standard basis in $mathbb R^m$ (why?). Can you use the echelon form for this? Do you actually need to get the standard basis or is some other set of vectors ok?
A function is onto if its codomain is equal to its image, that is, $f: X to Y$ is onto if $f(X) = Y$.
Assuming you are working with real matrices, a matrix $A$ is usually seen as mapping vectors from $mathbb R^n$ to $mathbb R^m$. You want to show that for all $vec y in mathbb R^m$ there exists a $vec x in mathbb R^n$ such that $A vec x = vec y$. In some cases this is not possible depending on $m$ and $n$. How does $m$ and $n$ relate to the size of the matrix?
One way to check that a matrix is onto is to check if you can find vectors $vec v_1, vec v_2, dots, vec v_m$ such that $Avec v_i = vec e_i$ for $i = 1, 2, dots, m$, where $vec e_1, vec e_2, dots, vec e_m$ is the standard basis in $mathbb R^m$ (why?). Can you use the echelon form for this? Do you actually need to get the standard basis or is some other set of vectors ok?
edited Mar 11 '11 at 12:01
answered Mar 11 '11 at 11:43
Calle
6,36412342
6,36412342
The question is sort of an open one, it says discuss the concepts of one-to-one and onto with regard to the two following matrices.
– Algific
Mar 11 '11 at 12:11
Well, to start of, check if they are one-to-one and onto. Does a matrix have to be one-to-one to be onto? Does it have to be onto to be one-to-one?
– Calle
Mar 11 '11 at 12:17
The problem is that the theorems doesn't mention the other state. Funny enough.
– Algific
Mar 11 '11 at 16:06
add a comment |
The question is sort of an open one, it says discuss the concepts of one-to-one and onto with regard to the two following matrices.
– Algific
Mar 11 '11 at 12:11
Well, to start of, check if they are one-to-one and onto. Does a matrix have to be one-to-one to be onto? Does it have to be onto to be one-to-one?
– Calle
Mar 11 '11 at 12:17
The problem is that the theorems doesn't mention the other state. Funny enough.
– Algific
Mar 11 '11 at 16:06
The question is sort of an open one, it says discuss the concepts of one-to-one and onto with regard to the two following matrices.
– Algific
Mar 11 '11 at 12:11
The question is sort of an open one, it says discuss the concepts of one-to-one and onto with regard to the two following matrices.
– Algific
Mar 11 '11 at 12:11
Well, to start of, check if they are one-to-one and onto. Does a matrix have to be one-to-one to be onto? Does it have to be onto to be one-to-one?
– Calle
Mar 11 '11 at 12:17
Well, to start of, check if they are one-to-one and onto. Does a matrix have to be one-to-one to be onto? Does it have to be onto to be one-to-one?
– Calle
Mar 11 '11 at 12:17
The problem is that the theorems doesn't mention the other state. Funny enough.
– Algific
Mar 11 '11 at 16:06
The problem is that the theorems doesn't mention the other state. Funny enough.
– Algific
Mar 11 '11 at 16:06
add a comment |
Thanks for contributing an answer to Mathematics Stack Exchange!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
Use MathJax to format equations. MathJax reference.
To learn more, see our tips on writing great answers.
Some of your past answers have not been well-received, and you're in danger of being blocked from answering.
Please pay close attention to the following guidance:
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f26371%2fis-a-linear-tranformation-onto-or-one-to-one%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
What is your definition of an onto matrix? Full (row) rank?
– wildildildlife
Mar 11 '11 at 11:00
I'll update when I get home :)
– Algific
Mar 11 '11 at 11:19