My Math Forum  

Go Back   My Math Forum > College Math Forum > Linear Algebra

Linear Algebra Linear Algebra Math Forum

Thanks Tree1Thanks
  • 1 Post By romsek
LinkBack Thread Tools Display Modes
December 8th, 2016, 07:46 PM   #1
Joined: Dec 2016
From: Downtown

Posts: 1
Thanks: 0



Stuck on this question, no idea what to do or how to solve it
Peaya is offline  
December 8th, 2016, 08:21 PM   #2
Senior Member
romsek's Avatar
Joined: Sep 2015
From: USA

Posts: 1,862
Thanks: 968


$V=(v_1,~v_2,~v_3)$ be a $3 \times 3$ matrix whose columns are your 3 linearly independent $v$'s


$F$ be a $3 \times 3$ matrix representing your invertible linear transformation $f$

finally let

$B = (f(v_1),~f(v_2),~f(v_3)) = (Fv_1,~Fv_2,~Fv_3)=FV$

i.e. $B$ is the matrix whose columns are the $v$'s transformed by $F$

Now suppose the columns of $B$ are not linearly independent.

Then $\exists x \ni Bx =0$ so


But $F$ is invertible so multiply both sides by $F^{-1}$

$F^{-1}F V x = 0$

$Vx = 0$

and thus the column vectors of $V$ are linearly dependent.

but we are given they are linearly independent and thus we have a contradiction and it must be that the columns of $B$ are linearly independent.

I'll let you think about (b), it's important
Thanks from Peaya
romsek is offline  

  My Math Forum > College Math Forum > Linear Algebra


Thread Tools
Display Modes

Similar Threads
Thread Thread Starter Forum Replies Last Post
vector as product of matrices, exp function of vector whitegreen Linear Algebra 1 June 9th, 2015 06:11 AM
vector product of a vector and a scalar? 71GA Algebra 1 June 3rd, 2012 01:20 PM
Determining a 3D Vector B after a rotation of 3D Vector A babarorhum Algebra 0 October 20th, 2011 03:53 PM
Vector Calculus Divergence of a Vector Field MasterOfDisaster Calculus 2 September 26th, 2011 09:17 AM
Spherical Vector to Cartesian Vector tsa256 Algebra 3 August 20th, 2010 07:33 AM

Copyright © 2018 My Math Forum. All rights reserved.