Linear Algebra- independence and dependence

Status
Not open for further replies.

Snitch

Junior Member level 3
Joined
Jun 5, 2012
Messages
27
Helped
0
Reputation
0
Reaction score
0
Trophy points
1,281
Location
Sri Lanka
Activity points
1,469
Let u and v be non-zero vectors in V. Prove or disprove the following claim. u and v are linearly dependent ⟹ (u+v) and (u−v) are linearly dependent.

Is the following proof correct? Otherwise can someone answer the given question? Thanks in advance.

Here is my answer.

Since u,v are two linearly dependent vectors, au+bv=0 with a,b≠0
we can write, [(a+b)/2]⋅(u+v)+[(a−b)/2]⋅(u−v)=0
case 1 (|a|=|b|) :
if a=b then (a+b)/2≠0⟹(u+v) and (u−v) are linearly dependent.

if a=−b then (a−b)/2≠0⟹(u+v) and (u−v) are linearly dependent.

case 2 (|a|≠|b|) :
then (a+b)/2,(a−b)/2≠0⟹(u+v) and (u−v) are linearly dependent.

So, u and v are linearly dependent ⟹(u+v) and (u−v) are linearly dependent.
 

Status
Not open for further replies.
Cookies are required to use this site. You must accept them to continue using the site. Learn more…