+0  
 
0
365
1
avatar

solve ab=a+b=a/b

 Nov 14, 2016
 #1
avatar
0

First I chose two of the equations: ab=a/b

(ab)/b = (a/b)/b

a = a/b^2

(a)/a = (a/b^2)/a

1 = a/b^2/a

1 = 1/b^2

b^2 = 1

b = sqrt(1)

b = 1

 

Plugged it back into the equations

a(1) = a + 1 = a/1

a = a + 1 = a/1

a = a + 1 is false

There is no solution.

 

The closest that's possible is 0, where:

(0)(0) = 0 + 0 = 0/0

0 = 0 = undefined

 

0/0 is undefined and thus there is no solution.

 

If you want reasoning for 0/0 being undefined:

It is understood that any number over itself equals 1 (such as 1/1 = 1, 2/2 = 1, 27/27 = 1)

It is also understood that zero divided by any number equals 0 (such as 0/1 = 1, 0/2 = 0, and 0/27 = 0)

 

Because both arguments stand, you cannot define 0/0 as either 1 or 0, leaving it as undefined.

 Nov 14, 2016

0 Online Users