answersLogoWhite

0

Let's start out with the basic inequality 1 < 2 < 4.

Now, we'll take the square root of this inequality:

1 < √2 < 2.

If you subtract all numbers by 1, you get:

0 < √2 - 1 < 1.

If √2 is rational, then it can be expressed as a fraction of two integers, m/n. This next part is the only remotely tricky part of this proof, so pay attention. We're going to assume that m/n is in its most reduced form; i.e., that the value for n is the smallest it can be and still be able to represent √2. Therefore, √2n must be an integer, and n must be the smallest multiple of √2 to make this true. If you don't understand this part, read it again, because this is the heart of the proof.

Now, we're going to multiply √2n by (√2 - 1). This gives 2n - √2n. Well, 2n is an integer, and, as we explained above, √2n is also an integer; therefore, 2n - √2n is an integer as well. We're going to rearrange this expression to (√2n - n)√2 and then set the term (√2n - n) equal to p, for simplicity. This gives us the expression √2p, which is equal to 2n - √2n, and is an integer.

Remember, from above, that 0 < √2 - 1 < 1.

If we multiply this inequality by n, we get 0 < √2n - n < n, or, from what we defined above, 0 < p < n. This means that p < n and thus √2p < √2n. We've already determined that both √2p and √2n are integers, but recall that we said n was the smallest multiple of √2 to yield an integer value. Thus, √2p < √2n is a contradiction; therefore √2 can't be rational and so must be irrational.

Q.E.D.

User Avatar

Wiki User

14y ago

What else can I help you with?

Related Questions

Prove that square root of 2 is an irrational number?

The square root of 2 is 1.141..... is an irrational number


Prove the square root of 2 not irrational?

This is impossible to prove, as the square root of 2 is irrational.


How do you prove that root 2 root 3 is an irrational number?

It is known that the square root of an integer is either an integer or irrational. If we square root2 root3 we get 6. The square root of 6 is irrational. Therefore, root2 root3 is irrational.


Prove that root 2 is irrational number?

I linked a good resource that explains what you asked below.


Is the square root of 6 rational?

No; you can prove the square root of any positive number that's not a perfect square is irrational, using a similar method to showing the square root of 2 is irrational.


Is the square root a 2 an irrational number?

Yes, the square root of 2 is an irrational number.


Is the square root of a 143 a irrational number?

Yes. The square root of a positive integer can ONLY be either:* An integer (in this case, it isn't), OR * An irrational number. The proof is basically the same as the proof used in high school algebra, to prove that the square root of 2 is irrational.


Is a square root of 2 rational or irrational?

The square root of 2 is an irrational number


How is root 2 irrational?

2 is a prime number and its square root is an irrational number that cannot be expressed as a fraction


Is the square root of 2 an irrational number or a rational number?

irrational


Is the square root of a irrational number always rational?

No, the square root of an irrational number is not always rational. In fact, the square root of an irrational number is typically also irrational. For example, the square root of 2, which is an irrational number, is itself irrational. However, there are exceptions, such as the square root of a perfect square of an irrational number, which can be rational.


Can multiplying two irrational numbers give you an irrational number?

Yes. For example, the square root of 3 (an irrational number) times the square root of 2(an irrational number) gets you the square root of 6(an irrational number)