2/9 = 0.222… 7/9 = 0.777…
0.222… + 0.777… = 0.999… 2/9 + 7/9 = 1
0.999… = 1
No, it equals 1.
Similarly, 1/3 = 0.3333…
So 3 times 1/3 = 0.9999… but also 3/3 = 1
Another nice one:
Let x = 0.9999… (multiply both sides by 10)
10x = 9.99999… (substitute 0.9999… = x)
10x = 9 + x (subtract x from both sides)
9x = 9 (divide both sides by 9)
x = 1
My favorite thing about this argument is that not only are you right, but you can prove it with math.
That’s more convoluted than the 1/3, 2/3, 3/3 thing.
3/3 = 0.99999…
3/3 = 1
If somebody still wants to argue after that, don’t bother.
Sure, when you start decoupling the numbers from their actual values. The only thing this proves is that the fraction-to-decimal conversion is inaccurate. Your floating points (and for that matter, our mathematical model) don’t have enough precision to appropriately model what the value of 7/9 actually is. The variation is negligible though, and that’s the core of this, is the variation off what it actually is is so small as to be insignificant and, really undefinable to us - but that doesn’t actually matter in practice, so we just ignore it or convert it. But at the end of the day 0.999… does not equal 1. A number which is not 1 is not equal to 1. That would be absurd. We’re just bad at converting fractions in our current mathematical understanding.
Edit: wow, this has proven HIGHLY unpopular, probably because it’s apparently incorrect. See below for about a dozen people educating me on math I’ve never heard of. The “intuitive” explanation on the Wikipedia page for this makes zero sense to me largely because I don’t understand how and why a repeating decimal can be considered a real number. But I’ll leave that to the math nerds and shut my mouth on the subject.
You are just wrong.
The rigorous explanation for why 0.999…=1 is that 0.999… represents a geometric series of the form 9/10+9/10^2+… by definition, i.e. this is what that notation literally means. The sum of this series follows by taking the limit of the corresponding partial sums of this series (see here) which happens to evaluate to 1 in the particular case of 0.999… this step is by definition of a convergent infinite series.
The only thing this proves is that the fraction-to-decimal conversion is inaccurate.
No number is getting converted, it’s the same number in both cases but written in a different representation. 4 is also the same number as IV, no conversion going on it’s still the natural number elsewhere written S(S(S(S(Z))))
. Also decimal representation isn’t inaccurate, it just happens to have multiple valid representations for the same number.
A number which is not 1 is not equal to 1.
Good then that 0.999… and 1 are not numbers, but representations.
It still equals 1, you can prove it without using fractions:
x = 0.999…
10x = 9.999…
10x = 9 + 0.999…
10x = 9 + x
9x = 9
x = 1
There’s even a Wikipedia page on the subject
If they aren’t equal, there should be a number in between that separates them. Between 0.1 and 0.2 i can come up with 0.15. Between 0.1 and 0.15 is 0.125. You can keep going, but if the numbers are equal, there is nothing in between. There’s no gap between 0.1 and 0.1, so they are equal.
What number comes between 0.999… and 1?
(I used to think it was imprecise representations too, but this is how it made sense to me :)
My brother. You are scared of infinities. Look up the infinite hotel problem. I will lay it out for you if you are interested.
Image you are incharge of a hotel and it has infinite rooms. Currently your hotel is at full capacity… Meaning all rooms are occupied. A new guest arrives. What do you do? Surely your hotel is full and you can’t take him in… Right? WRONG!!! You tell the resident of room 1 to move to room 2, you tell the resident of room 2 to move to room 3 and so on… You tell the resident of room n to move to room n+1. Now you have room 1 empty
But sir… How did I create an extra room? You didn’t. The question is the same as asking yourself that is there a number for which n+1 doesn’t exist. The answer is no… I can always add 1.
Infinity doesn’t behave like other numbers since it isn’t technically a number.
So when you write 0.99999… You are playing with things that aren’t normal. Maths has come with fuckall ways to deal with stuff like this.
Well you may say, this is absurd… There is nothing in reality that behaves this way. Well yes and no. You know how the building blocks of our universe obey quantum mechanics? The equations contain lots of infinities but only at intermediate steps. You have to “renormalise” them to make them go away. Nature apparently has infinities but likes to hide the from us.
The infinity problem is so fucked up. You know the reason physics people are unable to quantize gravity? Surely they can do the same thing to gravity as they did to say electromagnetic force? NOPE. Gravitation doesn’t normalise. You get left with infinities in your final answer.
Anyways. Keep on learning, the world has a lot of information and it’s amazing. And the only thing that makes us human is the ability to learn and grow from it. I wish you all the very best.
It’s a correct proof.
One way to think about this is that we represent numbers in different ways. For example, 1 can be 1.0, or a single hash mark, or a dot, or 1/1, or 10/10. All of them point to some platonic ideal world version of the concept of the number 1.
What we have here is two different representations of the same number that are in a similar representation. 1 and 0.999… both point to the same concept.
I strongly agree with you, and while the people replying aren’t wrong, they’re arguing for something that I don’t think you said.
1/3 ≈ 0.333… in the same way that approximating a circle with polygons of increasing side number has a limit of a circle, but will never yeild a circle with just geometry.
0.999… ≈ 1 in the same way that shuffling infinite people around an infinite hotel leaves infinite free rooms, but if you try to do the paperwork, no one will ever get anywhere.
Decimals require you to check the end of the number to see if you can round up, but there never will be an end. Thus we need higher mathematics to avoid the halting problem. People get taught how decimals work, find this bug, and then instead of being told how decimals are broken, get told how they’re wrong for using the tools they’ve been taught.
If we just accept that decimals fail with infinite steps, the transition to new tools would be so much easier, and reflect the same transition into new tools in other sciences. Like Bohr’s Atom, Newton’s Gravity, Linnaean Taxonomy, or Comte’s Positivism.