this post was submitted on 01 Dec 2023
334 points (98.0% liked)

Programmer Humor

32050 readers
1619 users here now

Post funny things about programming here! (Or just rant about your favourite programming language.)

Rules:

founded 5 years ago
MODERATORS
 

I know this isn't any kind of surprise, and yet, well...

you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 42 points 9 months ago (7 children)

2100 and 2400 will be a shitshow

[–] [email protected] 35 points 9 months ago (2 children)
[–] [email protected] 10 points 9 months ago

Yeah that's a different shitshow but agreed it is likely to be worse - like y2k the effects are smeared out before and after the date.

[–] [email protected] 2 points 9 months ago (2 children)
[–] [email protected] 8 points 9 months ago
[–] [email protected] 6 points 9 months ago* (last edited 9 months ago) (1 children)

32bit systems will stop working. The Unix timestamp, which increases by 1 every second and started the first second of 1970, will reach the max of 32 bit integers. Bad things will follow.

[–] [email protected] 10 points 9 months ago

This has already been patched on all 64 bit OSes though - whatever 32 bit systems are still in existence in another 15 years will just roll their dates back 50 years and add another layer of duct tape to their jerry-rigged existence

[–] [email protected] 16 points 9 months ago

2038 will certainly be a shit show

[–] [email protected] 8 points 9 months ago

Yeah but I'll be dead so not my problem lmao

[–] [email protected] 3 points 9 months ago (1 children)

Nah.

Same thing happened in 2000 and it was a mouse’s fart.

[–] [email protected] 18 points 9 months ago (1 children)

Because of months of preparation. I know, I was doing it.

[–] [email protected] 3 points 9 months ago (1 children)

And now that every time library has been updated, we're safe until our grandchildren reimplement those bugs in a language that has not yet been invented.

[–] [email protected] 8 points 9 months ago (2 children)

I've already seen reimplementation of 2 digit dates here and there.

[–] [email protected] 5 points 9 months ago (1 children)
[–] [email protected] 4 points 9 months ago (1 children)

Fortunately I will not be involved. Hopefully I can make something from 2038 though.

[–] [email protected] 2 points 9 months ago

You’re not the only one forseeing a nice consultant payday there.

[–] [email protected] 3 points 9 months ago

I went to uni in the mid 90s when Y2K prep was all the rage, went back to do another degree 20 years later. It was interesting to see the graffiti in the CS toilets. Two digits up to about 1996, four digits for a decade, then back to two.

[–] [email protected] 3 points 9 months ago

Luckily, none of us will be there.

[–] [email protected] 3 points 9 months ago (2 children)
[–] [email protected] 16 points 9 months ago (1 children)

2100 not a leap year (divisible by 100). 2400 is a leap year (divisible by 400). Developing for dates is a minefield.

[–] [email protected] 1 points 9 months ago* (last edited 9 months ago)

Now imagine working on non Georgian, and the year is 2060

[–] [email protected] 8 points 9 months ago

Because they're not leap years but are 0 === year % 4

[–] [email protected] 1 points 9 months ago (1 children)

Won’t the computer’s clock reset every time you go to sleep and stop cranking the power generator?

[–] [email protected] 3 points 9 months ago

Yeah who knows if our computers are sticks by either date