this post was submitted on 01 Dec 2023
334 points (98.0% liked)

Programmer Humor

32453 readers
713 users here now

Post funny things about programming here! (Or just rant about your favourite programming language.)

Rules:

founded 5 years ago
MODERATORS
 

I know this isn't any kind of surprise, and yet, well...

you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 42 points 11 months ago (7 children)

2100 and 2400 will be a shitshow

[–] [email protected] 35 points 11 months ago (2 children)
[–] [email protected] 10 points 11 months ago

Yeah that's a different shitshow but agreed it is likely to be worse - like y2k the effects are smeared out before and after the date.

[–] [email protected] 2 points 11 months ago (2 children)
[–] [email protected] 8 points 11 months ago
[–] [email protected] 6 points 11 months ago* (last edited 11 months ago) (1 children)

32bit systems will stop working. The Unix timestamp, which increases by 1 every second and started the first second of 1970, will reach the max of 32 bit integers. Bad things will follow.

[–] [email protected] 10 points 11 months ago

This has already been patched on all 64 bit OSes though - whatever 32 bit systems are still in existence in another 15 years will just roll their dates back 50 years and add another layer of duct tape to their jerry-rigged existence

[–] [email protected] 16 points 11 months ago

2038 will certainly be a shit show

[–] [email protected] 8 points 11 months ago

Yeah but I'll be dead so not my problem lmao

[–] [email protected] 3 points 11 months ago (1 children)

Nah.

Same thing happened in 2000 and it was a mouse’s fart.

[–] [email protected] 18 points 11 months ago (1 children)

Because of months of preparation. I know, I was doing it.

[–] [email protected] 3 points 11 months ago (1 children)

And now that every time library has been updated, we're safe until our grandchildren reimplement those bugs in a language that has not yet been invented.

[–] [email protected] 8 points 11 months ago (2 children)

I've already seen reimplementation of 2 digit dates here and there.

[–] [email protected] 5 points 11 months ago (1 children)
[–] [email protected] 4 points 11 months ago (1 children)

Fortunately I will not be involved. Hopefully I can make something from 2038 though.

[–] [email protected] 2 points 11 months ago

You’re not the only one forseeing a nice consultant payday there.

[–] [email protected] 3 points 11 months ago

I went to uni in the mid 90s when Y2K prep was all the rage, went back to do another degree 20 years later. It was interesting to see the graffiti in the CS toilets. Two digits up to about 1996, four digits for a decade, then back to two.

[–] [email protected] 3 points 11 months ago

Luckily, none of us will be there.

[–] [email protected] 3 points 11 months ago (2 children)
[–] [email protected] 16 points 11 months ago (1 children)

2100 not a leap year (divisible by 100). 2400 is a leap year (divisible by 400). Developing for dates is a minefield.

[–] [email protected] 1 points 11 months ago* (last edited 11 months ago)

Now imagine working on non Georgian, and the year is 2060

[–] [email protected] 8 points 11 months ago

Because they're not leap years but are 0 === year % 4

[–] [email protected] 1 points 11 months ago (1 children)

Won’t the computer’s clock reset every time you go to sleep and stop cranking the power generator?

[–] [email protected] 3 points 11 months ago

Yeah who knows if our computers are sticks by either date