Redkey

joined 1 year ago
[–] [email protected] 3 points 2 months ago* (last edited 2 months ago)

Whoops! When I looked at the second time that the shift value is calculated, I wondered if it would be inverted from the first time, but for some reason I decided that it wouldn't be. But looking at it again it's clear now that (1 - i) = (-i + 1) = ((~i + 1) + 1), making bit 0 the inverse. Then I wondered why there wasn't more corruption and realized that the author's compiler must perform postfix increments and decrements immediately after the variable is used, so the initial shift is also inverted. That's why the character pairs are flipped, but they still decode correctly otherwise. I hope this version works better:

long main () {
    char output;
    unsigned char shift;
    long temp;
    
    if (i < 152) {
        shift = (~i & 1) * 7;
        temp = b[i >> 1] >> shift;
        i++;
        output = (char)(64 & temp);
        output += (char)((n >> (temp & 63)) & main());
        printf("%c", output);
    }

    return 63;
}

EDIT: I just got a chance to compile it and it does work.

[–] [email protected] 10 points 2 months ago (3 children)

I first learned about Java in the late 90s and it sounded fantastic. "Write once, run anywhere!" Great!

After I got past "Hello world!" and other simple text output tutorials, things took a turn for the worse. It seemed like if you wanted to do just about anything beyond producing text output with compile-time data (e.g. graphics, sound, file access), you needed to figure out what platform and which edition/version of Java your program was being run on, so you could import the right libraries and call the right functions with the right parameters. I guess that technically this was still "write once, run anywhere".

After that, I learned just enough Java to squeak past a university project that required it, then promptly forgot all of it.

I feel like Sun was trying to hit multiple moving targets at the same time, and failing to land a solid hit on any of them. They were laser-focused on portable binaries, but without standardized storage or multimedia APIs at a time when even low-powered devices were starting to come with those capabilities. I presume that things are better now, but I've never been tempted to have another look. Even just trying to get my machines set up to run other people's Java programs has been enough to keep me away.

[–] [email protected] 8 points 2 months ago* (last edited 2 months ago) (4 children)

I don't know if this will work or even compile, but I feel like I'm pretty close.

long main () {
    char output;
    unsigned char shift;
    long temp;
    
    if (i < 152) {
        shift = (i & 1) * 7;
        temp = b[i >> 1] >> shift;
        i++;
        output = (char)(64 & temp);
        output += (char)((n >> (temp & 63)) & main());
        printf("%c", output);
    }

    return 63;
}
[–] [email protected] 4 points 3 months ago

This genie must've read or watched Brewster's Millions.

[–] [email protected] 5 points 3 months ago* (last edited 3 months ago)

I saw it at the cinema and vaguely remember enjoying it well enough. It's not a great movie, but it's not awful, either. I didn't know that it was supposed to be terrible; it looks like reviewers gave it a slightly better than average score.

I don't expect ever to watch it a second time, if that helps.

Lara Croft and the Cradle of Life, though... All I can remember about it now is that afterwards, my friends and I agreed that we should've trusted our instincts and just walked out after about 30 minutes.

[–] [email protected] 4 points 3 months ago

That's still newer than any of my daily-use laptops that are all running full-featured Linux distros just fine. I got 'em all cheap secondhand, and just pumped up the RAM (12-16GB) and installed SSDs.

[–] [email protected] 11 points 4 months ago* (last edited 4 months ago) (1 children)

Did you read all the way to the end of the article? I did.

At the very bottom of the piece, I found that the author had already expressed what I wanted to say quite well:

In my humble opinion, here’s the key takeaway: just write your own fucking constructors! You see all that nonsense? Almost completely avoidable if you had just written your own fucking constructors. Don’t let the compiler figure it out for you. You’re the one in control here.

The joke here isn't C++. The joke is people who expect C++ to be as warm, fuzzy, and forgiving as JavaScript.

[–] [email protected] 36 points 4 months ago (1 children)

Yeah, I'm sure that almost all of us have felt this way at one time or another. But the thing is, every team behind every moronic, bone-headed interface "update" that you've ever hated also sees themselves in the programmer's position in this meme.

[–] [email protected] 7 points 4 months ago (1 children)

Since you seem earnest, probably play_my_game or possibly gamedev.

[–] [email protected] 16 points 4 months ago (3 children)

I reserve further comments until I know whether you posted this in this community: a) deliberately but seriously, b) deliberately and sarcastically, or c) by accident.

[–] [email protected] 3 points 5 months ago (1 children)

Any time I need to learn something about JS, I go to W3Schools to wrap my head around the basics, then over to MDN for current best practice.

view more: next ›