this post was submitted on 02 Oct 2024
171 points (95.2% liked)

Technology

59374 readers
7416 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 97 points 1 month ago (5 children)

Telegram was built to protect activists and ordinary people from corrupt governments and corporations – we do not allow criminals to abuse our platform to evade justice.

So who gets to pick what's a lawful request and criminal activity? It's criminal in some states to seek an abortion or help with an abortion, so would they hand out the IPs of those "criminals"? Because depending on who you ask some will tell you they're basically murderers. And that's just one example.

Good privacy apps have nothing to hand out to any government, like Signal.

[–] [email protected] 33 points 1 month ago (1 children)

Exactly. The strive for zero knowledge is the proper way to be going.

[–] [email protected] 12 points 1 month ago (1 children)

But then you can't sell your customer's data for profit. Even if you don't now, you still have that option in the future.

[–] [email protected] 6 points 1 month ago

Exactly. Which is the entire reason you should do it. Since you can't sell your customers for profit, that means you have to profit off of your customers. And another business could start up and compete with you. Also, your customers will trust you more.

[–] [email protected] 10 points 1 month ago* (last edited 1 month ago) (1 children)

The second I went to sign up and learned a phone number was absolutely required, I knew that their privacy was pure bullshit. That little declaration at the end here is an absolute slap to the face.

[–] [email protected] 18 points 1 month ago* (last edited 1 month ago) (1 children)

Signal requires that as well. Their privacy is definitely not bullshit. As far as I can tell, it's a spam mitigation method. But yeah, Telegram is pretty much the very bottom of privacy. Even Meta now encrypts all messages across all platforms.

[–] [email protected] 1 points 1 month ago (3 children)

It's bad for privacy no matter how you sell it. Unless you have a good amount of disposable income to buy up burner numbers all the time, a phone number tends to be incredibly identifying. So if a government agency comes along saying "Hey, we know this account sent this message and you have to give us everything you have about this account," for the average person, it doesn't end up being that different than having given them your full id.

[–] [email protected] 9 points 1 month ago (1 children)

Another aspect is the social graph. It's targeted for normies to easily switch to.

Very few people want to install a communication app, open the compose screen for the first time, and be met by an empty list of who they can communicate with.

https://signal.org/blog/private-contact-discovery/

By using phone numbers, you can message your friends without needing to have them all register usernames and tell them to you. It also means Signal doesn't need to keep a copy of your contact list on their servers, everyone has their local contact list.

This means private messages for loads of people, their goal.

Hey, we know this account sent this message and you have to give us everything you have about this account

It's a bit backwards, since your account is your phone number, the agency would be asking "give us everything you have from this number". They've already IDed you at that point.

[–] [email protected] 0 points 1 month ago

Yep, at that point they're just fishing for more which, hey, why wouldn't they.

It's a give and take for sure, requiring a real phone number makes it harder for automated spam bots to use the service, but at the same time, it puts the weight of true privacy on the shoulders and wallets of the users, and in a lesser way, incentives the use of less than reputable services, should a user want to truly keep their activities private.

And yeah, there's an argument to be made for keeping crime at bay, but that also comes with risks itself. If there was some way to keep truly egregious use at bay while not risking a $10,000 fine on someone for downloading an episode of Ms. Marvel, I think that would be great.

[–] [email protected] 7 points 1 month ago* (last edited 1 month ago) (1 children)

It's bad for privacy no matter how you sell it.

I mean it's not ideal but as long as it's not tied to literally any other information, the way Signal does it, it's "fine", and certainly not "bad" and especially not "pure bullshit".

So if a government agency comes along saying "Hey, we know this account sent this message and you have to give us everything you have about this account,"

They have done this several times, they give them nothing because they have nothing.

[–] [email protected] 0 points 1 month ago* (last edited 1 month ago) (1 children)

Says right there in the subpoena "You are required to provide all information tied to the following phone numbers." This means that the phone number requirement has already created a leak of private information in this instance, Signal simply couldn't add more to it.

Additionally, that was posted in 2021. Since then, Signal has introduced usernames to "keep your phone number private." Good for your average Joe Blow, but should another subpoena be submitted, now stating "You are required to provide all information tied to the following usernames," this time they will have something to give, being the user's phone number, which can then be used to tie any use of Signal they already have proof of back to the individual.

Yeah, it's great that they don't log what you send, but that doesn't help if they get proof in any other way. The fact is, because of the phone number requirement, anything you ever send on Signal can easily be tied back to you should it get out, and that subpoena alone is proof that it does.

[–] [email protected] 9 points 1 month ago* (last edited 1 month ago) (1 children)

This means that the phone number requirement has already created a leak of private information

What information? The gov already had the phone number. They needed it to make the request.

Additionally, that was posted in 2021.

Here's a more recent one.. Matter of fact, here's a full list of all of them. Notice the lack of any usernames provided.

Also note that a bunch of the numbers they requested weren't even registered with Signal, so the gov didn't even know if they were using the app and were just throwing shit at the wall and seeing what sticks.

You are required to provide all information tied to the following usernames

They can't respond to requests for usernames because they don't know any of them. From Signal: "Once again, Signal doesn’t have access to your messages; your calls; your chat list; your files and attachments; your stories; your groups; your contacts; your stickers; your profile name or avatar; your reactions; or even the animated GIFs you search for – and it’s impossible to turn over any data that we never had access to in the first place."

What else ya got?

but that doesn't help if they get proof in any other way.

If they're getting evidence outside of Signal, that's outside the scope of this discussion.

because of the phone number requirement, anything you ever send on Signal can easily be tied back to you should it get out

...no. It can't.

that subpoena alone is proof that it does.

It's proof that it doesn't.

[–] [email protected] 5 points 1 month ago

Guys like you see privacy as a monolith, that it never is. Unusable privacy is meanigless as email had shown. Privacy of communications does not mean privacy of communicators and usable authentication can be more important then anonymity.

And all this has to be realised on real-world servers, that are always in reach of real world goverment.

[–] [email protected] 1 points 1 month ago* (last edited 1 month ago)

In the US, agents must petition a judge for a search warrant. If granted, the agent may then compel an IT company to produce. If they are able, they must comply. It isn't up to the CEO to decide what he feels is right.

Look for services that allow your data to be encrypted, but it must also clearly state the service provider does not have the encryption keys -- you do. Apple does this, I believe.

[–] [email protected] 0 points 1 month ago

So who gets to pick what’s a lawful request and criminal activity?

Probably Telegram themselves. Durov was forced into exile by Putin.