megaman

joined 1 year ago
[–] [email protected] 1 points 1 week ago

it is indeed infrequent, but the modern world has trained me to expect convenience and instant-ness. Last time i wanted a 12-year-old email I was in the car with friends and and to pull it up. it wasn't anything important at all, to be clear, but i'm hoping to search my 12-year-old emails with the same convenience as last month's.

[–] [email protected] 2 points 1 week ago

I think that that is right that I fundamentally want an archive, not what a normal mail server provides. Part of my thought on looking at mail servers is that those would integrate directly with whatever other front-end/client that I'd normally use, whereas an archive maybe would not.

And regarding archive-specific stuff, I am seeing some things on a search, but I guess i'm wondering if folks here have any recommendations. When I look at , for example, nothing comes up for email archive, just for email servers. That, plus what I see when searching, makes me think that the archive-specific stuff is either oriented to business or oriented to a CLI (like NotMuch, which was mentioned in the discussion here and does look cool).

[–] [email protected] 2 points 1 week ago (2 children)

This looks like a good backend for sure, but the web frontends look a little lacking and I'm not seeing anything about a mobile frontend (other than if a web one was up, which would be fine). Have you tried any of the web frontends?

 

Yet another question about self-hosting email, but I haven't found the answer at least phrased in a way that makes sense with my question.

I've got ~15 GBs of old gmail data that I've already downloaded, and google is on my ass about "91% full" and we know I'm not about to pay them for storage (I'll sooner spend 100 hours trying to solve it myself before I pay them $3/month).

What I want is to have the same (or relatively close to the same) access and experience to find stuff in those old emails when they are stored on my hardware as I do when they are in my gmail. That is, I want to have a website and/or app that i search for emails from so-and-so, in some date-range, keywords. I don't actually want to send any emails from this server or receive anything to it (maybe I would want gmail to forward to it or something, but probably I'd just do another archive batch every year).

What I've tried so far, which is sort of working, is that I've set up docker-mailserver on my box, and that is working and accessible. I can connect to it via Thunderbird or K-9 mail. I also converted big email download from google, which was a .mbox, into maildir using mb2md (apt install mb2md on debian was nice). This gave me a directory with ~120k individual email files.

When I check this out in Thunderbird, I see all those emails (and they look like they have the right info) (as a side - I actually only moved 1k emails into the directory that docker-mailserver has access to, just for testing, and Thunderbird only sees that 1k then). I can do some searching on those.

When I open in K-9, it by default looks like it just pulls in 100 of them. I can pull in more or refresh it sort of thing. I don't normally use K-9, so I may just be missing how the functionality there is supposed to work.

I also just tried connecting to the mail server with Nextcloud Mail, which works in the sense that it connects but it (1) seems like it is struggling, and (2) is putting 'today' as the date for all the emails rather than when they actually came through. I don't really want to use Nextcloud Mail here...

So, I think my question here is now really around search and storage. In Thunderbird, I think that the way it works (I don't normally use Thunderbird much either) is that it downloads all the files locally, and then it will search them locally. In K-9 that appears to be the same, but with the caveat that it doesn't look like it really wants to download 120k emails locally (even if I can).

What I think I want to do, though, is have the search running on the server. Like I don't want to download 15GBs (and another 9 from gmail soon enough) to each client. I want it all on the server and just put in my search and the server do the query and give me a response.

docker-mailserver has a page for setting up Full-Text Search with Xapian, where it'll make all the indices and all that. I tinkered with this and think I got it set up. This is another sort of thing where I would want the search to be utilizing the server rather than client since the server is (hopefully) optimizing for some of this stuff.

Should I be using a different server for what I want here? I've poked around at different ones and am more than open to changing to something else that is more for what I need here.

For clients, should I be using Roundcube or something else? Will that actually help with this 'use the server to search' question? For mobile, is there any way to avoid downloading all the emails to the client?

Thanks for the help.

[–] [email protected] 1 points 3 weeks ago

This article isnt about how emails associated with logins got released in a breach, but that documents that are uploaded to the archive are stamped with the email address of the account that uploaded it and that can be viewed by anyone who downloads the document.

So in standard, everyday use of the site, email addresses are being revealed and are associated with the actions of that person. Like if I upload a copy of the manual for my washing machine or something, which is a more benign example, my email is linked to that document now.

Then combine this with (1) the internet archive says in multiple spots that they dont reveal this info anywhere, and (2) the issue has been raised to the organization, and it becomes more of a specific negligence from them.

[–] [email protected] 4 points 3 weeks ago

This article isnt about how emails associated with logins got released in a breach, but that documents that are uploaded to the archive are stamped with the email address of the account that uploaded it and that can be viewed by anyone who downloads the document.

So in standard, everyday use of the site, email addresses are being revealed and are associated with the actions of that person. Like if I upload a copy of the manual for my washing machine or something, which is a more benign example, my email is linked to that document now.

Then combine this with (1) the internet archive says in multiple spots that they dont reveal this info anywhere, and (2) the issue has been raised to the organization, and it becomes more of a specific negligence from them.

[–] [email protected] 2 points 1 month ago

The nsa wants to watch people who are watching the pornhub video of someone else watching porn. The third level there is more difficult to find

[–] [email protected] 6 points 3 months ago

The folks who found it are presenting at Defcon this weekend, according to the article.

I imagine some of the industry press (i.e. Wired) are just looking through the Defcon agenda to figure out what to write. I saw two or three other articles about hacks or exploits and things like that that also mentioned it was bring presented at Defcon.

[–] [email protected] 3 points 3 months ago (1 children)

Unrelated to your actual post (plan to read later), but is your RSS busted? The rss link on the webpage gives a 404 and my RSS reader is erroring on it as well...

[–] [email protected] 5 points 4 months ago* (last edited 4 months ago)

Ive got this working with Caddy and Adguard

I use Caddy as my reverse proxy. It is running on the machine in the basement with all the different docker-container-services on different ports. My registrar is set up so that *.my-domain.com goes to my IP.

Caddy is then configured for 'service-a.my-domain.com' to port 1234, and the others going to their ports. This is just completely standard reverse proxy.

For some subdomains (i.e. different services) ive whitelisted only the local network. There is some config for that.

Im pretty sure that I also have to have adguard do a dns rewrite on the local network as well. That is, adguard has a rewrite for '*.my-domain.com' to go to 192.168.0.22 (the local machine with caddy). I think i had to do this to ensure that when the request gets to caddy it is coming from the local whitelisted network rather than my public IP (which changes every couple months, but could be more).

[–] [email protected] 9 points 4 months ago (4 children)

Everyone who downvoted me didnt read the article, or didnt read what i said, or didnt read op, or something, i dont remember what they didnt read but they cannot be real because the only way to disagree with me is to not have read something or other (or did read it, cant remember which)

[–] [email protected] 12 points 4 months ago (6 children)

I read the fun blogpost that is not an academic paper and ive downvoted you. Does that mean i dont actually exist or that u dont actually exist???

[–] [email protected] 21 points 4 months ago (1 children)

When i was doing a headless install, i spend a hour or two trying to figure out how to pre setup configs for the debian installer or how to do it over network or what before i finally lugged the new machine to the other room and plugged it into the monitor and keyboard of the main rig, installed it all (and set up ssh so i can later get into from the main rig), and unplugged it.

My point is, even if it isnt trivial to have the keyboard and monitor, it may be much easier to get them than to really do an install without them.

 

I've got my main house server that has a number of dockerized applications, including nextcloud-aio. Nextcloud-AIO comes with a built-in backup system using BorgBackups. I've had this running and doing my backups, it is probably fine. Notable, it does encrypt the backup.

Now, I recently setup a separate machine to use rsnapshot to backup the things from the main machine that need backing up. It is SSHing on a schedule to do that, and backing up the folders I've listed.

When I set that up, I skipped the nextcloud borg backup, because that is already backing up; however, it is not a remote backup, so is of limited use (granted, my 'official' backup computer is using about 18 inches away from the main server, so also of limited use).

I can easily just include the nextcloud-borg-directory on the rsnapshot list, but does anyone know if it will properly handle just the updates?

That is, both Borg and Rsnapshot are set up so that each backup isn't a complete backup but just incremental changes, so that you don't fill your whole disk in two weeks. But if Borg does that first on the nextcloud data, will rsnapshot just not work and then try to backup the full 50GBs every day? Or just do the incremental changes? Will the borg encryption jack up the ability of rsnapshot to see the changes?

If no one knows, I will just do it anyway and report back in a few days if my disk is completely full or not.

 

Hey, all.

Is it possible to skip this 'register your server' step when creating a self-hosted Rocketchat instance? I just don't want to, ya know? Regular websearching is just giving a lot about how to disable user registration rather than skipping the server registration with Rocketchat HQ.

view more: next ›