lal309

joined 1 year ago
[–] [email protected] 2 points 2 months ago

I was under the impression that Google retired the "app password" workflow and moved to Gmail API within Google Cloud. I have the API set up and that's what I'm using in the Vikunja configs but like I mentioned in the post, at this point I don't care if its Gmail or something else. I just need the email functionality to work so I will use whatever service works well with Vikunja.

18
submitted 2 months ago* (last edited 2 months ago) by [email protected] to c/[email protected]
 

Does anyone have a working Vikunja instance sending emails through Gmail? I’ve enabled the mailer options and entered the info but the test_email function times out. I’ve checked all the information and even tried different ports.

Honestly at this point it doesn’t have to be Gmail (I’m just most familiar with this workflow). I just need my Vikunja instance to send emails.

Edit: I was able to solve my issue. You can only create Gmail app passwords if you have 2FA enabled. I also had the wrong address (it’s smtp.gmqil.com not smtp.google.com)

[–] [email protected] 2 points 4 months ago

I don’t have an answer for you but I have one instead. When I attempted to do swarm my biggest challenge was shared storage. I was attempting to run a swarm with shared storage on a NAS. Literally could not run apps, ran into a ton of problems running stacks (NAS share tried SMB and NFS). How did you get around this problem?

[–] [email protected] 2 points 5 months ago (1 children)

Well she stopped using it so I deleted the instance (just too busy with small children). For the time that she was using it Bookstack seemed to have me her needs once a cohesive breakdown was established (translating Bookstack hierarchy and matching it up with her topics).

[–] [email protected] 2 points 6 months ago

Following. Sounds interesting.

[–] [email protected] 26 points 6 months ago

Lemme fix the headline for you.

HashiCorp joins the list of companies and software killed by IBM.

[–] [email protected] 1 points 7 months ago

Not quiet. I was running gitea before so my mount was ./gitea:/data but since switching over to forgejo, I renamed my ./gitea directory to ./forgejo. Adjusted my compose file to have a mount of ./forgejo:/data.

Now inside of that renamed forgejo directory, there are a bunch of gitea references and even one more directory called gitea. When I migrated everything worked right away but since I wanted a cleaner transition, I renamed and switched all gitea references to forgejo but went I brought the stack back online, it went belly up.

As a troubleshooting step, I recreated my compose file and created a new empty ./forgejo on a different machine just to see what a new and fresh install would look like and the forgejo stack itself created all kinds of gitea references and gitea directory once I brought it up. So to fix my original deployment, I reverted all the references back from forgejo to gitea and everything worked again.

For fun, I went out to codeberg to look at the Dockerfile and saw that they had a bunch of gitea things within their own Dockerfile so nothing I can do for now

[–] [email protected] 1 points 7 months ago

This looks interesting too

[–] [email protected] 5 points 7 months ago (1 children)

Love that username tho!! Yeah might just do RSS. I already run FreshRSS and it’s ability to filter stuff would probably come in handy too

[–] [email protected] 5 points 7 months ago

This sounds like the simplest and most effective solution. Thanks!

[–] [email protected] 1 points 7 months ago (1 children)

Interesting… do you like this way more or the rss route more?

[–] [email protected] 2 points 7 months ago (3 children)

Was not aware of Diun, will check out!

[–] [email protected] 3 points 7 months ago

After you mentioned it, I looked it up too and stumbled on a similar answer to that link. Thanks!

 

TL;DR: is there an app that can alert me when a new version of some other app is available?

I have about 12 - 15 services (freshrss, heimdall, photoprism, Wordpress, etc) running using docker compose spread across 4 hosts. Through my self-hosting journey I’ve been burned a few times using “latest” images so I now pin app image versions within compose.

The problem then becomes that every couple of weeks, I have to go out to different GitHub’s, docker hub, etc. to see if a new update for that service is available. It gets a bit tedious with 12-15 services every couple of weeks so I need a centralized and more efficient way of “keeping up”.

Is there some type of app that can track whether an app/service has a new version available? Ideally it can send me some type of notification, self-hostable, and ideally not Portainer?

 

I’m running a very small business and now I have a need to start tracking my sales and expenses for the business. Not looking for a full blown Quickbooks type of thing but if that’s all that’s available then no big deal, I can just use what I need and ignore the rest.

Obviously, I have to self host this. Hardware available varies but I have several raspberry pi’s laying around not doing much (3, 4 & 5). Ideally dockerized. My research shows GnuCash, Akaunting and Odoo.

What does this awesome community recommend?

P.S. Tried spinning up Akaunting on an rpi 5 and encountered a breaking bug (already reported to their github).

 

My apologies for the long post.

I have a single server running Unraid with about 12 services (Pihole, Wordpress, Heimdall, Jellyfin, etc.) all running on Docker. This server is also acting as my home lab NAS. Everything runs fine for my use case (at least for right now) but I’ve been thinking about creating some type of compute cluster for my services instead of a single server.

Recently, I saw a discussion about Proxmox, Docker, LXD and Incus where a user felt that Incus was a better option to all the others. Curious, I started reading up on Incus and playing around with it and contemplating switching all my services from Docker in Unraid to an Incus cluster (I’m thinking around 3 nodes) and leaving the Unraid server to serve as a NAS only.

In a nutshell Incus/LXD appear to be (to me) a lightweight version of a VM in which case I would have to manually install and configure each service I have running. Based on the services I have running, that seems like a ton of work to switch to Incus when I could just do 3 physical servers (Debian) in docker swarm mode or a Proxmox cluster with 3 Debian VMs with docker in swarm mode. I’d all possible, I would like to keep my services containerized rather then actual VMs.

What has me thinking that a switch to Incus may be worth it is performance. If the performance of my services is significantly better in Incus/LXDs as compared to Docker, then that’s worth it to me. I have not been able to find any type of performance comparison between Incus/LXD and Docker. I don’t know if there are other reasons as to “Incus over Proxmox and Docker” which is why I’m asking the greater community.

Here’s my question:

Based on your experience and taking into consideration my use case (home lab/home use), do the pros and cons of Incus outweigh accomplishing my goal by creating standalone hosts cluster or Proxmox cluster?

 

TL;DR - What are you running as a means of “antivirus” on Linux servers?

I have a few small Debian 12 servers running my services and would like to enhance my security posture. Some services are exposed to the internet and I’ve done quite a few things to protect the services and the hosts. When it comes to “antivirus”, I was looking at ClamAV as it seemed to be the most recommended. However, when I read the documentation, it stated that the recommended RAM was at least 2-4 gigs. Some of my servers have more power than other but some do not meet this requirement. The lower powered hosts are rpi3s and some Lenovo tinys.

When I searched for alternatives, I came across rkhunter and chrootkit, but they seem to no longer be maintained as their latest release was several years ago.

If possible, I’d like to run the same software across all my servers for simplicity and uniformity.

If you have a similar setup, what are you running? Any other recommendations?

P.S. if you are of the mindset that Linux doesn’t need this kind of protection then fine, that’s your belief, not mine. So please just skip this post.

 

TL;DR - had to rebuild my PhotoPrism database and now my originals count is off by ~5,000. Can I do a full sync of my devices and have it only upload what is missing?

Hello gurus,

I’ve been running Photoprism for quite some time and I’m happy with it.

I ran to an unrelated issue with my database (MariaDB) and has to rebuild the database. PhotoPrism uses this instance of MariaDB so naturally the metadata was gone.

The original pictures (originals) were stored in a separate array so at a minimum I still have all my pictures. I rebuilt the database and PhotoPrism (docker container) and pointed it to the array for the originals. Once that was done, I logged in to the PhotoPrism UI and perform a complete rescan and index of my originals. Once it was done, I noticed that my originals count was 27,000 but i should have 31,000 objects (according to a picture I took of the PhotoPrism UI I took the night before rebuilding the database). So I started digging a bit.

  • The array itself (where my originals are stored) is showing 27,000 objects.

  • The pictures I took the night before rebuilding the database and PhotoPrism containers said that the count of originals was ~31,000.

  • The two main devices backing media to PhotoPrism is my phone and my wife’s phone. My phone shows ~4,500 and my wife’s sores ~26,500.

  • Since these two phones are previously fully backups a few weeks before the rebuild I should have ~31,000 objects in the originals.

My question is, can I redo a full backups sync of both phones (through PhotoSync) and have it only copy the objects that are not in the originals?

Since the database has to be rebuilt, I fear that if I do another full sync, it will just copy everything again and I end up with ~60,000 objects rather than the ~31,000 I should have.

What can I do to see which objects are missing between my devices and PhotoPrism and how can I only copy those over to PhotoPrism?

 

For those of you running Unraid and backing up their data to Backblaze, how are you doing it?

I’ve been playing a bit with KopiaUI but what is the easiest and most straight forward way?

Bonus points if I can have the same “client/software/utility” backup data from non-servers (Windows, macOS and Linux) in my network to said Unraid server.

I don’t want to complicate the setup with a bunch of dependencies and things that would make the recovery process long and tedious or require in-depth knowledge to use it for the recovery process.

To recap:

Workstations/laptop > back up data with what? > Unraid

Unraid > back up data with what? > Backblaze

view more: next ›