What’s up, what’s down and what are you not sure about?
Let us know what you set up lately, what kind of problems you currently think about or are running into, what new device you added to your homelab or what interesting service or article you found.
Personally I’m finally reaping the fruits of my labour and enjoy my stable homelab without doing much. One node went down recently and the other took over until I restarted so I was not in a hurry to fix things. Enjoying family time and only running updates that aren’t automated (yet). I’m about to dig a bit deeper into logging, probably setting up central log collection like Loki at some point, but not yet.
Realized today that borgbackup failed for almost 2 months straight on one of my servers (was a simple case of a lock being stuck). Finally setup push notifications via Pushover to notify on success/fail.
This is worth it. Had this happen on OS backup. Lost my data. Notifs should be default.
Healthchecks is incredibly nice for this kind of thing, it’ll notify you if it doesn’t receive a ‘success’ ping on whatever interval you specify.
I use it for all my Restic backups.
Same. I’d rather be alerted because something expected didn’t happen, not silence because something failed so hard it didn’t even send an alert.
Yeah that sounds even better. What service do you use?
Backblaze B2 for storage, and I host Healthchecks myself at home.
Found out Ghost 6.0 is out today and now it supports ActivityPub. It’s time to set up a new blog I’ll never write once more!
Oh exciting, finally!
Another glorious day of not having to worry about my nice and stable Debian server. It runs on an old Dell thin client I got on ebay, which isn’t much, but it gets the job done.
I wanna get into it but man, the mountain of knowledge I need to even understand what people are talking about is hard to climb. I’m trying to just get some stuff running in docker and it fails to launch and I’m like… How?! Isn’t that the whole point of docker lol. Baby steps I guess
I felt exactly the same when i started - the learning curve is real! Try TrueCharts.org or linuxserver.io for reliable docker templates with good docs that actually work, saved me so much troubleshooting headache.
Thanks will do!
I’ve learnt it from scratch in my week off, spending 2 or 3 hours on it every night for a week (although this might be underselling it as I had become familiar with desktop Linux over the past year and had a superficial idea of Docker containers with my Synology NAS). But still it’s not as big a deal as you think once you find some good resources. I’m going to comment about my setup after this in this thread… Have a look.
Main resource that helped me was Marius Hosting and ChatGPT got me out of trouble when I got stuck by deciphering logs for me when things didn’t work.
Thanks. Yeah I’m just trying to work at it slowly in my downtime instead of just watching YouTube all night.
It’s messy. Dockers superpower: You can write a crazy ass python application that needs dozens of dependencies and weird software configured. You put it into a container, you can update and publish the container with a single script call. Other people can install that, set some variables and not have to install the dozens of other pieces of software. They also don’t have to worry about updates.
But that’s not to say you don’t have to worry about networks, storage and ports.
Then the simplicity of the configuration of containers depends upon the person that made the container. Maybe they wanted to be very flexible and there are dozens of things you need to set. Maybe they didn’t include the data store internally in that container and you need your own data store in another container.
Check out Cosmos, I struggled piecing things together but when I restarted from scratch with this as the base is has been SO much easier to get services working, while still being able to see how things work under the hood.
It’s basically a docker manager with integrated reverse proxy and OpenID SSO capability, with optional VPN and storage management
Im at the level where I don’t know what SSO means. I can follow instructions to change a DNS. But what a DNS actually is I don’t know. Which is fine, until I need to work out what’s broken
SSO is “single sign on”. DNS is “domain name service”, which is just a way to turn a hostname (like www.google.com) into an IP address. It’s sort of like a phone directory, but for the Internet.
SSO is single sign on, so you don’t need individual username and password for every service. It’s a bit more advanced so don’t worry about it until you have what you want working properly for a while.
DNS is like the yellow pages of the internet - when you type www.google.com your computer uses a DNS server to look up what actual IP address corresponds to the website name. The point of Adguard or pihole is that when a website tries to load an ad your custom DNS server just says it doesn’t recognize the address
Oh like a custom yellowpages, sick!
Docker should be trivial to run. Hopefully it gives you some useful messages in the logs.
Realized last week that my fail2ban settings are too strict – I get banned immediately if I visit my funkwhale (music server) domain without being logged in. In fact, I think much of my “downtime” might have actually just been me banning myself for 15 minutes now and then…
I was thinking about getting rid of Grafana, which is overkill for my server, and replacing it with Logdy this weekend, but didn’t get around to it.
I installed a new server at home and went with NixOS. It looks super cool but it takes so much time to learn everything. The only thing keeping me from going back to Debian is how easy it was to permanently mount drives (and save a configuration for any future install or mishaps).
(I.e. mount,
nixos-generate-config
,nixos-rebuild switch
and done!)NixOS […] learn everything
I don’t think it’s possible to learn everything for NixOS as a casual user / admin. It’s massive. I was luckily able to sneak a NixOS project into work which gave me some paid time on the topic. But there’s always room to learn more about it. Which is a good thing - by its nature, it’s just more powerful than conventional distributions.
More powerful = more mental burden and capacity used to know how to run and manage its unique syntax and structure.
Sincerely NIX user daily. Switching away from nix and off to fedora kinoite.
I recently started setting up home server on Raspberry Pi 5. Having issues with raid1. I have 2 nvme PCIe gen 4 SSDs. There was power outages while writing. Now second disk keeps randomly falling. Though, I’m not sure if that’s the reason because I don’t know what was raid status before outage, also disk passes checks. First time it degraded, it tried to recover and it failed. I removed that disk from raid, recreated partition run some test using nvme-cli. Disk looked healthy. I re-added disk, rebuild started and completed successfully. Then I’ve written around 500Gbs of data and it degraded again. At that point I took a break.
There are two things I’m yet to try:
- Change configs to use gen 2 PCI, currently it’s set to gen 3, but AFAIK pi 5 does not support gen 3 officially.
- Remove, format and write data to problematic disk directly. I hope this will give me an idea is this hardware issue or software issue
I’m frustrated and will appreciate any hints.
Getting ready to move from out of the woods and back to civilization with my partner.
Not looking forward to having neighbors above or below me but I’m very excited to have internet that doesnt fucking suck.
Once were moved and a bit more settled, I’m gonna start really digging into to selfhosting things. I have the hardware, a couple HP mini PCs that will run home assistant and probably a server for various docker things. Nextcloud and immich seem to be the things I’ve found i wanna use so far. I already have a NAS set up, but was having am issue with it not booting if a monitor isnt plugged in. I bought a dummy plug for it but haven’t tried it out yet.
Will also be setting up an AI server for local LLM use. Hope to train one to fit my needs once I pull the trigger on 3060 12GB card but need to figure out what other parts I’ll use. Might upgrade my main rig and use the parts from that, or maybe I’ll buy a old dell and fix it up. Not sure yet.
Lots of ideas, so little time lol.
Might want a bigger GPU, I have a 3080ti and the 12gb is pretty limiting in terms of how large a model you can use, or like one thing I was hoping to do was essentially replace Google Assistant/Gemini and can’t realistically run a good model and the STT/TTS off the one gpu.
Let us know what you set up lately, what kind of problems you currently think about or are running into
I noticed that my link collector nears perfection (for my use case) - not much stuff required to be done lately. Which is a good thing.
Everything is just peachy this week except that I’m still trying to sort out why my I’m unable to access the internet when I’m connected to my unraid wireguard instance.
I am also finally ready to ditch my plex instance, too. Got some self-inflicted permissions issues sorted and it’s been smooth sailing for long enough that I’m ready to make the switch
Finally retired proxmox (actually I just removed pve packages and repos). Left the nfs export on there and hardened the whole thing.
Now I’m slowly working to get all my installs into layered ansible playbooks. Fortunately, there exists an incus ansible module.
With separate, mounted, persistent data, it’s getting very close to docker in easy deployment.
I’ve set up Pangolin on my VPS and had no problems accessing docker services on my homelab remotely. However, I don’t know how I am supposed to SSH or SFTP to my homelab. Will I connect to my VPS instead? Would I need to break Pangolin or expose a vulnerability to do so?
Honestly I am in need of a proper networking tutorial at this point.
According to the Pangolin docs it supports raw TCP and UDP connections.
For SSH you can also try to use the VPS as a jump host like this:
$ ssh user@vpn-homelab-ip -J user@vps-ip
I would never have found this on my own otherwise. I feel any amount of gratitude would fall short of compensating for how much time and effort it has saved me. Thank you regardless.
If possible, can you share how I can achieve the same effect with SFTP?
Either use the
sftp
command, it also supports the-J
option, or use SSH tunneling. For example here I bind the homelab port 4533 to my local port 8080.$ ssh -L 8080:vpn-homelab-ip:4533 user@vps-ip (user@vps) $
I can now open a new shell and run:
$ curl http://localhost:8080/ <a href="/app/">Found</a>.
You could also do it this way:
$ ssh -L 8080:localhost:4533 user@vpn-homelab-ip -J user@vps-ip (user@homelab) $
Thanks a ton!
I looked into VyOS to replace my main firewall/IPS system (IPfire) with, as I would like to switch to running it in a VM, which is not recommended with IPfire. Seems pretty good so far with the new gratis semi-stable Stream releases.
And I set up Unified Push notifications with my Ejabberd server. Works great.
Opnsense is also great, and has a webUI for easier setup.
That is what I started with originally, and I don’t want to go back. The WebUI is super convoluted and for anything other than the basics it does more harm than good in my experience… and well, FreeBSD is just not my thing.
Any particular reason you are looking for a virtualized VM? Just to be less reliant on a single piece of hardware?
Does anyone know how to get a static IP for their server when their ISP doesn’t allow it. I’ve found out how to use duckdns, but I want to set up my own DNS server from anywhere but I’m pretty sure it requires using a static IP.
I use duckdns, and thus have a xyz.duckdns.org domain, that points to the dynamic ipv4 address of my server. I do not host my own DNS server, rather I rely on a cheap Website / Mail / domain bundle. There I can enter my duckdns domain as a CNAME DNS entry. Thus every DNS lookup that is not for the remote hosted Website will resolve the duckdns domain and finally end at my server.
I am not sure where you want to host your DNS server or also for what specific reason… If you don’t have a domain, you kind of don’t need to host a DNS, and every domain provider I had, also offered a DNS server with it.
I wanted to run a pi hole to use as a DNS so that I can be ad free on any device. The problem is that with my computer or with my phone, I need to put in a specific IP address when I want to change DNS on that device.
So I wouldn’t put Pihole on the internet, but instead set up a Wireguard VPN on your devices and access Pihole via that.
Then you can use the dynamic DNS hostname for Wireguard, and a direct IP for Pihole.
Alternatively you could run Adguard Home instead, as it supports being a DoT and DoH server, both of which work over a hostname on your devices (ie; Android uses DoT for its secure DNS option).
Ooh wireguard sounds like a great option
If you drop the “from anywhere” part, you can set up a pihole with a static address that you can use from within your LAN, without any involvement from your ISP.
Read section “Assign your Raspberry Pi a static IP address” of https://www.raspberrypi.com/tutorials/running-pi-hole-on-a-raspberry-pi/
Yes exactly, additionally you probably don’t want to host your pi hole for external use (mobile phone or laptop in a different network) for the reason of latency.
The delay that is imposed by visiting your pi hole at home for each DNS request is going to be very unpleasant.
Rather rely on an external dns provider that provides pi hole like funticionality.
But this does not mean that you can’t also host your pihole for internal use. I use it not just for removing ads, but also to allow the access of local domains.
Honestly I never thought about the latency issue, but I probably won’t do it because of that now that you mention it. Much appreciated.
Started looking at Gemini-protocol over the weekend. (It’s like a newer version of gopher) Now I’m looking for a problem to fit the tool.
I started writing a science fiction, choose your own adventure, short story to fit the platform But that’ll take ages to finish.
I’m also eyeing a meshtastic client proxy. But you only get about 200 bytes per message so I’m not entirely sure it’s worth it.
The last thing, it would be kind of cool is a Zim tie-in. It would be cool to have a canned Wikipedia that could be accessed via Gemini protocol.