1
177

Due to the large number of reports we've received about recent posts, we've added Rule 7 stating "No low-effort posts. This is subjective and will largely be determined by the community member reports."

In general, we allow a post's fate to be determined by the amount of downvotes it receives. Sometimes, a post is so offensive to the community that removal seems appropriate. This new rule now allows such action to be taken.

We expect to fine-tune this approach as time goes on. Your patience is appreciated.

2
358
submitted 2 years ago* (last edited 2 years ago) by devve@lemmy.world to c/selfhosted@lemmy.world

Hello everyone! Mods here 😊

Tell us, what services do you selfhost? Extra points for selfhosted hardware infrastructure.

Feel free to take it as a chance to present yourself to the community!

🦎

3
29
4
50
submitted 23 hours ago* (last edited 7 hours ago) by kuvwert@lemmy.dbzer0.com to c/selfhosted@lemmy.world

Double Check Your NFS timeouts to your NAS arent an NFS problem. They might be a dirty page writeback problem.

I'm really Sorry in advance for the wall of text here. I debated trimming this down but honestly the whole reason I spent months stuck on this is because nothing about it was obvious. The symptoms point you at NFS, your mount options, your network, everything except whats actually wrong. And because the defaults that cause it ship with basically every linux distro, Id bet money theres a ton of people out there with the same problem right now just blaming thier NAS or Jellyfin or whatever. For all I know this is common knowledge and I'm just the last person to figure it out, but on the off chance somebody else is out there googling the same NFS timeout errors I was, heres the full story. (TL;DR Below)

Ive been chasing NFS issues on my Proxmox cluster for months now and I finally found the actual cause, and it wasnt anything Id seen anyone talk about online. Figured Id write it up because I guarantee other people are hitting this exact same wall.

The setup: half a dozen VMs on Proxmox, all mounting a Synology NAS over NFS. Jellyfin, Audiobookshelf, Sonarr, Radarr, the usual self-hosted media stack. Things would work fine for a while and then randomly go sideways. Jellyfin stops mid-playback. Audiobookshelf loses track of where you were. Sonarr tries to import a downloaded episode and the entire container locks up. dmesg fills with nfs: server 192.168.1.50 not responding, timed out and youre rebooting things again.

The part that kept me going in circles for so long is that it was never consistent. An audiobook would stream for hours without a hiccup, but then Sonarr would try to move a 4GB episode file and the whole mount would go down. I could ls the mount and browse around just fine even while Sonarr was hung. Small file operations worked. Large writes didnt. But not always, Sometimes a big import would go through without a problem, and Id convince myself whatever Id just changed in my mount options had fixed it.

I went through all the usual advice. Switched from NFSv4 to NFSv3, which I was especially convinced was the fix because the timing lined up with when Id been experimenting with v4. It wasnt. I toggled nolock, tuned rsize and wsize down from 128K to 32K, tried soft vs hard mounts, checked the Synologys HDD hibernation settings, disabled TCP offloading on the virtio NIC. Nothing actually fixed it. Every time I thought I had it, the next import that was over the threshold would fail and i would scream.

Then at one point I gave a couple of the VMs more RAM, thinking the media workloads could use the headroom. Everything got worse after that. Like, measurably worse. I didnt connect the two at the time.

What finally cracked it was running a dd test to write a 2GB file to the NFS mount and actually watching the numbers. With the 32K buffer mount options, the write reported 2.1 GB/s. On a gigabit link. Obviously that data is not going to the NAS. The kernel was eating the entire write into the VMs page cache, saying "yep, done!" and then trying to flush 2+ GB of dirty pages to the Synology all at once. The NAS gets hit with a wall of data it cant process fast enough, NFS RPC calls start timing out, and everything goes to hell.

The default value for vm.dirty_ratio is 20, meaning the kernel will let 20% of your RAM fill up with dirty pages before it forces a writeback. On my 13GB VM thats 2.6GB of buffered writes. So the kernel would happily sit there absorbing data into RAM, and then try to shove 2.6 gigs down a gigabit pipe to the NAS all at once. And when I "upgraded" VMs with more RAM, I was literally raising the ceiling on how big that buffer could get. Thats why things got worse. The inconsistency made sense too. A 700MB file might stay under the background flush threshold and trickle out fine. A 4GB season pack would blow past it and trigger the whole mess.

The fix

Two sysctl values:

sysctl -w vm.dirty_bytes=67108864
sysctl -w vm.dirty_background_bytes=33554432

This caps the dirty page buffer at 64MB and starts background writeback at 32MB. Instead of hoarding gigabytes and flushing all at once, the kernel now pushes data out to the NAS continuously in small batches. Make it persistent:

# For distros using /etc/sysctl.d/ (Debian 12+, Ubuntu, etc.)
echo -e 'vm.dirty_bytes=67108864\nvm.dirty_background_bytes=33554432' > /etc/sysctl.d/99-nfs-dirty-pages.conf
sysctl -p /etc/sysctl.d/99-nfs-dirty-pages.conf

# For distros using /etc/sysctl.conf
echo 'vm.dirty_bytes=67108864' >> /etc/sysctl.conf
echo 'vm.dirty_background_bytes=33554432' >> /etc/sysctl.conf

Before: 2GB dd writes at 101 MB/s, dies at the 2GB mark with NFS timeouts and I/O errors. After: same test, steady 11.4 MB/s start to finish, zero NFS timeouts, completes cleanly. OK oK Yeah, the throughput number is lower, but Ill take a transfer that actually finishes over one that crashes every time.

I applied this across all six of my VMs that mount the NAS and the whole fleet has been stable since. Theyd all been independently building up multi-gigabyte write backlogs and dumping them onto the Synology simultanously. I was basically DDoSing my own nas from six directions every time anything tried to write a big file.

Then I checked the Proxmox host itself. 128GB of RAM. Four NFS mounts to the same Synology, including the one Proxmox writes VM backups to. All hard mounts with default dirty ratio. Thats a 25GB dirty page ceiling on the hypervisor. Every scheduled backup was potentially building up a 25 gigabyte write buffer and then hosing the NAS with it in one shot. And because the mounts were hard, if the Synology choked during the flush, the hypervisor itself would hang, not just a VM. I dont even want to think about how many weird backup failures and unexplained freezes this was behind.

Since applying the fix Ive also noticed that Jellyfin library scans are completing reliably now. They used to hang constantly and Id just accepted that as normal Jellyfin-over-NFS jank. The scans were generating thumbnails and writing metadata, building up dirty pages, and triggering the same flush that would take down the mount mid-scan. Audiobookshelf was doing the same thing. It would scan libraries and randomly lose connection to the mounted paths. That one was harder to pin down because audiobook files and cover art are small enough that the writes wouldnt always push past the threshold on their own. But if another VM had already half-filled the NASs tolerance with its own flush, Audiobookshelf tipping it over would be enough. Same underlying bug in every case, and I spent months blaming three different applications for it.

If youre running a media stack on VMs with NFS mounts to a NAS and youve been tearing your hair out over random timeouts, check your vm.dirty_ratio and do the math against your RAM. Bet you its higher than you think.

TLDR; If your NFS mounts to a NAS randomly time out during large writes, your VMs are probably buffering gigabytes of dirty pages in RAM and then flushing them all at once, overwhelming the nas. Symptoms in my case were Jellyfin stopping mid-playback and hanging during library scans, Audiobookshelf losing connection to mounted paths and forgetting playback position, and Sonarr/Radarr locking up completely when trying to import episodes. Set vm.dirty_bytes=67108864 and vm.dirty_background_bytes=33554432 on every VM (and the hypervisor) to cap the buffer at 64MB and force continuous small writebacks instead.


Edit 1: @deadcade pointed out that 11.4 MB/s is suspiciously close to a 100 Mbps link ceiling and they were right. Checked the NAS LAN1 network status and it was negotiating at 100 Mbps... The NAS was plugged into my router which has gigabit ports but was apparently negotiating down due to what i must assume is an issue with the router.

SO the real solution: I went to Bestbuy and grabbed a $20 gigabit switch, plugged the NAS and Proxmox host into it directly, and the Synology came up at 1000 Mbps immediately. Same 2GB dd test now completes at 107 MB/s from the host and 115 MB/s from the VM, no timeouts, totally clean.

So if i actually understand wtf is going on here... it was actually two problems stacked on top of each other this entire time.

The 100 Mbps link was the speed ceiling between the router and the NAS. The dirty page defaults were what turned that speed limitation into a catastrophic failure. The kernel would buffer gigabytes of writes and then try to flush them through a 100 Mbps pipe where the NFS RPCs would time out long before the data finished arriving. The sysctl fix worked because it accidentally rate-limited the client to roughly what the 100 Mbps link could handle. Fixing the link speed solved the actual bottleneck.

THANKS for the insight deadcade!

Both fixes stay though. 64MB dirty page cap on a gigabit link still saturates the connection at 115 MB/s and there's no reason to let a 128GB Proxmox host build up a 25GB write buffer aimed at a consumer NAS. Also check your link speeds.

Edit 2: Thanks again to everyone who chimed in with your fantastic insights and ideas.

5
36

Probably a silly question but the .uk domain is really cheap. If I'm not in the UK can I still use that domain for my server without issue?

Its like 50 bucks for a ten year lease

6
112

I recently made a huge mistake. My self-hosted setup is more of a production environment than something I do for fun. The old Dell PowerEdge in my basement stores and serves tons of important data; or at least data that is important to me and my family. Documents, photos, movies, etc. It's all there in that big black box.

A few weeks ago, I decided to migrate from Hyper-V to Proxmox VE (PVE). Hyper-V Server 2019 is out of mainstream support and I'm trying to aggressively reduce my dependence on Microsoft. The migration was a little time consuming but overall went over without a hitch.

I had been using Veeam for backups but Veeam's Proxmox support is kind of "meh" and it made sense to move to Proxmox Backup Server (PBS) since I was already using their virtualization system. My server uses hardware raid and has two virtual disk arrays. One for VM virtual disk storage and one for backup storage. Previously, Veeam was dumping backups to the backup storage array and copying them to S3 storage offsite. I should note that storing backups on the same host being backed up is not advisable. However, sometimes you have to make compromises, especially if you want to keep costs down, and I figured that as long as I stayed on top of the offsite replications, I would be fine in the event of a major hardware failure.

With the migration to Proxmox, the plan was to offload the backups to a PBS physical server on-site which would then replicate those to another PBS host in the cloud. There were some problems with the new on-site PBS server which left me looking for a stop-gap solution.

Here's where the problems started. Proxmox VE can backup to storage without the need for PBS. I started doing that just so I had some sort of backups. I quickly learned that PBS can replicate storage from other PBS servers. It cannot, however, replicate storage from Proxmox VE. I thought, "Ok. I'll just spin up a PBS VM and dump backups to the backup disk array like I was doing with Veeam."

Hyper-V has a very straight forward process for giving VM's direct access to physical disks. It's doable in Proxmox VE (which is built on Debian) but less straight forward. I spun up my PBS VM, unmounted the backup disk array from the PVE host, and assigned it as mapped storage to the new PBS VM. ...or at least I thought that's what I did.

I got everything configured and started running local backups which ran like complete and utter shit. I thought, "Huh. That's strange. Oh well, it's temporary anyways." and went on with my day. About two days later, I go to access Paperless-ngx and it won't come up. I check the VM console. VM is frozen. I hard reset it aaaannnnddd now it won't boot. I start digging into it and find that the virtual HDD is corrupt. fsck is unable to repair it and I'm scratching my head trying to figure out what is going on.

I continued investigating until I noticed something. The physical disk id that's mapped to the PBS VM is the same as the id of the host VM storage disk. At that point, I realize just how fucked I actually am. The host server and the PBS VM have been trying to write to the same disk array for the better part of two days. There's a solid chance that the entire disk is corrupt and unrecoverable. VM data, backups, all of it. I'm sweating bullets because there are tons of important documents, pictures of my kids, and other stuff in there that I can't afford to lose.

Half a day working the physical disk over with various data recovery tools confirmed my worst fears: Everything on it is gone. Completely corrupted and unreadable.

Then I caught a break. After I initially unmounted the [correct] backup array from PVE it's just been sitting there untouched. Every once in a great while, my incompetence works out to my advantage I guess. All the backups that were created directly from PVE, without PBS, were still in tact. A few days old at this point but still way better than nothing. As I write this, I'm waiting on the last restore to finish. I managed to successfully restore all the other VM's.

What's really bad about this is I'm a veteran. I've been in IT in some form for almost 20 years. I know better. Making mistakes is OK and is just part of learning. You have to plan for the fact that you WILL make mistakes and systems WILL fail. If you don't, you might find yourself up shit creek without a paddle.

So what did I do wrong in this situation?

  • First, I failed to adequately plan ahead. I knew there were risks involved but I failed to appreciate the seriousness of those risks, much less mitigate them. What I should have done was go and buy a high capacity external drive, using it to make absolutely sure I had a known good backup of everything stored separately from my server. My inner cheapskate talked me out of it. That was a mistake.
  • Second, I failed to verify, verify, verify, and verify again that I was using the correct disk id. I already said this once but I'll repeat it: storing backups on the host being backed up is ill advised. In an enterprise environment, it would be completely unacceptable. With self-hosting, it's understandable, especially given that redundancy is expensive. If you are storing backups on the server being backed up, even if it's on removable storage, you need to make sure you have a redundant offsite backup and that it is fully functional.

Luck is not a disaster recovery plan. That was a close call for me. Way too close for my comfort.

7
-33

I Built a Python script that uses a local Ollama LLM to automatically find and add movies to Radarr.

It picks random films from your library, asks Ollama for similar suggestions based on theme and atmosphere, validates against OMDb, scores with plot embeddings, then adds the top results to Radarr automatically.

Examples:

  • Whiplash → La La Land, Birdman, All That Jazz
  • The Thing → In the Mouth of Madness, It Follows, The Descent
  • In Bruges → Seven Psychopaths, Dead Man's Shoes

Features:

  • 100% local, no external AI API
  • --auto mode for daily cron/Task Scheduler
  • --genre "Horror" for themed movie nights
  • Persistent blacklist, configurable quality profile
  • Works on Windows, Linux, Mac

GitHub: https://github.com/nikodindon/radarr-movie-recommender

8
32

Hi everyone,

I wanted to share an open-source project that I think the self-hosted community will appreciate, especially given the recent shift toward corporate-controlled data.

QST (Quiz/Survey/Test) is a complete, non-commercial assessment platform licensed under GPLv2. Unlike "free" cloud tools that harvest user data, QST is designed to be hosted on your own hardware (Windows, Linux, or Mac).

Why it’s relevant in 2026:

    Zero Data Leakage: Since it's self-hosted, student/taker data never touches a third-party server or AI training set.

    Scalability: It uses a multithreaded architecture that can handle thousands of users on modest hardware.

    Interoperability: Supports QTI, Moodle XML, and Word XML imports—no vendor lock-in for your question banks.

    No "Enshittification": It’s a standalone tool, not a "lite" version of a paid service.

It’s great for anyone who needs to run exams, certifications, or private surveys without the overhead of a massive LMS.

Source & Info: https://sourceforge.net/projects/qstonline/

Documentation: https://qstonline.org/

I'm happy to answer any questions about the setup or the Perl-based architecture!
9
827
submitted 3 days ago* (last edited 2 days ago) by Ek-Hou-Van-Braai@piefed.social to c/selfhosted@lemmy.world
10
56
submitted 2 days ago* (last edited 1 day ago) by Inkstainthebat@pawb.social to c/selfhosted@lemmy.world

I've been thinking about finally getting myself a proper domain for my server, but a friend told me that to get one I either need a VPS with a public ip (which just takes all the fun out of selfhosting) or purchase a static ip, which is beyond what I'm willing to spend for a hobby. Do I have any good options or should I just let it go?

Also, if this isn't the correct community for this, I'd appreciate being pointed to the right one, thank you

Update: after reading the comments the two main options I'm considering now are either a cheap VPS to use as proxy for my network via wireguard, or DynamicDNS. I'll see if I can figure out the rest from here, thank you!

11
21

I'm looking for a community management tool I could #selfhosting

Our Club try to move from Mail to messanger as communication tool.

Our userbase use Whatsapp, #Threema, #Signal
The usage is not regulated and not all boardmembers use all tools or wish to do that.

Could someone recommens selfhosting tools for that. I woukd host it on my #Docker
At moment I only have #Zammad on my list. But it could not use Threema and is not exactly for community management.

@selfhosted

12
90
submitted 3 days ago* (last edited 2 days ago) by KillianLarcher@lemmy.world to c/selfhosted@lemmy.world

Hi community,

I’m one of the maintainers of Portabase, and this is my first time sharing about it on Lemmy.

Portabase is an open-source platform for database backup and restore.

It’s designed to be simple, reliable, and lightweight, without exposing your databases to public networks. It works via a central server and edge agents (like Portainer), making it perfect for self-hosted or edge environments.

It currently supports 7 databases:

PostgreSQL, MariaDB, MySQL, SQLite, MongoDB, Redis and Valkey

Repository: https://github.com/Portabase/portabase

(we hit 500 stars recently!)

Key features:

  • Logical backups for PostgreSQL, MySQL, MariaDB, MongoDB, SQLite, Redis, Valkey
  • Multiple storage backends: local filesystem, S3, Cloudflare R2, Google Drive
  • Notifications via Discord, Telegram, Slack, webhooks, etc.
  • Cron-based scheduling with flexible retention strategies
  • Agent-based architecture for secure, edge-friendly deployments
  • Ready-to-use Docker Compose setup and Helm Chart

What’s coming next:

  • Increasing test coverage
  • Extending database support

I’d love to hear from you: which database would you like to see supported next in Portabase?

13
263
submitted 4 days ago* (last edited 4 days ago) by Tiritibambix@lemmy.ml to c/selfhosted@lemmy.world

I AM NOT THE DEV

Hi all,

There are plenty of CalDAV servers out there, but surprisingly very few good self-hosted web interfaces to actually access and manage your calendars.

For the past couple of years I've been following a project called Luna, and I think it deserves a bit more visibility here: https://github.com/Opisek/luna

It can pull calendars from multiple sources (CalDAV, Google, iCal links, etc.) and bring them together in a single interface.

For people like me with a constantly changing, busy, and somewhat chaotic schedule, having everything in one place like this is incredibly useful.

The project is still young, but it's progressing steadily, and the developer has been very patient and responsive whenever I've interacted with him.

Just wanted to share it here and show some support.

PS: I think the developer didn’t really expect anyone to promote the project or draw attention to it yet 🙂 Screenshots will probably come in time. It’s still a young project, so a little patience goes a long way.

14
120
submitted 3 days ago by Lem453@lemmy.ca to c/selfhosted@lemmy.world

This is a hugely requested feature for many years and a huge hole in my entire self hosted ecosystem. Every self-hosted app I have connects to my Authentik system for user management... Except home assistant. Arguably one of the apps I need it for the most for the whole family to use with their accounts.

Devs have been resistant for some reason.

There is now a community integratation that allows user management for HA to be via any openID backend (authentik, keycloak etc).

I've been running it for a few days and it works perfectly. Very easy to setup if you already have a working authentik setup and know how to use it with other apps like immich.

15
78
submitted 3 days ago by dont@lemmy.world to c/selfhosted@lemmy.world

🤣 sure, I'll use a reverse proxy / waf that has a release change log "I don't remember lol" (Yes, it's in alpha, but still...)

Is anyone here using it? Are you scared?

16
395

I mostly lurk here, and I know we've had this discussion come up a number of times since Discord's age verification changes were announced, but I figured this video offers value for the walkthrough and comparative analysis. Like me, the video authors aren't seasoned self-hosters, and I've still got a lot to learn. Stoat and Fluxer both look appealing to me for my needs, but Stoat seemingly needs self-hosted servers to route through their master server (unless I'm missing something stupid) and I replicated the 404 for Fluxer's self-hosting documentation seen in the video, so it's looking like I'm leaning toward a Matrix server of some kind. Hopefully everyone looking for the Discord exit ramp is closer to finding it after this video.

17
181
submitted 4 days ago* (last edited 3 days ago) by minoche@lemmy.world to c/selfhosted@lemmy.world

According to the official Discord, "ACX has made the decision to close Booklore and step away." Some contributors are working together on an unnamed replacement project.

For those not in the loop, Booklore was an app for selfhosting book libraries. It had a nice UI. It was able to store metadata separately from the download files, so you could have an organized library without duplication. In recent weeks, there have been conflicts about AI code, licensing, and general Discord nastiness.

RIP

Edit: The discord, website and github are all gone. I found a copy of the announcement:

Announcement📢 A note on where things stand

ACX has made the decision to close BookLore and step away. He has a partner, a new chapter of his life ahead of him, and honestly - building something that reached 10k stars and thousands of daily users is something to be proud of. We wish him well.

That said - this community, and this project, is bigger than any one person. That's the whole point of open source.

So here's what's happening next:

A group of the original contributors - the people who built a lot of what you've been using - are continuing the work under a new name. [PROJECT NAME TBD] is that continuation. Same mission. Better foundation. Governed the way an open source project should be: transparently, collaboratively, and with the community at the center.

We're not starting from zero. We're starting from everything this community has already built together.

If you want to be part of what comes next, come join us: 👉 https://discord.gg/FwqHeFWk

More details - name, repo, roadmap - coming very shortly. Thank you for your patience, and thank you for giving a damn about this project. That's exactly why it's worth continuing.

18
15

So, I have a VPS running some stuff, local proxmox-setup running something and then the 'normal' computers (laptop mostly) which I'd like to get a bit better backup solution for.

Proxmox VMs are taken care of by proxmox backup server and hetzner storagebox + nas at the separated garage, so they are decently protected against hardware failures. Workstations keep important files synced to nextcloud and the VPS has it's own nightly snapshots at the provider, so there's some reundancy in the system already. However, as we all know, two is one and one is none with backups, so I'd like to add a separate backup server in the mix.

As there's devices which are not online all the time I'm leaning towards an agent-based solution where devices push their data to the backup server instead of server pulling the data in. Also as I have some spare capacity I'd like to have an option to offer backup storage for friends as well where agent-based solution is a practically requirement.

But now the difficult thing is to decide software for it. Veeam offers something for hobbyists, but I'd rather have more open solution. Bacula seems promising, but based on a quick read it doesn't seem to be that simple to set up. Amanda looked good too, but that seems to be more or less abandoned project. Borg Backup would fill my own needs, but as friends tend to have either Windows or OSX that's not quite what I'm after.

Any suggestions on what route I should take?

19
76
submitted 4 days ago by gblues@lemmy.zip to c/selfhosted@lemmy.world

Hello everyone. Need some opinions here. Does it worth all the trouble to make things like jellyfin and immich run with HTTPS for services that are only accesible in the LAN? I ask it 'cause, as far as I know, there is no way to put a valid certificate like let's encrypt for a service that is not accessible from the net and I don't plan to buy a certificate for myself. But I have some trouble with the rest of my family having issue with their browsers complaining about the lack of https every time a browser is updated. So, what would be the best solution?

20
307
submitted 4 days ago* (last edited 4 days ago) by plankton@programming.dev to c/selfhosted@lemmy.world

is this Little Bobby Tables?

21
34
submitted 3 days ago by InsightSeeker to c/selfhosted@lemmy.world

Hey everyone,

Quick question out of curiosity.

I work as a manager in a consulting firm, and a lot of my day goes into communicating across platforms like Slack, WhatsApp, Teams, LinkedIn messages, etc. Switching between all of them sometimes feels a bit messy.

A couple of things I personally struggle with are important tasks getting buried in chats and constantly jumping between apps to keep up with conversations.

Would be great to hear how you handle this in your day-to-day work.

22
88

Hi there, I’m looking to get into self-hosting for privacy reasons and I wanted to ask y’all: how inadvisable is it to utilize an ISP-owned router/modem? I feel like they’re able to track everything I do online with their more than likely integrated spyware.

23
67

Hey selfhosters 👋

0.7 is out. Biggest release so far.

Quick context if you missed the previous posts: Ideon is a self-hosted visual workspace for developers. Infinite canvas, everything about a project in one place: notes, Git repos, tasks, files, links, kanbans.

The problem was never too many tools. It was that none of them talk to each other.

What's new:

  • Obsidian import: drop a vault into a project, notes become blocks on the canvas
  • Vercel integration: deploy and check deployment status without leaving the canvas
  • Excalidraw block: killed the old Sketch block, Excalidraw is just better
  • Folder block: group and collapse blocks, helps a lot on bigger projects
  • Drag and drop import: files, folders, text, Excalidraw sketches, drop anything on the canvas and Ideon picks the right block type
  • Vim mode for Note and Snippet blocks, opt-in from Account settings
  • Audio and video playback in File blocks

300+ stars now, more than I expected when I first posted here. Every issue, every comment pushed this forward, thank you.

Anyway. Worth a try if you haven't yet :)

GitHub: https://github.com/3xpyth0n/ideon

Docs: https://www.theideon.com/docs

24
12

Got a new PC handed down to me. And now have my old one collecting dust. It has a dedicated GPU (GTX 1060 6GB VRAM) i guess the most obvious thing would be an AI model or maybe jellyfin (which is currently running on a raspi 5 just fine for), but was wondering if you maybe had other suggestions?

25
51
submitted 4 days ago* (last edited 1 day ago) by N0cT4rtle@lemmy.zip to c/selfhosted@lemmy.world

Can you guys suggest some reliable and secure selfhosted IM service? I'm kinda in a very bad spot right now, so any centralized messaging wouldn't really work. And yeah, state sponsored mass surveillance is a question of concern. Sorry for odd phrasing, just really at a loss.

I heard of matrix, XMPP (heard good things about snikket.org), SimpleX and even some IRC wizardry over TOR. And I actually tried matrix (synapse server), but found it not reliable enough - sometimes skips a notification, periodic troubles with logging in, weird lack of voice calls on mobile client, and some other irritating, tiny hiccups. I'm open to any suggestion, really, even open to trying matrix once again. Just, please, describe why you think one option is better than the other.

And just FYI, use case is simply texting with friends and family, while avoiding state monitoring. Nothing nefarious

Update: Thank you, everyone! After some discussions with the lads, we decided to use XMPP (Prosody) as our main server, we also chose to use DeltaChat as a fallback. Both working great for a couple of days now. Again, thank you, everyone, for suggestions (and configs :) ), really helped me out

view more: next ›

Selfhosted

57687 readers
386 users here now

A place to share alternatives to popular online services that can be self-hosted without giving up privacy or locking you into a service you don't control.

Rules:

  1. Be civil: we're here to support and learn from one another. Insults won't be tolerated. Flame wars are frowned upon.

  2. No spam posting.

  3. Posts have to be centered around self-hosting. There are other communities for discussing hardware or home computing. If it's not obvious why your post topic revolves around selfhosting, please include details to make it clear.

  4. Don't duplicate the full text of your blog or github here. Just post the link for folks to click.

  5. Submission headline should match the article title (don’t cherry-pick information from the title to fit your agenda).

  6. No trolling.

  7. No low-effort posts. This is subjective and will largely be determined by the community member reports.

Resources:

Any issues on the community? Report it using the report flag.

Questions? DM the mods!

founded 2 years ago
MODERATORS