Self hosted instances - Lemmy.world

I am considering creating my own lemmy instance as my current instance (lemmy.world) is often down, and I enjoy this kind of a project. That being said I haven’t been able to find any resources on hw needed for this? If I have a raspberry pi 4b 8g with 128gb is that enough for just me using it? Do I need to pay for a cloud server to host it? How can federation work if I only have <10 users tops? Anything else I might need to know?

Consider object storage for pictures, that’s all I can say. That, and the community seeder bot is extremely helpful.
Is there a guide on setting up object storage?
It’s thru pictrs, but yes. This is a solid guide that explains how to do it out of the box without migration. literature.cafe/post/67882 <- (copy this link into the search of your instance, it’ll federate it over)
Pro-tip: Self-hosting Lemmy? You can use object storage to back pict-rs (image hosting) to save a lot of money - literature.cafe

Just thought I’d share this since it’s working for me at my home instance of federate.cc [http://federate.cc], even though it’s not documented in the Lemmy hosting guide. The image server used by Lemmy, pict-rs, recently added support for object storage like Amazon S3, instead of serving images directly off the disk. This is potentially interesting to you because object storage is orders of magnitude cheaper than disk storage with a VM. By way of example, I’m hosting my setup on Vultr, but this applies to say Digital Ocean or AWS as well. Going from a 50GB to a 100GB VM instance on Vultr will take you from $12 to $24/month. Up to 180GB, $48/month. Of course these include CPU and RAM step-ups too, but I’m focusing only on disk space for now. Vultr’s object storage by comparison is $5/month for 1TB of storage and includes a separate 1TB of bandwidth that doesn’t count against your main VM, plus this content is served off of Vultr’s CDN instead of your instance, meaning even less CPU load for you. This is pretty easy to do. What we’ll be doing is diverging slightly from the official Lemmy ansible setup [https://github.com/LemmyNet/lemmy-ansible] to add some different environment variables to pict-rs. After step 5, before running the ansible playbook, we’re going to modify the ansible template slightly: cd templates/ cp docker-compose.yml docker-compose.yml.original Now we’re going to edit the docker-compose.yml with your favourite text editor, personally I like micro but vim, emacs, nano or whatever will do… favourite-editor docker-compose.yml Down around line 67 begins the section for pictrs, you’ll notice under the environment section there are a bunch of things that the Lemmy guys predefined. We’re going to add some here to take advantage of the new support for object storage in pict-rs 0.4+ [https://git.asonix.dog/asonix/pict-rs/#user-content-filesystem-to-object-storage-migration]: At the bottom of the environment section we’ll add these new vars: - PICTRS__STORE__TYPE=object_storage - PICTRS__STORE__ENDPOINT=Your Object Store Endpoint - PICTRS__STORE__BUCKET_NAME=Your Bucket Name - PICTRS__STORE__REGION=Your Bucket Region - PICTRS__STORE__USE_PATH_STYLE=false - PICTRS__STORE__ACCESS_KEY=Your Access Key - PICTRS__STORE__SECRET_KEY=Your Secret Key So your whole pictrs section looks something like this: https://pastebin.com/X1dP1jew [https://pastebin.com/X1dP1jew] The actual bucket name, region, access key and secret key will come from your provider. If you’re using Vultr like me then they are under the details after you’ve created your object store, under Overview -> S3 Credentials. On Vultr your endpoint will be something like sjc1.vultrobjects.com [http://sjc1.vultrobjects.com], and your region is the domain prefix, so in this case sjc1. Now you can install as usual. If you have an existing instance already deployed, there is an additional migration command you have to run to move your on-disk images into the object storage. [https://git.asonix.dog/asonix/pict-rs/#filesystem-to-object-storage-migration] You’re now good to go and things should pretty much behave like before, except pict-rs will be saving images to your designated cloud/object store, and when serving images it will instead redirect clients to pull directly from the object store, saving you a lot of storage, cpu use and bandwidth, and therefore money. Hope this helps someone, I am not an expert in either Lemmy administration nor Linux sysadmin stuff, but I can say I’ve done this on my own instance at federate.cc [https://federate.cc] and so far I can’t see any ill effects. Happy Lemmy-ing!

Is this ok to do for an instance already running?

copying the link to the search bar just shows me this post in search. What is it supposed to do?

It should be? It’s supposed just show a link to this post -> lemmy.federate.cc/post/4824
Pro-tip: Self-hosting Lemmy? You can use object storage to back pict-rs (image hosting) to save a lot of money - federate.cc

Just thought I’d share this since it’s working for me at my home instance of federate.cc [http://federate.cc], even though it’s not documented in the Lemmy hosting guide. The image server used by Lemmy, pict-rs, recently added support for object storage like Amazon S3, instead of serving images directly off the disk. This is potentially interesting to you because object storage is orders of magnitude cheaper than disk storage with a VM. By way of example, I’m hosting my setup on Vultr, but this applies to say Digital Ocean or AWS as well. Going from a 50GB to a 100GB VM instance on Vultr will take you from $12 to $24/month. Up to 180GB, $48/month. Of course these include CPU and RAM step-ups too, but I’m focusing only on disk space for now. Vultr’s object storage by comparison is $5/month for 1TB of storage and includes a separate 1TB of bandwidth that doesn’t count against your main VM, plus this content is served off of Vultr’s CDN instead of your instance, meaning even less CPU load for you. This is pretty easy to do. What we’ll be doing is diverging slightly from the official Lemmy ansible setup [https://github.com/LemmyNet/lemmy-ansible] to add some different environment variables to pict-rs. After step 5, before running the ansible playbook, we’re going to modify the ansible template slightly: cd templates/ cp docker-compose.yml docker-compose.yml.original Now we’re going to edit the docker-compose.yml with your favourite text editor, personally I like micro but vim, emacs, nano or whatever will do… favourite-editor docker-compose.yml Down around line 67 begins the section for pictrs, you’ll notice under the environment section there are a bunch of things that the Lemmy guys predefined. We’re going to add some here to take advantage of the new support for object storage in pict-rs 0.4+ [https://git.asonix.dog/asonix/pict-rs/#user-content-filesystem-to-object-storage-migration]: At the bottom of the environment section we’ll add these new vars: - PICTRS__STORE__TYPE=object_storage - PICTRS__STORE__ENDPOINT=Your Object Store Endpoint - PICTRS__STORE__BUCKET_NAME=Your Bucket Name - PICTRS__STORE__REGION=Your Bucket Region - PICTRS__STORE__USE_PATH_STYLE=false - PICTRS__STORE__ACCESS_KEY=Your Access Key - PICTRS__STORE__SECRET_KEY=Your Secret Key So your whole pictrs section looks something like this: https://pastebin.com/X1dP1jew [https://pastebin.com/X1dP1jew] The actual bucket name, region, access key and secret key will come from your provider. If you’re using Vultr like me then they are under the details after you’ve created your object store, under Overview -> S3 Credentials. On Vultr your endpoint will be something like sjc1.vultrobjects.com [http://sjc1.vultrobjects.com], and your region is the domain prefix, so in this case sjc1. Now you can install as usual. If you have an existing instance already deployed, there is an additional migration command you have to run to move your on-disk images into the object storage. [https://git.asonix.dog/asonix/pict-rs/#filesystem-to-object-storage-migration] You’re now good to go and things should pretty much behave like before, except pict-rs will be saving images to your designated cloud/object store, and when serving images it will instead redirect clients to pull directly from the object store, saving you a lot of storage, cpu use and bandwidth, and therefore money. Hope this helps someone, I am not an expert in either Lemmy administration nor Linux sysadmin stuff, but I can say I’ve done this on my own instance at federate.cc [https://federate.cc] and so far I can’t see any ill effects. Happy Lemmy-ing!