Note: I am not affiliated with this project in any way. I think it’s a very promising alternative to things like MinIO and deserves more attention.
Note: I am not affiliated with this project in any way. I think it’s a very promising alternative to things like MinIO and deserves more attention.
I remember when minio just started and it was small and easy to run. Nowadays, it’s a full-blown enterprise product, though, full of features you’ll never care about in a homelab eating on your cpu and ram.
Garage is small and easy to run. I’ve been toying with it for several months and I’m more than happy with its simple API and tiny footprint. I even run my (static html) blog off it because it’s just easier to deploy it to a S3-compatible API.
Minio now describes itself as “S3 & Kubernetes Native Object Storage for AI” - lol
Guess it’s time to look for alternatives if you’re not doing ML stuff
Of course it does AI now!
But seriously, the easiest guide for minio setup meant using their operator. The garage guide was: spin up this single deploy and it works from there.
S3 storage is simpler than local files? I think you need to elaborate
It is simpler when you’re doing stuff on the web and/or need to scale.
I disagree. Local files access is always superior. If you disagree, your target solution is likely poor to begin with
This is such a poor attempt at trolling. Don’t you have better things to do?
Bro, I’m an AWS Cloud Solution Architect and I seriously don’t know what you’re talking about. And, no, when I waste time on Lemmy, then there is literally nothing better to do.
AWS made S3. People built software to integrate S3 as a storage backend. Other people didn’t want to do AWS, and built single-node imitations of the S3 service. Now you use those services and think that is S3, while it is only a crude replica of what S3 really is. At this point the S3 API is redundant and you could just as well store your assets close to your application. You have no real, global S3 delivery service anyway. What’s the point?
Most people misuse AWS S3. Using stuff like minio is even more misguided.
S3 storage is simpler than running scp -r to a remote node, because you can copy files to S3 in a massively parallel way and scp is generally sequential. It’s very easy to protect the API too, as it’s just HTTP (and at it, it’s also significantly faster than WebDAV).
Nobody should be using SCP, use rsync.
S3 goes beyond the scope you describe. You disqualify yourself with such statements
Clearly I mean Garage in here when I write “S3.” It is significantly easier and faster to run
hugo deploy
and let it talk to Garage, then to figure out where on a remote node the nginx k8s pod has its data PV mounted and scp files into it. Yes, I could automate that. Yes, I could pin the blog’s pod to a single node. Yes, I could use a stable host path for that and use rsync, and I could skip the whole kubernetes insanity for a static html blog.But I somewhat enjoy poking the tech and yes, using Garage makes deploys faster and it provides me a stable well-known API endpoint for both data transfers and for serving the content, with very little maintenance required to make it work.
I don’t follow. S3 is an AWS service that these tools emulate locally by providing the same API. But I’m happy to accept that there’s just some misunderstanding 😃
In the context of my comments here, any mention of “S3” means “S3-compatible” in the way that’s implemented by Garage. I hope that clarifies it for you.