![](https://lemmy.world/pictrs/image/d7844a9a-d604-4147-8867-fb19b2db2b70.jpeg)
![](https://lemmy.world/pictrs/image/4271bdc6-5114-4749-a5a9-afbc82a99c78.png)
I was in the same place as you a few years ago - I liked swarm, and was a bit intimidated by kubernetes - so I’d encourage you to take a stab at kubernetes. Everything you like about swam kubernetes does better, and tools like k3s make it super simple to get set up. There _is& a learning curve, but I’d say it’s worth it. Swarm is more or less a dead end tech at this point, and there are a lot more resources about kubernetes out there.
The early twenties intermediate dev on my team was explaining the other week that if you remember a time before smartphones and broadband, you are old
I personally am familiar with 2 organisations with millions of dollars in annual revenue that deploy critical line of business applications like this in 2024
Very NSFW
They are, but I think the question was more “does the increased speed of an SSD make a practical difference in user experience for immich specifically”
I suspect that the biggest difference would be running the Postgres DB on an SSD where the fast random access is going to make queries significantly faster (unless you have enough ram that Postgres can keep the entire DB in memory where it makes less of a difference).
Putting the actual image storage on SSD might improve latency slightly, but your hard drive is probably already faster than your internet connection so unless you’ve got lots of concurrent users or other things accessing the hard drive a bunch it’ll probably be fast enough.
These are all Reckons without data to back it up, so maybe do some testing
Debian. When I have time to mess about with server stuff, I want to be doing the thing I want to do rather than fixing whatever broke in the most recent set of updates
Sounds like a great idea - I suspect the biggest obstacle will be finding someone at the home who is confident enough in what to do with it to be willing to accept it.
I’ve run into similar issues with schools where they are hesitant to accept donations of things like that because they don’t want to be saddled with equipment they don’t know how to use and maintain. Maybe worth seeing if you can raise a bit of money for a second hand Xbox or something?
Pretty much - I try and time it so the dumps happen ~an hour before restic runs, but it’s not super critical
pg_dumpall
on a schedule, then restic to backup the dumps. I’m running Zalando Postgres in kubernetes so scheduled tasks and intercontainer networking is a bit simpler, but should be able to run a sidecar container in your compose file
If you figure it out, I know several companies that would be more than willing to drop 7 figures a year to license the tech from you
git-annex maybe?
We would like to apologise for any unpleasant experiences or doubts caused by the miscommunication of these terms
We are sorry we got called out
Yeah, they are mostly designed for classification and inference tasks; given a piece of input data, decide which of these categories it belongs to - the sort of things you are going to want to do in near real time, where it isn’t really practical to ship off to a data centre somewhere for processing.
Dealing with this at the moment - in an org that’s been pretty lax at writing anything down about what and why as far as internal software goes, trying (with support from C-suite) to get people to actually write up any amount of detail in their requests is like pulling teeth.
I tend to take that position as well; if it’s not defined, I get to define it. If I ask for feedback or review and get silence, that means you approve.
That’s not how Neon works. Your install will upgrade itself once the team have finished rebuilding everything on top of 24.04 - it’s happening, but it takes a bit of time
https://blog.neon.kde.org/2024/05/09/kde-neon-rebasing-on-ubuntu-noble/
It’s coming. It takes a bit of work to make it happen
Because accountants mostly.
For large businesses, you essentially have two ways to spend money:
This leaves companies in a slightly odd spot where from an accounting standpoint, it might look better on the books to spend $3 million/year on cloud stuff than $10 million every 5 years on servers
Seems pretty reasonable. At the end of the day people have to eat, so projects like this either trundle on as hobby-and-spare-time projects for a few years until people get bored and burnt out, or you find a way to make working on the project a paid gig for the core people
“is not exactly tailored to my specific requirements, aesthetic preferences and built using technology I’m familiar with” = “sucks” apparently