I just stop my containers and tar gzip their compose files, their volumes and the /etc folder on the host
I like sysadmin, scripting, manga and football.
I just stop my containers and tar gzip their compose files, their volumes and the /etc folder on the host
Assuming you have all of them under a folder, I just run this lol
for f in *; do
echo "$f";
git -C "$f" pull;
git -C "$f" submodule update --recursive --remote;
echo "";
echo "#########################################################################";
echo "";
done
From what I understand its still restrained by the server in real time so downloading a 2 hour movie would still take two hours 🥲
Yeah, I have been eyeing upgrades to get avx512 anyway because lately I have been doing very heave very low preset av1 encodes but when they are a dick about it I just feel like postponing it.
lol for the past 15 years I have “rebuilt” my desktop every 5 years but I didn’t expect the would try to force me out of my 7 3700x right on the date
Your connection is 40MB/s I assume
5e is capable of getting the full 1Gbps of my connection so I easily see over 90MB/s. That being said I bought a big 100m bulk years ago and have been clipping it myself with care.
If you were indeed using leftover/ free cables of cuestionable quality it indeed could be a reason for poor perfomance
I read about the drop of the microservices container but not about the env changes. Thankfully I had not set them so defaults worked out lol
Is it for security? I think is mostly recommended because your home router is likely to have a dynamic address.
This is in regards to opening a port for WG vs a tunnel to a VPS. Of course directly exposing nginx on your router is bad.
remnant, partially because it’s a frankestein of second hand from wallapop and dusted pieces from my old computers, partially as a weeb reference to the world of RWBY lol
lol same I like to know exactly where the data is
I do this always for any service but also do a dump of the db in the top directory to keep a clean copy to could be version independent. They wont take much more space ho estly
It doesn’t cover permissions unless you are willing to setup http auth on your webserver but I really enjoy mdbooks. I looks clean and still is just markdown.
I guess that’s fair for single service composes but I don’t really trust composes with multiple services to gracefully handle only recreating one of the containers
I do it that way. Enable email notifications for new tagged releases, something arrives, check changelog, everything fine?
docker-compose pull; docker-compose down; docker-compose up -d
And we are done
You could not use the dns challenge?
I have no experience outside of blocky, but the configuration file is so damm simple and clean I have troubles even considering anything else.
Daily, usually keeping only the last week or so
If you keep the same filenames for the video files it should not redownload what already has.
For automatically I think is honestly easier to just run the command on a cronjob every 5 mins.
Just buy bigger disks 🫢
Arch Linu, Kubuntu, Supergrub, Tails, Kali, Windows