I’m not sure which German you expect to have conducted this research other than a former Nazi, seeing as basically everyone left was a former Nazi. And I don’t see how you can just dismiss the government report as Nazi lies when even the Soviets report a figure of 350 thousand dead.
Some editors can embed neovim, for example: vscode-neovim. Not sure how well that works though as I never tried it.
Well personally if a package is not on aur I first check if there’s an appimage available, or if there’s a flatpak. If neither exist, I generally make a package for myself.
It sounds intimidating, but for most software the package description is just gonna be a single file of maybe 10-15 lines. It’s a useful skill to learn and there’s lots of tutorials explaining how to get into it, as well as the arch wiki serving as documentation. Not to mention, every aur or arch package can be looked at as an example, just click the “view PKGBUILD” link on the side on the package view. You can even simply download an existing package with git clone and just change some bits.
Alternatively you can just make it locally and use it like that, i.e. just run make without install.
Aur and pacman are 90% of why I use arch.
Also fyi to OP: never install software system-wide without your package manager. No sudo make install
, no curl .. | sudo bash
or whatever the readme calls for. Not because it’s unsafe, but because eventually you’re likely to end up with a broken system, and then you’ll blame your distro for it, or just Linux in general.
My desktop install is about a decade old now, and never broke because I only ever use the package manager.
Of course in your home folder anything goes.
I think they meant you don’t know what the binary is called because it doesn’t match the package name. I usually list the package files to see what it put in /use/bin
in such cases.
What’s up with the abuse of the word open lately. I had a look at that project to see how they were doing the conversion, but I couldn’t find it. But I found this:
Short answer, yes! OpenScanCloud (OSC) is and will stay closed source…
Your data will be transferred through Dropbox and stored/processed on my local servers. I will use those image sets and resulting 3d models for further research, but none of your data will be published without your explicit consent!
I feel like I’d rather use Autodesk at that point. At least I know what I’m dealing with right out of the gate.
But check that it has all the features you need because it lags behind gitea in some aspects (like ci).
I mean I learned it in a few days and found it very intuitive as well. Far more intuitive than I found fusion when I tried that years later. Inventor and onshape also feel more pleasant to use.
The issue seems to be that the fusion interface is very non-standard when compared to other cad suites, so people that get used to it first find everything else unintuitive.
Podman quadlets have been a blessing. They basically let you manage containers as if they were simple services. You just plop a container unit file in /etc/containers/systemd/
, daemon-reload and presto, you’ve got a service that other containers or services can depend on.
I’ve been in love with the concept of ansible since I discovered it almost a decade ago, but I still hate how verbose it is, and how cumbersome the yaml based DSL is. You can have a role that basically does the job of 3 lines of bash and it’ll need 3 yaml files in 4 directories.
About 3 years ago I wrote a big ansible playbook that would fully configure my home server, desktop and laptop from a minimal arch install. Then I used said playbook for my laptop and server.
I just got a new laptop and went to look at the playbook but realised it probably needs to be updated in a few places. I got feelings of dread thinking about reading all that yaml and updating it.
So instead I’m just gonna rewrite everything in simple python with a few helper functions. The few roles I rewrote are already so much cleaner and shorter. Should be way faster and more user friendly and maintainable.
I’ll keep ansible for actual deployments.
Not sure what you’re on about, most package managers have a literal database of most package manager installed files. Debian and derivatives have dpkg --verify
or debsums
to verify the files, arch has paccheck
, I’m sure other distros have something similar. And fixing them is just a matter of reinstalling the package, which you can do from a chroot if the system won’t boot.
Or you can just run your system on a checksumming FS like btrfs which will instantly tell you when a file goes bad.
Someone found a way to weaponise bikeshedding.
Everyone just confirming aliteral’s point.
I guess it’s too much to ask the richest company on the planet to keep a list of a few accounts indefinitely. I’m sure that database is a whole gigabyte sized and maintaining it requires a whole person to check in on it once in a while. Obviously they can only afford that level of effort for a year or two. And we’re only taking about removing access from millions of people to something they paid good money for, and also doing it because. Yeah, I’m with you on this one, totally not their fault.
All public companies are, it’s just what Boeing makes things that fall out of the sky if they mess up, so it’s more obvious.
Just have NAS A send a rocket with the data to NAS B.
I remember the clusterfuck that existed before systemd, so I love systemd.
If this was done by multiple people, I’m sure the person that designed this delivery mechanism is really annoyed with the person that made the sloppy payload, since that made it all get detected right away.
Quite literally the first paragraph of the article:
Or in more detail lower down in the section titled Soviet statistics: