- cross-posted to:
- nixos@lemmit.online
- cross-posted to:
- nixos@lemmit.online
Some folks on the internet were interested in how I had managed to ditch Docker for local development. This is a slightly overdue write up on how I typically do things now with Nix, Overmind and Just.
You might be interested in this article that compares nix and docker. It explains why docker builds are not considered reproducible:
and why nix builds are reproducible a lot of the time:
Containerization has other advantages though (security) and you can actually use nix’s reproducible builds in combination with (docker) containers.
That seems like an argument for maintaining a frozen repo of packages, not against containers. You can only have a truly fully-reproducible build environment if you setup your toolchain to keep copies of every piece of external software so that you can do hermetic builds.
I think this is a misguided way to workaround proper toolchain setup. Nix is pretty cool though.
I am not arguing against containers, I am arguing that nix is more reproducible. Containers can be used with nix and are useful in other ways.
This is essentially what nix does. In addition it verifies that the packages are identical to the packages specified in your flake.nix file.
This is essentially what Nix does, except Nix verifies the external software is the same with checksums. It also does hermetic builds.
Nix is indeed cool. I just see it as less practical than maintaining a toolchain for devs to use. Seems like reinventing the wheel, instead of airing-up the tires. I could well be absolutely wrong there - my experience is mainly enterprise software and not every process or tool there is used because it is the best one.
@nickwitha_k @uthredii I’d like to think a better analogy would be that nix is like using a 3D model of a wheel instead of a compass and a straightedge to make wheels hehe 🙃
I quite like the sound of Nix, every time I touch on it but haven’t really dug in yet. You’re making me really want to though.
There are definately some things preventing Nix adoption. What are the reasons you see it as less practical than the alternatives?
What are alternative ways of maintaining a toolchain that achieves the same thing?
I see it as less practical mainly due to the extant tooling and age/maturity of the project.
The ways that I’m most familiar with are use of software like Artifactory - basically a multi-repo. Using such a tool, any package or artifact can be readily retained for future use. Then, for builds, one only needs to ensure that it is used as the package source, regardless of type (PyPy, Docker image, binary, RPM, etc).
Alternatively, one can use individual repos for any relevant package type but that’s a bit more overhead to manage.
I’ll certainly give this a read!
Are you saying that nix will cache all the dependencies within itself/its “container,” or whatever its container replacement would be called?
Yep, sort of.
It saves each version of your dependencies to the /nix/store folder with a checksum prefixing the program name. For example you might have the following Firefox programs
Because of this you can largely avoid dependency conflicts. For example a program A could depend on
/nix/store/cm1bdi4hp8g8ic5jxqjhzmm7gl3a6c46-firefox-108.0.1
and a program B could depend on/nix/store/rfr0n62z21ymi0ljj04qw2d7fgy2ckrq-firefox-114.0.1
and both programs would work as both have dependencies satisfied. AFAIK using other build systems you would have to break program A or program B (or find versions of program A and program B where both dependencies are satisfied).