I'd say avoid pulling in all the python and containers required and just download the gguf from huggingface website directly in a browser rather than doing is programmatically. That sidesteps a lot of this project's complexity since nothing about llama.cpp requires those heavy deps or abstractions.
Yeah, I should have mentioned that in the post because I know that not everyone likes to tinker as much as I do.
I'll add a note about this site in a revision soon.
However, I still find it useful to know how to do that manually.