gentoo.mahdi.cz
dev-python
:airllm
AirLLM allows single 4GB GPU card to run 70B large language models without quantization, distillation or pruning.
https://github.com/lyogavin/Anima/tree/main/air_llm
airllm-2.8.3
download
~amd64 ~x86
pypi