gentoo.mahdi.cz

  

dev-python:airllm


AirLLM allows single 4GB GPU card to run 70B large language models without quantization, distillation or pruning.
https://github.com/lyogavin/Anima/tree/main/air_llm

airllm-2.8.3
download~amd64 ~x86pypi