Repository Showcase

A curated collection of interesting GitHub repositories

View the Project on GitHub tom-doerr/repo_posts

lyogavin/airllm

Run 70B LLMs on a 4GB GPU with layer-wise inference and memory optimization, quantization optional