A curated collection of interesting GitHub repositories
View the Project on GitHub tom-doerr/repo_posts
Run huge LLMs with 100k context on an 8GB GPU using SSD offload, no quantization