lyogavin/airllm
Run 70B LLMs on a 4GB GPU with layer-wise inference and memory optimization, quantization optional

View on index · View in 3D Map
// SURVEILLANCE FEED
Discovered repositories from the open source frontier
Run 70B LLMs on a 4GB GPU with layer-wise inference and memory optimization, quantization optional

View on index · View in 3D Map