I created this resource to help me quickly see which models I can run on certain VRAM constraints.
Check it out here: https://imraf.github.io/ai-model-reference/
I’d like this to be as comprehensive as possible. It’s on GitHub and contributions are welcome!
💬 Discussion r/LocalLLaMA (151 points, 41 commentaires) 🔗 Source