Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

One of the comments on reddit

> i have no clue why everyone isn't talking about this all the time.

Talking maybe, not doing it because it's still too out of reach for most of us? I've got a laptop with 32 GB RAM and a 2GB NVIDIA card. At which speed can I run that model, if I can run it at all?



2gb vram is not enough for running LLM but you can run on CPU gpt4all provides gui just click click download model click and you are ready, just give it a try

start with `mpt-chat7b`, my suggestion

https://gpt4all.io/index.html


Thank you for that, I’m installing now


2GB VRAM means you can run things comparable to GPT2, a glorified Markov chain. On your CPU you could run much larger models at far from real time speeds.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: