this post was submitted on 13 Feb 2025
7 points (88.9% liked)
LocalLLaMA
2585 readers
7 users here now
Community to discuss about LLaMA, the large language model created by Meta AI.
This is intended to be a replacement for r/LocalLLaMA on Reddit.
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Mistral
I personally run models on my laptop. I have 48 GB of ram and a i5-12500U. It runs a little slow but usable
My gear is an old:
I7-4790 16GB RAM
How many tokens by second?
The biggest bottleneck is going to be memory. I would just stick with GPU only since your GPU memory has the most bandwidth.