Comment Re:Open Source Model? (Score 2) 20
You can run it just fine if you have enough RAM. In this case that would be a 256 GiB system, of witch 231 GiB will be used to run the model. You don't need a GPU, but you need fast memory. ollama generates about one token (1 word +/-) for each time is scans the entire model's memory. It's more bound to memory performance than to CPU / GPU performance. GPUs have much faster memory, but less memory typically, but server processors with fast DDR5 and 8 or 12 channels should work just fine also. For smaller models you need less RAM, of course. There are 40 GiB and 4.7 GiB versions which you can run on your laptop quite easily, if you have 64 GiB or even 8 GiB RAM.