Your laptop won’t just cry – it’ll beg for mercy before bursting into flames…
~5GB of free disk space isn’t enough even for just downloading Deepseek R1 8b.
As for the CPU and GPU specs, the LLM will have to be handled all by the CPU, which the Ryzen 3 3250U will struggle with.
But even with a much more powerful CPU, 6GB of RAM isn’t enough to hold Deepseek R1 8b (the “8B” is short for “8 billion tokens”) in memory while it runs without thrashing swap space.
The “1.5b” Deepseek R1 model would fit within your laptop’s hardware specs, but it’s not going to be fast.
Regarding the disk space thingy, the website lists the size for the 8B model as 5.2 GB. How’s it not enough?
For context, I have several gigs of free space. I just said I wanted a model around 5GB big, not “significantly” more, e.g., not models of size 8 GB, 10GB…
Is the installed size of the model bigger than the listed size of the model on the website?
Ah, 60GB of free space is fine then. The 8b model might not load depending on what the swap setup is.
For comparison, I’ve been running Google DeepMind’s 7b “Gemma” model on an AMD EPYC CPU (no GPU assist) with ~16x the multi-core performance of the Ryzen 3 3250U. It’s usable, but takes 6-10 seconds to answer “What is Linux?”.
I run a lot of AI models locally - doing everything from chatting with LLMs to producing images and music. My main rig for AI use has an i7 CPU, 32 GB RAM, and a 12 GB RTX 3000 series GPU. That makes everything run fast and smoothly.
My main workstation desktop has been a trusty old Optiplex with 16 GB of RAM and a modern i7 CPU. That latter computer will run AI models locally, but it’s not hugely fast. I don’t normally run anything over a 4B model with it. For your setup, I definitely would aim for an LLM that is listed as 1.5B (or no more than 3B) range. Your CPU will produce an answer very slowly, though I’m not sure you have enough RAM to run both your OS and an LLM.
If you do get it running, it helps if you tell it something like, “Limit your answer to 200 words” so that it gives you an answer that is produced more quickly.
For those who are new to running local LLMs, LM Studio is a bit more user friendly than ollama.