Here are the steps to get a model running on your not-so-powerful computer:
- Install llama.cpp (you can also build it from source with CMake):
brew install llama.cpp
other ways are available here
- Download the Gemma 3 model from unsloth (https://huggingface.co/unsloth). The 1-billion-parameter version should work with most CPUs, (faster with gpus in which case you need llama.cpp with gpu support):