Forked from mberman84/gist:9b3c281ae5e3e92b7e946f6a09787cde
Last active
March 7, 2024 21:15
-
-
Save sutr90/e8dba4b2c20049c1cd85cd29ce6bb1ca to your computer and use it in GitHub Desktop.
PrivateGPT Installation on WSL2
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
# Install CUDA runtime from https://docs.nvidia.com/cuda/wsl-user-guide/index.html | |
# Install pyenv from https://bgasparotto.com/install-pyenv-ubuntu-debian | |
# Clone the repo | |
git clone https://github.com/imartinez/privateGPT | |
cd privateGPT | |
# Install Python 3.11 | |
pyenv install 3.11 | |
pyenv local 3.11 | |
# Install Poetry | |
pip install poetry | |
# Install dependencies | |
poetry install --with ui,local | |
# Download Embedding and LLM models | |
poetry run python scripts/setup | |
# To Run on GPU (with CUDA 12) | |
CUDACXX=/usr/local/cuda-12/bin/nvcc CMAKE_ARGS="-DLLAMA_CUBLAS=on -DCMAKE_CUDA_ARCHITECTURES=native" FORCE_CMAKE=1 pip install llama-cpp-python --no-cache-dir --force-reinstall --upgrade | |
# Run the local server | |
PGPT_PROFILES=local make run | |
# Note: on Mac with Metal you should see a ggml_metal_add_buffer log, stating GPU is | |
being used | |
# Navigate to the UI and try it out! | |
http://localhost:8001/ |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment