Skip to content

Latest commit

 

History

History
41 lines (24 loc) · 1.35 KB

llm-llama-cpp-windows.md

File metadata and controls

41 lines (24 loc) · 1.35 KB

Setup llm-inference using LLAMA-CPP

This is a short guide to setup llm-inference project to run on your Linux machine using llama-cpp-python package.

NOTE: Python 3.12 breaks torch instllation. Please use Python 3.10

Setup llm-inference using CPU on windows

This is a short guide to setup llm-inference project to run on your Windows machine using CPU.

  • Install Windows build tools from:
    https://visualstudio.microsoft.com/visual-cpp-build-tools/
    You are looking for Desktop Development with C++

  • Create Python Virtual Environment:

    • python -m venv venv
  • Activate the virtual environment:

    • .\venv\Scripts\activate
  • Install pytorch with cpu support:

  • Install required packages:

  • Install Bitsandbytes - windows compatible version

    • pip3 install git+https://github.com/Keith-Hon/bitsandbytes-windows.git
  • Create .env file based on .env.example or env-samples/env.llamacpp.example

    • Change the Model path and config then Run the server:
      • python main.py --multiprocess

Back to main doc