Quickstart#
This package targets Python 3.12+ and follows a standard src/ layout.
See Compatibility And Start Here for the tested package combination and install-path
guidance.
Local development setup:
python -m venv .venv
source .venv/bin/activate
make dev
make test
Run the bundled example:
python -m pip install "llama-cpp-python[server]" huggingface-hub
make run-example
make examples-test
make run-example executes the canonical umbrella-level walkthrough that
uses a real packaged problem, a live model-backed workflow agent, canonical
experiment artifacts, and exported event-table validation.
The walkthrough uses a managed llama.cpp client. Install
llama-cpp-python[server] before running it. If you want the default GGUF
download path to work, also install huggingface-hub; otherwise set
LLAMA_CPP_MODEL to a specific local GGUF file.
make examples-test stays deterministic and offline-first by default. It
exercises the two smaller recipe-first examples and skips the live
walkthrough unless RUN_LIVE_EXAMPLE=1.
Build the docs:
make docs