llama-cpp | skill guide | OpenClaw Study

Run LLM inference with llama.cpp on CPU, Apple Silicon, AMD/Intel GPUs, or NVIDIA — plus GGUF model conversion and quantization (2–8 bit with K-quants and…

Run LLM inference with llama.cpp on CPU, Apple Silicon, AMD/Intel GPUs, or NVIDIA — plus GGUF model conversion and quantization (2–8 bit with K-quants and im...

This page belongs to the OpenClaw Skills learning hub with install guides, category navigation, and practical links.

简体中文 繁體中文 日本語 Español Português