← Registry

gguf-quantization

Community

GGUF format and llama.cpp quantization for efficient CPU/GPU inference. Use when deploying models on consumer hardware, Apple Silicon, or when needing flexible quantization from 2-8 bit without GPU requirements.

Install

skillpm install gguf-quantization

Format score

85/100

Spec

v1.0

Installs

0

Published

April 1, 2026