Runs LLM inference on CPU, Apple Silicon, and consumer GPUs without NVIDIA hardware. Use for edge deployment, M1/M2/M3 Macs, AMD/Intel GPUs, or when CUDA is unavailable. Supports GGUF quantization (1.5-8 bit) for reduced memory and 4-10x speedup vs PyTorch on CPU.
/plugin marketplace add zechenzhangAGI/AI-research-SKILLs/plugin install llama-cpp@zechenzhangAGI/AI-research-SKILLs