Skip to main content

llama.cpp (GGUF and GGML)

AIKit utilizes and depends on llama.cpp, which provides inference of Meta's LLaMA model (and others) in pure C/C++, for the llama backend.

This is the default backend for aikit. No additional configuration is required.

This backend:

  • provides support for GGUF (recommended) and GGML models
  • supports both CPU (avx2, avx or fallback) and CUDA runtimes

Example

warning

Please make sure to change syntax to #syntax=ghcr.io/sozercan/aikit:latest in the examples below.

CPU

https://github.com/sozercan/aikit/blob/main/test/aikitfile-llama.yaml

GPU (CUDA)

https://github.com/sozercan/aikit/blob/main/test/aikitfile-llama-cuda.yaml