Skip to main content

Backends

llama-node currently supports llm-rs, llama.cpp and rwkv.cpp backends.

llm-rs can supported multiple inference at same time.

llama.cpp and rwkv.cpp will treat async inference (in concurrent) as sequential requests.


  • To use llama.cpp backend, run
npm install @llama-node/llama-cpp
  • To use llm-rs backend, run
npm install @llama-node/core
  • To use rwkv.cpp backend, run
npm install @llama-node/rwkv-cpp