LocalAI

LocalAI is the free, Open Source OpenAI alternative. LocalAI act as a drop-in replacement REST API that’s compatible with OpenAI API specifications for local inferencing. It allows you to run LLMs, generate images, audio (and not only) locally or on-prem with consumer grade hardware, supporting multiple model families and architectures.

Refer to the LocalAI documentation to install & configure it.

Once you have the LocalAI up & running, you'll need to configure the unCtl:

llm_config:
  - provider: LocalAI
    models:
      - name: thebloke__codellama-13b-gguf__codellama-13b.q6_k.gguf
        config:
          temperature: 0.01
          tokenizer_type: llama
    config:
      endpoint: http://localhost:8080

llm_provider: LocalAI
llm_model: thebloke__codellama-13b-gguf__codellama-13b.q6_k.gguf

endpoint - an address to a running LocalAI instance

Also, note the model name. The name should be the same as installed in the LocalAI.

Last updated