Skip to content

Use with Ollama

Ollama allows you to run LLMs on your local machine.

Refer to the Ollama documentation to get Ollama up and running.

Then follow the steps below to connect InfernoAI to Ollama.

1. Set OLLAMA_ORIGINS to *

Follow the steps for your operating system.

Mac

If Ollama is run as a macOS application, environment variables should be set using launchctl:

  1. For each environment variable, call launchctl setenv.

    Terminal window
    launchctl setenv OLLAMA_ORIGINS "*"
  2. Restart Ollama application.

Linux

If Ollama is run as a systemd service, environment variables should be set using systemctl:

  1. Edit the systemd service by calling systemctl edit ollama.service. This will open an editor.

  2. For each environment variable, add a line Environment under section [Service]:

    [Service]
    Environment="OLLAMA_ORIGINS=*"
  3. Save and exit.

  4. Reload systemd and restart Ollama:

    Terminal window
    systemctl daemon-reload
    systemctl restart ollama

Windows

On windows, Ollama inherits your user and system environment variables.

  1. First Quit Ollama by clicking on it in the task bar

  2. Edit system environment variables from the control panel

  3. Set OLLAMA_ORIGINS to *

  4. Click OK/Apply to save

  5. Run ollama from a new terminal window

2. Add custom model in InfernoAI

Now that Ollama is allowing connections from any origin, add the Ollama model as a custom model in InfernoAI’s model settings.

Manage models

Open the “Manage models” dialog in the bottom left.

Add custom model

Click “Add custom model”

Custom model

Complete the form:

  • Name: A recognisable name for the model.
  • Endpoint URL: The Ollama chat completions endpoint URL. By default this is https://localhost:11434/v1/
  • Model ID: The exact ID for the model you want to access through Ollama.

Click “Create” and close the dialog. Now you can select the custom model in the models menu in the top right and start chatting.