From 94a5af66eb4babc219d0283c523b8662f0b0d81e Mon Sep 17 00:00:00 2001 From: shamoon <4887959+shamoon@users.noreply.github.com> Date: Wed, 14 Jan 2026 15:36:01 -0800 Subject: [PATCH] Fix default llama3.1 --- docs/configuration.md | 2 +- src/paperless_ai/client.py | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/docs/configuration.md b/docs/configuration.md index 6a3624d363..b7b24d3133 100644 --- a/docs/configuration.md +++ b/docs/configuration.md @@ -1873,7 +1873,7 @@ using the OpenAI API. This setting is required to be set to use the AI features. #### [`PAPERLESS_AI_LLM_MODEL=`](#PAPERLESS_AI_LLM_MODEL) {#PAPERLESS_AI_LLM_MODEL} : The model to use for the AI backend, i.e. "gpt-3.5-turbo", "gpt-4" or any of the models supported by the -current backend. If not supplied, defaults to "gpt-3.5-turbo" for OpenAI and "llama3" for Ollama. +current backend. If not supplied, defaults to "gpt-3.5-turbo" for OpenAI and "llama3.1" for Ollama. Defaults to None. diff --git a/src/paperless_ai/client.py b/src/paperless_ai/client.py index 2b66bfd9ba..1f52c56c72 100644 --- a/src/paperless_ai/client.py +++ b/src/paperless_ai/client.py @@ -23,7 +23,7 @@ class AIClient: def get_llm(self) -> Ollama | OpenAI: if self.settings.llm_backend == "ollama": return Ollama( - model=self.settings.llm_model or "llama3", + model=self.settings.llm_model or "llama3.1", base_url=self.settings.llm_endpoint or "http://localhost:11434", request_timeout=120, ) -- 2.47.3