From 6594c94cbdd4e9c65f492e72145254e0a6aa26f4 Mon Sep 17 00:00:00 2001 From: Liam Dowd <101684827+itsliamdowd@users.noreply.github.com> Date: Wed, 28 Aug 2024 08:17:16 -0700 Subject: [PATCH] Update settings.py --- private_gpt/settings/settings.py | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/private_gpt/settings/settings.py b/private_gpt/settings/settings.py index 13ad9d8..6525884 100644 --- a/private_gpt/settings/settings.py +++ b/private_gpt/settings/settings.py @@ -114,7 +114,7 @@ class LLMSettings(BaseModel): "mock", "ollama", "gemini", - "mistral", + "mistralai", ] max_new_tokens: int = Field( 256, @@ -137,7 +137,7 @@ class LLMSettings(BaseModel): description="The temperature of the model. Increasing the temperature will make the model answer more creatively. A value of 0.1 would be more factual.", ) prompt_style: Literal[ - "default", "llama2", "llama3", "tag", "mistral", "chatml" + "default", "llama2", "llama3", "tag", "mistralai", "chatml" ] = Field( "llama2", description=( @@ -197,7 +197,7 @@ class HuggingFaceSettings(BaseModel): class EmbeddingSettings(BaseModel): mode: Literal[ - "huggingface", "openai", "azopenai", "sagemaker", "ollama", "mock", "gemini", "mistral" + "huggingface", "openai", "azopenai", "sagemaker", "ollama", "mock", "gemini", "mistralai" ] ingest_mode: Literal["simple", "batch", "parallel", "pipeline"] = Field( "simple",