mirror of
https://github.com/zylon-ai/private-gpt.git
synced 2025-12-22 10:45:42 +01:00
fix: mypy
This commit is contained in:
parent
0a9c57447b
commit
c096a42aa1
2 changed files with 14 additions and 13 deletions
|
|
@ -68,10 +68,10 @@ class EmbeddingComponent:
|
|||
)
|
||||
case "ollama":
|
||||
try:
|
||||
from ollama import Client, AsyncClient # type: ignore
|
||||
from llama_index.embeddings.ollama import ( # type: ignore
|
||||
OllamaEmbedding,
|
||||
)
|
||||
from ollama import Client # type: ignore
|
||||
except ImportError as e:
|
||||
raise ImportError(
|
||||
"Local dependencies not found, install with `poetry install --extras embeddings-ollama`"
|
||||
|
|
@ -79,9 +79,9 @@ class EmbeddingComponent:
|
|||
|
||||
ollama_settings = settings.ollama
|
||||
|
||||
# calculate embedding model. If not provided tag, it will be use latest model
|
||||
# Calculate embedding model. If not provided tag, it will be use latest
|
||||
model_name = (
|
||||
ollama_settings.embedding_model + ':latest'
|
||||
ollama_settings.embedding_model + ":latest"
|
||||
if ":" not in ollama_settings.embedding_model
|
||||
else ollama_settings.embedding_model
|
||||
)
|
||||
|
|
@ -98,11 +98,11 @@ class EmbeddingComponent:
|
|||
host=ollama_settings.embedding_api_base,
|
||||
timeout=ollama_settings.request_timeout,
|
||||
)
|
||||
installed_models = [model['name'] for model in client.list().get("models", {})]
|
||||
installed_models = [
|
||||
model["name"] for model in client.list().get("models", {})
|
||||
]
|
||||
if model_name not in installed_models:
|
||||
logger.info(
|
||||
f"Pulling model {model_name}. Please wait..."
|
||||
)
|
||||
logger.info(f"Pulling model {model_name}. Please wait...")
|
||||
client.pull(model_name)
|
||||
logger.info(f"Model {model_name} pulled successfully")
|
||||
except Exception as e:
|
||||
|
|
|
|||
|
|
@ -146,9 +146,9 @@ class LLMComponent:
|
|||
"repeat_penalty": ollama_settings.repeat_penalty, # ollama llama-cpp
|
||||
}
|
||||
|
||||
# calculate llm model. If not provided tag, it will be use latest model
|
||||
# calculate llm model. If not provided tag, it will be use latest
|
||||
model_name = (
|
||||
ollama_settings.llm_model + ':latest'
|
||||
ollama_settings.llm_model + ":latest"
|
||||
if ":" not in ollama_settings.llm_model
|
||||
else ollama_settings.llm_model
|
||||
)
|
||||
|
|
@ -164,11 +164,12 @@ class LLMComponent:
|
|||
|
||||
if ollama_settings.autopull_models:
|
||||
try:
|
||||
installed_models = [model['name'] for model in llm.client.list().get("models", {})]
|
||||
installed_models = [
|
||||
model["name"]
|
||||
for model in llm.client.list().get("models", {})
|
||||
]
|
||||
if model_name not in installed_models:
|
||||
logger.info(
|
||||
f"Pulling model {model_name}. Please wait..."
|
||||
)
|
||||
logger.info(f"Pulling model {model_name}. Please wait...")
|
||||
llm.client.pull(model_name)
|
||||
logger.info(f"Model {model_name} pulled successfully")
|
||||
except Exception as e:
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue