Commit graph

62 commits

Author SHA1 Message Date
Nikhil Shrestha
1ea0663a3b Merge branch 'main' into Global-docker
# Conflicts:
#	poetry.lock
#	pyproject.toml
#	settings-local.yaml
#	settings.yaml
2024-05-04 17:16:25 +05:45
Saurab-Shrestha
1963190d16 Updated the llm component 2024-05-02 10:58:03 +05:45
Saurab-Shrestha
bc343206cc Updated docker settings 2024-04-30 17:45:51 +05:45
Patrick Peng
9d0d614706
fix: Replacing unsafe eval() with json.loads() (#1890)
Some checks failed
release-please / release-please (push) Has been cancelled
tests / setup (push) Has been cancelled
tests / ${{ matrix.quality-command }} (black) (push) Has been cancelled
tests / ${{ matrix.quality-command }} (mypy) (push) Has been cancelled
tests / ${{ matrix.quality-command }} (ruff) (push) Has been cancelled
tests / test (push) Has been cancelled
tests / all_checks_passed (push) Has been cancelled
2024-04-30 09:58:19 +02:00
icsy7867
e21bf20c10
feat: prompt_style applied to all LLMs + extra LLM params. (#1835)
Some checks are pending
release-please / release-please (push) Waiting to run
tests / setup (push) Waiting to run
tests / ${{ matrix.quality-command }} (black) (push) Blocked by required conditions
tests / ${{ matrix.quality-command }} (mypy) (push) Blocked by required conditions
tests / ${{ matrix.quality-command }} (ruff) (push) Blocked by required conditions
tests / test (push) Blocked by required conditions
tests / all_checks_passed (push) Blocked by required conditions
* Updated prompt_style to be moved to the main LLM setting since all LLMs from llama_index can utilize this.  I also included temperature, context window size, max_tokens, max_new_tokens into the openailike to help ensure the settings are consistent from the other implementations.

* Removed prompt_style from llamacpp entirely

* Fixed settings-local.yaml to include prompt_style in the LLM settings instead of llamacpp.
2024-04-30 09:53:10 +02:00
Saurab-Shrestha
f9a454861d Updated docker compose 2024-04-29 20:08:06 +05:45
Saurab-Shrestha
3f99b0996f Merged with dev 2024-04-28 11:29:26 +05:45
Saurab-Shrestha
c7aac53cd9 Added new docker files 2024-04-28 11:25:38 +05:45
Saurab-Shrestha9639*969**9858//852
1d6fc7144a Added llama3 prompt 2024-04-24 17:15:13 +05:45
imartinez
49ef729abc Allow passing HF access token to download tokenizer. Fallback to default tokenizer. 2024-04-19 15:38:25 +02:00
Pablo Orgaz
347be643f7
fix(llm): special tokens and leading space (#1831) 2024-04-04 14:37:29 +02:00
Saurab-Shrestha9639*969**9858//852
ee0e1cd839 Updated chat history and items id with uuid 2024-04-04 12:02:12 +05:45
Saurab-Shrestha9639*969**9858//852
4bc9dd7870 Added chat history and chat item 2024-04-03 17:58:27 +05:45
Nikhil Shrestha
9f929cf4f3 Merge branch 'main' into dev
# Conflicts:
#	docker-compose.yaml
#	poetry.lock
#	pyproject.toml
#	settings.yaml
2024-04-03 17:46:42 +05:45
Robin Boone
b3b0140e24
feat(llm): Ollama LLM-Embeddings decouple + longer keep_alive settings (#1800) 2024-04-02 16:23:10 +02:00
Iván Martínez
6f6c785dac
feat(llm): Ollama timeout setting (#1773)
* added request_timeout to ollama, default set to 30.0 in settings.yaml and settings-ollama.yaml

* Update settings-ollama.yaml

* Update settings.yaml

* updated settings.py and tidied up settings-ollama-yaml

* feat(UI): Faster startup and document listing (#1763)

* fix(ingest): update script label (#1770)

huggingface -> Hugging Face

* Fix lint errors

---------

Co-authored-by: Stephen Gresham <steve@gresham.id.au>
Co-authored-by: Ikko Eltociear Ashimine <eltociear@gmail.com>
2024-03-20 21:33:46 +01:00
Brett England
134fc54d7d
feat(ingest): Created a faster ingestion mode - pipeline (#1750)
* Unify pgvector and postgres connection settings

* Remove local changes

* Update file pgvector->postgres

* postgresql should be postgres

* Adding pipeline ingestion mode

* disable hugging face parallelism.  Continue on file to doc transform failure

* Semaphore to limit docq async workers. ETA reporting
2024-03-19 21:24:46 +01:00
Saurab-Shrestha
f7de7c3b54 Update with verify function to check the maker request for documents 2024-03-17 17:58:42 +05:45
Otto L
1efac6a3fe
feat(llm - embed): Add support for Azure OpenAI (#1698)
* Add support for Azure OpenAI

* fix: wrong default api_version

Should be dashes instead of underscores.
see: https://learn.microsoft.com/en-us/azure/ai-services/openai/reference

* fix: code styling

applied "make check" changes

* refactor: extend documentation

* mention azopenai as available option and extras
* add recommended section
* include settings-azopenai.yaml configuration file

* fix: documentation
2024-03-15 16:49:50 +01:00
Brett England
63de7e4930
feat: unify settings for vector and nodestore connections to PostgreSQL (#1730)
* Unify pgvector and postgres connection settings

* Remove local changes

* Update file pgvector->postgres
2024-03-15 09:55:17 +01:00
Brett England
68b3a34b03
feat(nodestore): add Postgres for the doc and index store (#1706)
* Adding Postgres for the doc and index store

* Adding documentation.  Rename postgres database local->simple.  Postgres storage dependencies

* Update documentation for postgres storage

* Renaming feature to nodestore

* update docstore -> nodestore in doc

* missed some docstore changes in doc

* Updated poetry.lock

* Formatting updates to pass ruff/black checks

* Correction to unreachable code!

* Format adjustment to pass black test

* Adjust extra inclusion name for vector pg

* extra dep change for pg vector

* storage-postgres -> storage-nodestore-postgres

* Hash change on poetry lock
2024-03-14 17:12:33 +01:00
Saurab-Shrestha9639*969**9858//852
ba72507d1f Merge branch 'main' of https://github.com/QuickfoxConsulting/privateGPT into ollama 2024-03-12 11:15:57 +05:45
icsy7867
02dc83e8e9
feat(llm): adds serveral settings for llamacpp and ollama (#1703) 2024-03-11 22:51:05 +01:00
Saurab-Shrestha
e2bad96854 Added apis for documents update and user checker mode 2024-03-07 17:17:04 +05:45
Saurab-Shrestha
f011bb6a7a Documents assignment with departments 2024-03-07 13:23:59 +05:45
Iván Martínez
45f05711eb
feat: Upgrade to LlamaIndex to 0.10 (#1663)
* Extract optional dependencies

* Separate local mode into llms-llama-cpp and embeddings-huggingface for clarity

* Support Ollama embeddings

* Upgrade to llamaindex 0.10.14. Remove legacy use of ServiceContext in ContextChatEngine

* Fix vector retriever filters
2024-03-06 17:51:30 +01:00
Saurab-Shrestha
a87531c141 Updated single injector for pdf ocr 2024-02-29 12:22:25 +05:45
Saurab-Shrestha
77bf96cc42 added singleton design in table ocr 2024-02-29 10:57:47 +05:45
Saurab-Shrestha
484b6b7223 Try catch added on ingest 2024-02-27 10:38:44 +05:45
quick-karsth
a8a10e412e updated gpu for ocr 2024-02-27 10:30:26 +05:45
Saurab-Shrestha
b9556626b1 Updated with chunk file upload 2024-02-26 16:58:11 +05:45
TQ
cd40e3982b
feat(Vector): support pgvector (#1624) 2024-02-20 15:29:26 +01:00
Saurab-Shrestha
c45f5b993d Added api for document listing 2024-02-19 17:28:29 +05:45
Saurab-Shrestha
c8b39c898c Solved the ingestion issue for ocred pdf 2024-02-17 10:23:57 +05:45
Saurab-Shrestha
d849ee76f4 Added routes for pdf ocr 2024-02-15 17:39:07 +05:45
Saurab-Shrestha
91ebce47d4 Updated with routes for pdf_ocr 2024-02-15 14:32:13 +05:45
quick-karsth
a29e0f4253 Added OCRComponents 2024-02-15 11:46:06 +05:45
quick-karsth
7bba4d13eb Added OCRComponents 2024-02-15 11:33:37 +05:45
Ygal Blum
6bbec79583
feat(llm): Add support for Ollama LLM (#1526) 2024-02-09 15:50:50 +01:00
Naveen Kannan
869233f0e4
fix: Adding an LLM param to fix broken generator from llamacpp (#1519) 2024-01-17 18:10:45 +01:00
CognitiveTech
e326126d0d
feat: add mistral + chatml prompts (#1426) 2024-01-16 22:51:14 +01:00
Matthew Hill
2d27a9f956
feat(llm): Add openailike llm mode (#1447)
This mode behaves the same as the openai mode, except that it allows setting custom models not
supported by OpenAI. It can be used with any tool that serves models from an OpenAI compatible API.

Implements #1424
2023-12-26 10:26:08 +01:00
Iván Martínez
4780540870
feat(settings): Configurable context_window and tokenizer (#1437) 2023-12-21 14:49:35 +01:00
3ly-13
a072a40a7c
Allow setting OpenAI model in settings (#1386)
feat(settings): Allow setting openai model to be used. Default to GPT 3.5
2023-12-09 20:13:00 +01:00
Louis Melchior
a3ed14c58f
feat(llm): drop default_system_prompt (#1385)
As discussed on Discord, the decision has been made to remove the system prompts by default, to better segregate the API and the UI usages.

A concurrent PR (#1353) is enabling the dynamic setting of a system prompt in the UI.

Therefore, if UI users want to use a custom system prompt, they can specify one directly in the UI.
If the API users want to use a custom prompt, they can pass it directly into their messages that they are passing to the API.

In the highlight of the two use case above, it becomes clear that default system_prompt does not need to exist.
2023-12-08 23:13:51 +01:00
lopagela
56af625d71
Fix the parallel ingestion mode, and make it available through conf (#1336)
* Fix the parallel ingestion mode, and make it available through conf

Also updated the documentation to show how to configure the ingest mode.

* PR feedback: redirect to documentation
2023-11-30 11:41:55 +01:00
Gianni Acquisto
9c192ddd73
Added max_new_tokens as a config option to llm yaml block (#1317)
* added max_new_tokens as a configuration option to the llm block in settings

* Update fern/docs/pages/manual/settings.mdx

Co-authored-by: lopagela <lpglm@orange.fr>

* Update private_gpt/settings/settings.py

Add default value for max_new_tokens = 256

Co-authored-by: lopagela <lpglm@orange.fr>

* Addressed location of docs comment

* reformatting from running 'make check'

* remove default config value from settings.yaml

---------

Co-authored-by: lopagela <lpglm@orange.fr>
2023-11-26 19:17:29 +01:00
lopagela
bafdd3baf1
Ingestion Speedup Multiple strategy (#1309) 2023-11-25 20:12:09 +01:00
Iván Martínez
944c43bfa8
Multi language support - fern debug (#1307)
---------

Co-authored-by: Louis <lpglm@orange.fr>
Co-authored-by: LeMoussel <cnhx27@gmail.com>
2023-11-25 14:34:23 +01:00
Iván Martínez
510caa576b
Make qdrant the default vector db (#1285)
* Make qdrant the default vector db

---------

Co-authored-by: Pablo Orgaz <pabloogc@gmail.com>
Co-authored-by: lopagela <lpglm@orange.fr>
2023-11-20 16:19:22 +01:00