mirror of
https://github.com/zylon-ai/private-gpt.git
synced 2025-12-22 17:05:41 +01:00
update doc for ipex-llm
This commit is contained in:
parent
3b3e96ad6c
commit
220c42ee85
1 changed files with 6 additions and 0 deletions
|
|
@ -193,3 +193,9 @@ or
|
|||
|
||||
When the server is started it will print a log *Application startup complete*.
|
||||
Navigate to http://localhost:8001 to use the Gradio UI or to http://localhost:8001/docs (API section) to try the API.
|
||||
|
||||
### Using IPEX-LLM
|
||||
|
||||
For a fully private setup on Intel GPUs (such as a local PC with an iGPU, or discrete GPUs like Arc, Flex, and Max), you can use [IPEX-LLM](https://github.com/intel-analytics/ipex-llm).
|
||||
|
||||
To deploy Ollama and pull models using IPEX-LLM, please refer to [this guide](https://ipex-llm.readthedocs.io/en/latest/doc/LLM/Quickstart/ollama_quickstart.html). Then, follow the same steps outlined in the [Using Ollama](#using-ollama) section to create a `settings-ollama.yaml` profile and run the private-GPT server.
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue