mirror of
https://github.com/zylon-ai/private-gpt.git
synced 2025-12-22 20:12:55 +01:00
Per requested changes, used ternary for response style within each mode
This commit is contained in:
parent
079761ea3c
commit
78fa938ed8
1 changed files with 12 additions and 20 deletions
|
|
@ -170,12 +170,6 @@ class PrivateGptUi:
|
||||||
role=MessageRole.SYSTEM,
|
role=MessageRole.SYSTEM,
|
||||||
),
|
),
|
||||||
)
|
)
|
||||||
def draw_methods(service_type):
|
|
||||||
service = getattr(self, f'_{service_type}_service')
|
|
||||||
return {
|
|
||||||
True: getattr(service, f'stream_{service_type}'),
|
|
||||||
False: getattr(service, f'{service_type}')
|
|
||||||
}
|
|
||||||
match mode:
|
match mode:
|
||||||
case Modes.RAG_MODE:
|
case Modes.RAG_MODE:
|
||||||
# Use only the selected file for the query
|
# Use only the selected file for the query
|
||||||
|
|
@ -190,18 +184,17 @@ class PrivateGptUi:
|
||||||
docs_ids.append(ingested_document.doc_id)
|
docs_ids.append(ingested_document.doc_id)
|
||||||
context_filter = ContextFilter(docs_ids=docs_ids)
|
context_filter = ContextFilter(docs_ids=docs_ids)
|
||||||
|
|
||||||
methods = draw_methods('chat')
|
query_stream = self._chat_service.stream_chat(
|
||||||
query_stream = methods.get(self._response_style, self._chat_service.stream_chat)(
|
all_messages, use_context=False
|
||||||
messages=all_messages,
|
) if self._response_style else self._chat_service.chat(
|
||||||
use_context=True,
|
all_messages, use_context=False
|
||||||
context_filter=context_filter
|
|
||||||
)
|
)
|
||||||
yield from (yield_deltas(query_stream) if self._response_style else [query_stream.response])
|
yield from (yield_deltas(query_stream) if self._response_style else [query_stream.response])
|
||||||
case Modes.BASIC_CHAT_MODE:
|
case Modes.BASIC_CHAT_MODE:
|
||||||
methods = draw_methods('chat')
|
llm_stream = self._chat_service.stream_chat(
|
||||||
llm_stream = methods.get(self._response_style, self._chat_service.stream_chat)(
|
all_messages, use_context=False
|
||||||
messages=all_messages,
|
) if self._response_style else self._chat_service.chat(
|
||||||
use_context=False
|
all_messages, use_context=False
|
||||||
)
|
)
|
||||||
yield from (yield_deltas(llm_stream) if self._response_style else [llm_stream.response])
|
yield from (yield_deltas(llm_stream) if self._response_style else [llm_stream.response])
|
||||||
|
|
||||||
|
|
@ -238,11 +231,10 @@ class PrivateGptUi:
|
||||||
)
|
)
|
||||||
yield from yield_tokens(summary_stream)
|
yield from yield_tokens(summary_stream)
|
||||||
'''
|
'''
|
||||||
methods = draw_methods('summarize')
|
summary_stream = self._summarize_service.summarize_stream(
|
||||||
summary_stream = methods.get(self._response_style, self._summarize_service.stream_summarize)(
|
all_messages, use_context=False
|
||||||
use_context=True,
|
) if self._response_style else self._summarize_service.summarize(
|
||||||
context_filter=context_filter,
|
all_messages, use_context=False
|
||||||
instructions=message
|
|
||||||
)
|
)
|
||||||
yield from yield_tokens(summary_stream) if response_style else summary_stream
|
yield from yield_tokens(summary_stream) if response_style else summary_stream
|
||||||
'''
|
'''
|
||||||
|
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue