Skip to content

Commit

Permalink
Pass streaming parameter to LLM loader
Browse files Browse the repository at this point in the history
Signed-off-by: Pavel Tisnovsky <[email protected]>
  • Loading branch information
tisnik committed Jan 27, 2025
1 parent 540a694 commit ae072eb
Show file tree
Hide file tree
Showing 4 changed files with 9 additions and 2 deletions.
1 change: 1 addition & 0 deletions ols/app/endpoints/ols.py
Original file line number Diff line number Diff line change
Expand Up @@ -376,6 +376,7 @@ def generate_response(
provider=llm_request.provider,
model=llm_request.model,
system_prompt=llm_request.system_prompt,
streaming=streaming,
)
history = CacheEntry.cache_entries_to_history(previous_input)
if streaming:
Expand Down
6 changes: 5 additions & 1 deletion ols/src/llms/llm_loader.py
Original file line number Diff line number Diff line change
Expand Up @@ -53,14 +53,18 @@ def resolve_provider_config(


def load_llm(
provider: str, model: str, generic_llm_params: Optional[dict] = None
provider: str,
model: str,
generic_llm_params: Optional[dict] = None,
streaming: Optional[bool] = None,
) -> LLM:
"""Load LLM according to input provider and model.
Args:
provider: The provider name.
model: The model name.
generic_llm_params: The optional parameters that will be converted into LLM-specific ones.
streaming: The optional parameter that enable streaming on LLM side if set to True.
Raises:
LLMConfigurationError: If the whole provider configuration is missing.
Expand Down
2 changes: 1 addition & 1 deletion ols/src/query_helpers/docs_summarizer.py
Original file line number Diff line number Diff line change
Expand Up @@ -42,7 +42,7 @@ def _prepare_llm(self) -> None:
GenericLLMParameters.MAX_TOKENS_FOR_RESPONSE: self.model_config.parameters.max_tokens_for_response # noqa: E501
}
self.bare_llm = self.llm_loader(
self.provider, self.model, self.generic_llm_params
self.provider, self.model, self.generic_llm_params, self.streaming
)

def _get_system_prompt(self) -> None:
Expand Down
2 changes: 2 additions & 0 deletions ols/src/query_helpers/query_helper.py
Original file line number Diff line number Diff line change
Expand Up @@ -23,6 +23,7 @@ def __init__(
generic_llm_params: Optional[dict] = None,
llm_loader: Optional[Callable[[str, str, dict], LLM]] = None,
system_prompt: Optional[str] = None,
streaming: Optional[bool] = None,
) -> None:
"""Initialize query helper."""
# NOTE: As signature of this method is evaluated before the config,
Expand All @@ -32,6 +33,7 @@ def __init__(
self.model = model or config.ols_config.default_model
self.generic_llm_params = generic_llm_params or {}
self.llm_loader = llm_loader or load_llm
self.streaming = streaming or False

self._system_prompt = (
(config.dev_config.enable_system_prompt_override and system_prompt)
Expand Down

0 comments on commit ae072eb

Please sign in to comment.