Skip to content
This repository has been archived by the owner on Nov 13, 2024. It is now read-only.

Commit

Permalink
Fix comments
Browse files Browse the repository at this point in the history
  • Loading branch information
izellevy committed Dec 18, 2023
1 parent 2a8c278 commit 0cc21d5
Show file tree
Hide file tree
Showing 5 changed files with 6 additions and 6 deletions.
2 changes: 1 addition & 1 deletion config/anyscale.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -27,7 +27,7 @@ chat_engine:
# Since Anyscale's LLM endpoint currently doesn't support function calling, we will use the LastMessageQueryGenerator
# --------------------------------------------------------------------
query_builder:
type: CondensedQueryGenerator # Options: [CondensedQueryGenerator, LastMessageQueryGenerator]
type: InstructionQueryGenerator # Options: [InstructionQueryGenerator, LastMessageQueryGenerator]

# -------------------------------------------------------------------------------------------------------------
# ContextEngine configuration
Expand Down
2 changes: 1 addition & 1 deletion config/config.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -59,7 +59,7 @@ chat_engine:
# The query builder is responsible for generating textual queries given user message history.
# --------------------------------------------------------------------
query_builder:
type: FunctionCallingQueryGenerator # Options: [FunctionCallingQueryGenerator, LastMessageQueryGenerator, CondensedQueryGenerator]
type: FunctionCallingQueryGenerator # Options: [FunctionCallingQueryGenerator, LastMessageQueryGenerator, InstructionQueryGenerator]
params:
prompt: *query_builder_prompt # The query builder's system prompt for calling the LLM
function_description: # A function description passed to the LLM's `function_calling` API
Expand Down
2 changes: 1 addition & 1 deletion src/canopy/chat_engine/query_generator/__init__.py
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
from .base import QueryGenerator
from .function_calling import FunctionCallingQueryGenerator
from .last_message import LastMessageQueryGenerator
from .condensed import CondensedQueryGenerator
from .instruction import InstructionQueryGenerator
Original file line number Diff line number Diff line change
Expand Up @@ -56,7 +56,7 @@ class ExtractionException(ValueError):
pass


class CondensedQueryGenerator(QueryGenerator):
class InstructionQueryGenerator(QueryGenerator):
_DEFAULT_COMPONENTS = {
"llm": OpenAILLM,
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,7 @@

import pytest

from canopy.chat_engine.query_generator import CondensedQueryGenerator
from canopy.chat_engine.query_generator import InstructionQueryGenerator
from canopy.llm import BaseLLM
from canopy.models.api_models import ChatResponse, _Choice, TokenCounts
from canopy.models.data_models import Query, UserMessage, AssistantMessage
Expand All @@ -16,7 +16,7 @@ def mock_llm():

@pytest.fixture
def query_generator(mock_llm):
query_gen = CondensedQueryGenerator(
query_gen = InstructionQueryGenerator(
llm=mock_llm,
)
return query_gen
Expand Down

0 comments on commit 0cc21d5

Please sign in to comment.