Skip to content

Commit a76f832

Browse files
committed
fix chat issue
1 parent 92cc528 commit a76f832

File tree

4 files changed

+10
-10
lines changed

4 files changed

+10
-10
lines changed

templates/components/engines/python/agent/engine.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@
88
from llama_index.core.tools.query_engine import QueryEngineTool
99

1010

11-
def get_chat_engine(filters=None, params=None, event_handlers=None):
11+
def get_chat_engine(filters=None, params=None, event_handlers=None, **kwargs):
1212
system_prompt = os.getenv("SYSTEM_PROMPT")
1313
top_k = int(os.getenv("TOP_K", 0))
1414
tools = []

templates/components/engines/python/chat/engine.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@
99
from llama_index.core.settings import Settings
1010

1111

12-
def get_chat_engine(filters=None, params=None, event_handlers=None):
12+
def get_chat_engine(filters=None, params=None, event_handlers=None, **kwargs):
1313
system_prompt = os.getenv("SYSTEM_PROMPT")
1414
citation_prompt = os.getenv("SYSTEM_CITATION_PROMPT", None)
1515
top_k = int(os.getenv("TOP_K", 0))

templates/types/streaming/fastapi/app/api/routers/chat.py

Lines changed: 4 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,10 +1,10 @@
11
import logging
22
from typing import List
33

4-
from fastapi import APIRouter, BackgroundTasks, Depends, HTTPException, Request, status
4+
from fastapi import APIRouter, BackgroundTasks, HTTPException, Request, status
55
from llama_index.core.agent import AgentRunner
66
from llama_index.core.chat_engine import CondensePlusContextChatEngine
7-
from llama_index.core.chat_engine.types import BaseChatEngine, NodeWithScore
7+
from llama_index.core.chat_engine.types import NodeWithScore
88
from llama_index.core.llms import MessageRole
99

1010
from app.api.routers.events import EventCallbackHandler
@@ -86,11 +86,12 @@ async def chat(
8686
@r.post("/request")
8787
async def chat_request(
8888
data: ChatData,
89-
chat_engine: BaseChatEngine = Depends(get_chat_engine),
9089
) -> Result:
9190
last_message_content = data.get_last_message_content()
9291
messages = data.get_history_messages()
9392

93+
chat_engine = get_chat_engine(filters=None, params=None)
94+
9495
response = await chat_engine.achat(last_message_content, messages)
9596
return Result(
9697
result=Message(role=MessageRole.ASSISTANT, content=response.response),

templates/types/streaming/fastapi/app/api/routers/vercel_response.py

Lines changed: 4 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,6 @@
88
from fastapi.responses import StreamingResponse
99
from llama_index.core.chat_engine.types import StreamingAgentChatResponse
1010

11-
from app.agents.single import AgentRunEvent, AgentRunResult
1211
from app.api.routers.events import EventCallbackHandler
1312
from app.api.routers.models import ChatData, Message, SourceNodes
1413
from app.api.services.suggestion import NextQuestionSuggestion
@@ -144,16 +143,16 @@ def _create_stream(
144143
self,
145144
request: Request,
146145
chat_data: ChatData,
147-
event_handler: AgentRunResult | AsyncGenerator,
148-
events: AsyncGenerator[AgentRunEvent, None],
146+
event_handler: "AgentRunResult" | AsyncGenerator,
147+
events: AsyncGenerator["AgentRunEvent", None],
149148
verbose: bool = True,
150149
):
151150
# Yield the text response
152151
async def _chat_response_generator():
153152
result = await event_handler
154153
final_response = ""
155154

156-
if isinstance(result, AgentRunResult):
155+
if isinstance(result, "AgentRunResult"):
157156
for token in result.response.message.content:
158157
final_response += token
159158
yield self.convert_text(token)
@@ -185,7 +184,7 @@ async def _event_generator():
185184
return combine
186185

187186
@staticmethod
188-
def _event_to_response(event: AgentRunEvent) -> dict:
187+
def _event_to_response(event: "AgentRunEvent") -> dict:
189188
return {
190189
"type": "agent",
191190
"data": {"agent": event.name, "text": event.msg},

0 commit comments

Comments
 (0)