diff --git a/backend/llm.py b/backend/llm.py index e4250d3..5b0d81e 100644 --- a/backend/llm.py +++ b/backend/llm.py @@ -114,6 +114,10 @@ async def stream_claude_response( # Return final message response = await stream.get_final_message() + + # Close the Anthropic client + await client.close() + return response.content[0].text @@ -178,6 +182,9 @@ async def stream_claude_response_native( f"Token usage: Input Tokens: {response.usage.input_tokens}, Output Tokens: {response.usage.output_tokens}" ) + # Close the Anthropic client + await client.close() + if not response: raise Exception("No HTML response found in AI response") else: