Skip to content
Merged
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
17 changes: 11 additions & 6 deletions lightrag/llm/anthropic.py
Original file line number Diff line number Diff line change
Expand Up @@ -100,8 +100,6 @@ async def anthropic_complete_if_cache(
)

messages: list[dict[str, Any]] = []
if system_prompt:
messages.append({"role": "system", "content": system_prompt})
messages.extend(history_messages)
messages.append({"role": "user", "content": prompt})

Expand All @@ -112,9 +110,16 @@ async def anthropic_complete_if_cache(
verbose_debug(f"System prompt: {system_prompt}")

try:
response = await anthropic_async_client.messages.create(
model=model, messages=messages, stream=True, **kwargs
)
create_params = {
'model': model,
'messages': messages,
'stream': True,
**kwargs
}
if system_prompt:
create_params['system'] = system_prompt
response = await anthropic_async_client.messages.create(**create_params)

except APIConnectionError as e:
logger.error(f"Anthropic API Connection Error: {e}")
raise
Expand All @@ -135,7 +140,7 @@ async def stream_response():
async for event in response:
content = (
event.delta.text
if hasattr(event, "delta") and event.delta.text
if hasattr(event, "delta") and hasattr(event.delta, "text") and event.delta.text
else None
)
if content is None:
Expand Down
Loading