Skip to content

Commit 24c90ef

Browse files
committed
Pass cached tokens to RequestUsage
1 parent 13e144e commit 24c90ef

File tree

3 files changed

+3
-0
lines changed

3 files changed

+3
-0
lines changed

python/packages/autogen-core/src/autogen_core/models/_types.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -86,6 +86,7 @@ class FunctionExecutionResultMessage(BaseModel):
8686
class RequestUsage:
8787
prompt_tokens: int
8888
completion_tokens: int
89+
cached_tokens: int
8990

9091

9192
FinishReasons = Literal["stop", "length", "function_calls", "content_filter", "unknown"]

python/packages/autogen-ext/src/autogen_ext/models/anthropic/_anthropic_client.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -685,6 +685,7 @@ async def create(
685685
usage = RequestUsage(
686686
prompt_tokens=result.usage.input_tokens,
687687
completion_tokens=result.usage.output_tokens,
688+
cached_tokens=result.usage.cache_read_input_tokens or 0,
688689
)
689690
serializable_messages: List[Dict[str, Any]] = [self._serialize_message(msg) for msg in anthropic_messages]
690691

python/packages/autogen-ext/src/autogen_ext/models/openai/_openai_client.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -711,6 +711,7 @@ async def create(
711711
# TODO backup token counting
712712
prompt_tokens=getattr(result.usage, "prompt_tokens", 0) if result.usage is not None else 0,
713713
completion_tokens=getattr(result.usage, "completion_tokens", 0) if result.usage is not None else 0,
714+
cached_tokens=getattr(result.usage.prompt_tokens_details, "cached_tokens", 0) if result.usage is not None else 0,
714715
)
715716

716717
logger.info(

0 commit comments

Comments
 (0)