Skip to content

Commit 3ec55de

Browse files
committed
docs: Update changelog for v0.0.13
1 parent 0fe83a9 commit 3ec55de

File tree

1 file changed

+21
-0
lines changed

1 file changed

+21
-0
lines changed

changelog.md

Lines changed: 21 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -6,6 +6,27 @@ The format is based on [Keep a Changelog](https://keepachangelog.com/en/1.0.0/),
66

77
## [Unreleased]
88

9+
## [0.0.13] - 2025-03-13
10+
11+
### Fixed - GPU Memory Leak and Duplicate Model Runs
12+
13+
**Critical: Prevents OOM errors and eliminates redundant inference calls**
14+
15+
- **GPU Memory Cleanup**: Added `_cleanup_gpu_memory()` function that runs `gc.collect()` and `torch.cuda.empty_cache()` between test iterations to prevent out-of-memory errors with local models
16+
- Only activates for `transformers` provider (GPU models)
17+
- Optional verbose mode shows memory stats after cleanup
18+
19+
- **Eliminate Duplicate Model Runs**: `analyze_streamed_steps()` now captures the response directly from `FinalAnswerStep.output`, removing the redundant `agent.run()` call
20+
- Previously, the model was invoked twice per test case — once via streaming and once via `agent.run()`
21+
- This halves inference time and GPU memory usage per test
22+
- Function now returns a 4-tuple: `(tools_used, final_answer_called, steps_count, response)`
23+
24+
**Files Modified:**
25+
- `smoltrace/core.py` - GPU cleanup + response capture from streaming
26+
- `tests/test_core.py` - Updated mocks for 4-tuple return
27+
- `tests/test_core_additional.py` - Updated mocks for 4-tuple return
28+
- `tests/test_final_coverage_push.py` - Updated mocks for 4-tuple return
29+
930
## [0.0.12] - 2025-01-30
1031

1132
### Fixed - Enable trust_remote_code by Default

0 commit comments

Comments
 (0)