Skip to content

Commit a80b7e5

Browse files
misc
1 parent 68aee31 commit a80b7e5

File tree

2 files changed

+4
-4
lines changed

2 files changed

+4
-4
lines changed

recipes/dev/generate_v2.py

+2-2
Original file line numberDiff line numberDiff line change
@@ -113,10 +113,10 @@ def log_metrics(self, total_time: int, tokens_per_second: float) -> None:
113113
f"Time for inference: {total_time:.02f} sec total, {tokens_per_second:.02f} tokens/sec"
114114
)
115115
self._logger.info(
116-
f"Bandwidth achieved: {model_size * tokens_per_second / 1024 / 1024 / 1024:.02f} GiB/s"
116+
f"Bandwidth achieved: {model_size * tokens_per_second / (1024**3):.02f} GiB/s"
117117
)
118118
self._logger.info(
119-
f"Max memory allocated: {torch.cuda.max_memory_allocated() / 1024 / 1024 / 1024:.02f} GiB"
119+
f"Max memory allocated: {torch.cuda.max_memory_allocated() / (1024**3):.02f} GiB"
120120
)
121121

122122
@torch.inference_mode()

recipes/dev/generate_v2_distributed.py

+2-2
Original file line numberDiff line numberDiff line change
@@ -154,10 +154,10 @@ def log_metrics(self, total_time: int, tokens_per_second: float) -> None:
154154
f"Time for inference: {total_time:.02f} sec total, {tokens_per_second:.02f} tokens/sec"
155155
)
156156
self._logger.info(
157-
f"Bandwidth achieved: {model_size * tokens_per_second / 1024 / 1024 / 1024:.02f} GiB/s"
157+
f"Bandwidth achieved: {model_size * tokens_per_second / (1024**3):.02f} GiB/s"
158158
)
159159
self._logger.info(
160-
f"Max memory allocated: {torch.cuda.max_memory_allocated() / 1024 / 1024 / 1024 :.02f} GiB"
160+
f"Max memory allocated: {torch.cuda.max_memory_allocated() / (1024**3):.02f} GiB"
161161
)
162162

163163
@torch.inference_mode()

0 commit comments

Comments
 (0)