mirror of
https://github.com/hpcaitech/ColossalAI.git
synced 2025-09-15 22:19:38 +00:00
fix typo with colossalai/trainer utils zero (#3908)
This commit is contained in:
@@ -25,7 +25,7 @@ class ChunkMemStatsCollector(MemStatsCollector):
|
||||
# override
|
||||
def record_model_data_volume(self) -> None:
|
||||
"""
|
||||
record model data volumn on cuda and cpu.
|
||||
record model data volume on cuda and cpu.
|
||||
"""
|
||||
if self._start_flag and not self.use_outside_memstats:
|
||||
cuda_mem = self._chunk_manager.total_mem['cuda']
|
||||
|
@@ -45,7 +45,7 @@ class MemoryMonitor:
|
||||
|
||||
class AsyncMemoryMonitor(MemoryMonitor):
|
||||
"""
|
||||
An Async Memory Monitor runing during computing. Sampling memory usage of the current GPU
|
||||
An Async Memory Monitor running during computing. Sampling memory usage of the current GPU
|
||||
at interval of `1/(10**power)` sec.
|
||||
|
||||
The idea comes from Runtime Memory Tracer of PatrickStar
|
||||
@@ -67,7 +67,7 @@ class AsyncMemoryMonitor(MemoryMonitor):
|
||||
async_mem_monitor.save('log.pkl')
|
||||
|
||||
Args:
|
||||
power (int, optional): the power of time interva. Defaults to 10.
|
||||
power (int, optional): the power of time interval. Defaults to 10.
|
||||
|
||||
.. _PatrickStar: Parallel Training of Pre-trained Models via Chunk-based Memory Management:
|
||||
https://arxiv.org/abs/2108.05818
|
||||
|
Reference in New Issue
Block a user