mirror of
https://github.com/hpcaitech/ColossalAI.git
synced 2025-09-18 07:31:19 +00:00
fix typo colossalai/auto_parallel autochunk fx/passes etc. (#3808)
This commit is contained in:
@@ -148,7 +148,7 @@ class MetaInfoProp:
|
||||
graph_info.fwd_tmp = buffer_tensors
|
||||
graph_info.fwd_out = output_tensors
|
||||
|
||||
# fetch other memory informations
|
||||
# fetch other memory information
|
||||
memory_cost = meta_info.memory_cost
|
||||
graph_info.fwd_mem_tmp = memory_cost.fwd.temp
|
||||
graph_info.fwd_mem_out = memory_cost.fwd.activation
|
||||
|
@@ -44,7 +44,7 @@ class BatchNormStrategyGenerator(StrategyGenerator):
|
||||
'''
|
||||
Compute the computation cost per device with this specific strategy.
|
||||
|
||||
Note: compute_cost need to be devided by TFLOPS, now it just shows the computation size.
|
||||
Note: compute_cost need to be divided by TFLOPS, now it just shows the computation size.
|
||||
'''
|
||||
# TODO: a constant coefficient need to be added.
|
||||
# 1D: (L) * N * Cin
|
||||
|
@@ -38,9 +38,9 @@ class ConvStrategyGenerator(StrategyGenerator):
|
||||
'''
|
||||
Compute the computation cost per device with this specific strategy.
|
||||
|
||||
Note: compute_cost need to be devided by TFLOPS, now it just shows the computation size.
|
||||
Note: compute_cost need to be divided by TFLOPS, now it just shows the computation size.
|
||||
'''
|
||||
# TODO: compute_cost need to be devided by TFLOPS, now it just shows the computation size.
|
||||
# TODO: compute_cost need to be divided by TFLOPS, now it just shows the computation size.
|
||||
# 1D: (L) * N * Cout * Cin * kernel
|
||||
# 2D: (H * W) * N * Cout * Cin * kernel
|
||||
# 3D: (H * W * D) * N * Cout * Cin * kernel
|
||||
|
@@ -34,9 +34,9 @@ class LayerNormGenerator(StrategyGenerator):
|
||||
'''
|
||||
Compute the computation cost per device with this specific strategy.
|
||||
|
||||
Note: compute_cost need to be devided by TFLOPS, now it just shows the computation size.
|
||||
Note: compute_cost need to be divided by TFLOPS, now it just shows the computation size.
|
||||
'''
|
||||
# TODO: compute_cost need to be devided by TFLOPS, now it just shows the computation size.
|
||||
# TODO: compute_cost need to be divided by TFLOPS, now it just shows the computation size.
|
||||
# TODO: a constant coefficient need to be added.
|
||||
|
||||
sharded_input_shape = strategy.sharding_specs[self.op_data['input']].get_sharded_shape_per_device()
|
||||
|
@@ -17,7 +17,7 @@ class NormalPoolStrategyGenerator(StrategyGenerator):
|
||||
"""
|
||||
NormalPoolStrategyGenerator is a generic class to generate strategies for pool operation like MaxPoolxd.
|
||||
The reason we call this normal pool is AvgPoolxd and MaxPoolxd are taking the kernel size element from image,
|
||||
and reduce them depening on the operation type.
|
||||
and reduce them depending on the operation type.
|
||||
"""
|
||||
|
||||
def validate(self) -> bool:
|
||||
@@ -35,9 +35,9 @@ class NormalPoolStrategyGenerator(StrategyGenerator):
|
||||
'''
|
||||
Compute the computation cost per device with this specific strategy.
|
||||
|
||||
Note: compute_cost need to be devided by TFLOPS, now it just shows the computation size.
|
||||
Note: compute_cost need to be divided by TFLOPS, now it just shows the computation size.
|
||||
'''
|
||||
# TODO: compute_cost need to be devided by TFLOPS, now it just shows the computation size.
|
||||
# TODO: compute_cost need to be divided by TFLOPS, now it just shows the computation size.
|
||||
# 1D: (Lout) * N * C * kernel
|
||||
# 2D: (H * W) * N * Cout * Cin * kernel
|
||||
# 3D: (H * W * D) * N * Cout * Cin * kernel
|
||||
|
Reference in New Issue
Block a user