[misc] update pre-commit and run all files (#4752)

* [misc] update pre-commit

* [misc] run pre-commit

* [misc] remove useless configuration files

* [misc] ignore cuda for clang-format
This commit is contained in:
Hongxin Liu
2023-09-19 14:20:26 +08:00
committed by GitHub
parent 3c6b831c26
commit 079bf3cb26
1268 changed files with 50037 additions and 38444 deletions

View File

@@ -7,7 +7,7 @@ from torch.distributed import ProcessGroup
from .parallel_module import ParallelModule
from .utils import create_randomizer_with_offset
__all__ = ['DropoutForParallelInput', 'DropoutForReplicatedInput']
__all__ = ["DropoutForParallelInput", "DropoutForReplicatedInput"]
class DropoutForParallelInput(ParallelModule, nn.Dropout):
@@ -31,8 +31,9 @@ class DropoutForParallelInput(ParallelModule, nn.Dropout):
self.randomizer = create_randomizer_with_offset(seed, process_group=process_group)
@staticmethod
def from_native_module(module: nn.Dropout,
process_group: Union[ProcessGroup, List[ProcessGroup]] = None) -> "DropoutForParallelInput":
def from_native_module(
module: nn.Dropout, process_group: Union[ProcessGroup, List[ProcessGroup]] = None
) -> "DropoutForParallelInput":
"""
Create a DropoutForParallelInput layer from a native dropout layer.
"""
@@ -68,8 +69,8 @@ class DropoutForReplicatedInput(ParallelModule, nn.Dropout):
@staticmethod
def from_native_module(
module: nn.Dropout,
process_group: Union[ProcessGroup, List[ProcessGroup]] = None) -> "DropoutForReplicatedInput":
module: nn.Dropout, process_group: Union[ProcessGroup, List[ProcessGroup]] = None
) -> "DropoutForReplicatedInput":
"""
Create a Dropout1D layer from a native dropout layer.
"""