mirror of
https://github.com/hpcaitech/ColossalAI.git
synced 2025-06-24 14:33:20 +00:00
update documentation
This commit is contained in:
parent
3245a69fc2
commit
3c7604ba30
@ -1,6 +1,8 @@
|
|||||||
# ColossalAI
|
# ColossalAI
|
||||||
|
|
||||||
An integrated large-scale model training system with efficient parallelization techniques
|
An integrated large-scale model training system with efficient parallelization techniques.
|
||||||
|
|
||||||
|
arXiv: [Colossal-AI: A Unified Deep Learning System For Large-Scale Parallel Training](https://arxiv.org/abs/2110.14883)
|
||||||
|
|
||||||
## Installation
|
## Installation
|
||||||
|
|
||||||
|
@ -12,6 +12,7 @@ from colossalai.registry import OPTIMIZERS
|
|||||||
class Lamb(Optimizer):
|
class Lamb(Optimizer):
|
||||||
r"""Implements Lamb algorithm.
|
r"""Implements Lamb algorithm.
|
||||||
It has been proposed in `Large Batch Optimization for Deep Learning: Training BERT in 76 minutes`_.
|
It has been proposed in `Large Batch Optimization for Deep Learning: Training BERT in 76 minutes`_.
|
||||||
|
|
||||||
Arguments:
|
Arguments:
|
||||||
params (iterable): iterable of parameters to optimize or dicts defining
|
params (iterable): iterable of parameters to optimize or dicts defining
|
||||||
parameter groups
|
parameter groups
|
||||||
@ -23,7 +24,8 @@ class Lamb(Optimizer):
|
|||||||
weight_decay (float, optional): weight decay (L2 penalty) (default: 0)
|
weight_decay (float, optional): weight decay (L2 penalty) (default: 0)
|
||||||
adam (bool, optional): always use trust ratio = 1, which turns this into
|
adam (bool, optional): always use trust ratio = 1, which turns this into
|
||||||
Adam. Useful for comparison purposes.
|
Adam. Useful for comparison purposes.
|
||||||
.. _Large Batch Optimization for Deep Learning: Training BERT in 76 minutes:
|
|
||||||
|
.. _Large Batch Optimization for Deep Learning\: Training BERT in 76 minutes:
|
||||||
https://arxiv.org/abs/1904.00962
|
https://arxiv.org/abs/1904.00962
|
||||||
"""
|
"""
|
||||||
|
|
||||||
|
@ -12,6 +12,7 @@ from colossalai.registry import OPTIMIZERS
|
|||||||
class Lars(Optimizer):
|
class Lars(Optimizer):
|
||||||
r"""Implements the LARS optimizer from `"Large batch training of convolutional networks"
|
r"""Implements the LARS optimizer from `"Large batch training of convolutional networks"
|
||||||
<https://arxiv.org/pdf/1708.03888.pdf>`_.
|
<https://arxiv.org/pdf/1708.03888.pdf>`_.
|
||||||
|
|
||||||
Args:
|
Args:
|
||||||
params (iterable): iterable of parameters to optimize or dicts defining
|
params (iterable): iterable of parameters to optimize or dicts defining
|
||||||
parameter groups
|
parameter groups
|
||||||
@ -35,7 +36,8 @@ class Lars(Optimizer):
|
|||||||
if momentum < 0.0:
|
if momentum < 0.0:
|
||||||
raise ValueError("Invalid momentum value: {}".format(momentum))
|
raise ValueError("Invalid momentum value: {}".format(momentum))
|
||||||
if weight_decay < 0.0:
|
if weight_decay < 0.0:
|
||||||
raise ValueError("Invalid weight_decay value: {}".format(weight_decay))
|
raise ValueError(
|
||||||
|
"Invalid weight_decay value: {}".format(weight_decay))
|
||||||
if eeta <= 0 or eeta > 1:
|
if eeta <= 0 or eeta > 1:
|
||||||
raise ValueError("Invalid eeta value: {}".format(eeta))
|
raise ValueError("Invalid eeta value: {}".format(eeta))
|
||||||
if epsilon < 0:
|
if epsilon < 0:
|
||||||
@ -48,6 +50,7 @@ class Lars(Optimizer):
|
|||||||
@torch.no_grad()
|
@torch.no_grad()
|
||||||
def step(self, closure=None):
|
def step(self, closure=None):
|
||||||
"""Performs a single optimization step.
|
"""Performs a single optimization step.
|
||||||
|
|
||||||
Arguments:
|
Arguments:
|
||||||
closure (callable, optional): A closure that reevaluates the model
|
closure (callable, optional): A closure that reevaluates the model
|
||||||
and returns the loss.
|
and returns the loss.
|
||||||
|
@ -1,5 +0,0 @@
|
|||||||
colossalai.nn.data.prefetcher
|
|
||||||
=============================
|
|
||||||
|
|
||||||
.. automodule:: colossalai.nn.data.prefetcher
|
|
||||||
:members:
|
|
@ -16,5 +16,3 @@ colossalai.nn.data
|
|||||||
colossalai.nn.data.base_dataset
|
colossalai.nn.data.base_dataset
|
||||||
colossalai.nn.data.caltech101_dataset
|
colossalai.nn.data.caltech101_dataset
|
||||||
colossalai.nn.data.cifar10_dataset
|
colossalai.nn.data.cifar10_dataset
|
||||||
colossalai.nn.data.prefetcher
|
|
||||||
colossalai.nn.data.wiki_dataset
|
|
||||||
|
@ -1,5 +0,0 @@
|
|||||||
colossalai.nn.data.wiki\_dataset
|
|
||||||
================================
|
|
||||||
|
|
||||||
.. automodule:: colossalai.nn.data.wiki_dataset
|
|
||||||
:members:
|
|
@ -1,5 +0,0 @@
|
|||||||
colossalai.nn.model.bert.bert
|
|
||||||
=============================
|
|
||||||
|
|
||||||
.. automodule:: colossalai.nn.model.bert.bert
|
|
||||||
:members:
|
|
@ -1,11 +0,0 @@
|
|||||||
colossalai.nn.model.bert
|
|
||||||
========================
|
|
||||||
|
|
||||||
.. automodule:: colossalai.nn.model.bert
|
|
||||||
:members:
|
|
||||||
|
|
||||||
|
|
||||||
.. toctree::
|
|
||||||
:maxdepth: 2
|
|
||||||
|
|
||||||
colossalai.nn.model.bert.bert
|
|
@ -7,7 +7,6 @@ colossalai.nn.model
|
|||||||
.. toctree::
|
.. toctree::
|
||||||
:maxdepth: 2
|
:maxdepth: 2
|
||||||
|
|
||||||
colossalai.nn.model.bert
|
|
||||||
colossalai.nn.model.vanilla_resnet
|
colossalai.nn.model.vanilla_resnet
|
||||||
colossalai.nn.model.vision_transformer
|
colossalai.nn.model.vision_transformer
|
||||||
|
|
||||||
|
5
docs/colossalai/colossalai.nn.optimizer.lars.rst
Normal file
5
docs/colossalai/colossalai.nn.optimizer.lars.rst
Normal file
@ -0,0 +1,5 @@
|
|||||||
|
colossalai.nn.optimizer.lars
|
||||||
|
============================
|
||||||
|
|
||||||
|
.. automodule:: colossalai.nn.optimizer.lars
|
||||||
|
:members:
|
@ -13,6 +13,7 @@ colossalai.nn.optimizer
|
|||||||
colossalai.nn.optimizer.fused_lamb
|
colossalai.nn.optimizer.fused_lamb
|
||||||
colossalai.nn.optimizer.fused_sgd
|
colossalai.nn.optimizer.fused_sgd
|
||||||
colossalai.nn.optimizer.lamb
|
colossalai.nn.optimizer.lamb
|
||||||
|
colossalai.nn.optimizer.lars
|
||||||
colossalai.nn.optimizer.loss_scaler
|
colossalai.nn.optimizer.loss_scaler
|
||||||
colossalai.nn.optimizer.zero_redundancy_optimizer_level_1
|
colossalai.nn.optimizer.zero_redundancy_optimizer_level_1
|
||||||
colossalai.nn.optimizer.zero_redundancy_optimizer_level_2
|
colossalai.nn.optimizer.zero_redundancy_optimizer_level_2
|
||||||
|
Loading…
Reference in New Issue
Block a user