mirror of
https://github.com/hpcaitech/ColossalAI.git
synced 2025-09-08 12:30:42 +00:00
[doc] update nvme offload documents. (#3850)
This commit is contained in:
@@ -78,8 +78,9 @@ from transformers.models.gpt2.modeling_gpt2 import GPT2LMHeadModel
|
||||
|
||||
import colossalai
|
||||
from colossalai.nn.optimizer import HybridAdam
|
||||
from colossalai.zero import zero_model_wrapper, zero_optim_wrapper
|
||||
from colossalai.utils.model.colo_init_context import ColoInitContext
|
||||
from colossalai.booster import Booster
|
||||
from colossalai.booster.plugin import GeminiPlugin
|
||||
```
|
||||
|
||||
Then we define a loss function:
|
||||
@@ -192,17 +193,23 @@ def train_gemini_cpu(nvme_offload_fraction: float = 0.0):
|
||||
optimizer = HybridAdam(model.parameters(), nvme_offload_fraction=nvme_offload_fraction)
|
||||
print(f'Model numel: {get_model_numel(model) / 1024**3:.3f} B')
|
||||
|
||||
gemini_config = dict(strict_ddp_mode=True, device=torch.cuda.current_device(),
|
||||
placement_policy='cpu', pin_memory=True, hidden_dim=config.n_embd)
|
||||
model = zero_model_wrapper(model, zero_stage=3, gemini_config=gemini_config)
|
||||
optimizer = zero_optim_wrapper(model, optimizer, initial_scale=2**5)
|
||||
plugin = GeminiPlugin(
|
||||
strict_ddp_mode=True,
|
||||
device=torch.cuda.current_device(),
|
||||
placement_policy='cpu',
|
||||
pin_memory=True,
|
||||
hidden_dim=config.n_embd,
|
||||
initial_scale=2**5
|
||||
)
|
||||
booster = Booster(plugin)
|
||||
model, optimizer, criterion, _* = booster.boost(model, optimizer, criterion)
|
||||
|
||||
start = time.time()
|
||||
for step in range(3):
|
||||
data = get_data(4, 128, config.vocab_size)
|
||||
outputs = model(**data)
|
||||
loss = criterion(outputs.logits, data['input_ids'])
|
||||
optimizer.backward(loss)
|
||||
booster.backward(loss, optimizer)
|
||||
optimizer.step()
|
||||
optimizer.zero_grad()
|
||||
print(f'[{step}] loss: {loss.item():.3f}')
|
||||
|
Reference in New Issue
Block a user