mirror of
https://github.com/hpcaitech/ColossalAI.git
synced 2025-09-05 11:02:05 +00:00
[hotfix] Fix examples no pad token & auto parallel codegen bug; (#5606)
* fix no pad token bug * fixed some auto parallel codegen bug, but might not run on torch 2.1 --------- Co-authored-by: Edenzzzz <wtan45@wisc.edu>
This commit is contained in:
@@ -5,7 +5,7 @@ import torch
|
||||
import torch.nn as nn
|
||||
|
||||
from colossalai.utils import _cast_float
|
||||
from colossalai.zero.legacy.gemini.tensor_utils import free_storage
|
||||
from colossalai.utils.common import free_storage
|
||||
|
||||
from .region_manager import RegionManager
|
||||
from .util import GlobalRuntimeInfo
|
||||
|
@@ -3,7 +3,8 @@ from typing import Dict, List, Tuple
|
||||
import torch
|
||||
from torch.fx import Node
|
||||
|
||||
from colossalai.zero.legacy.gemini.tensor_utils import alloc_storage, free_storage
|
||||
from colossalai.utils.common import free_storage
|
||||
from colossalai.zero.gemini.chunk.chunk import alloc_storage
|
||||
|
||||
|
||||
class Region:
|
||||
|
Reference in New Issue
Block a user