mirror of
https://github.com/hpcaitech/ColossalAI.git
synced 2025-09-16 14:41:53 +00:00
[fx] fixed unit tests for torch 1.12 (#1327)
This commit is contained in:
@@ -34,7 +34,6 @@ def test_single_sentence_albert():
|
||||
trace_model_and_compare_output(model, data_gen)
|
||||
|
||||
|
||||
@pytest.mark.skip("error with pytorch 1.10")
|
||||
def test_multi_sentence_albert():
|
||||
config = transformers.AlbertConfig(hidden_size=128,
|
||||
num_hidden_layers=2,
|
||||
|
@@ -31,7 +31,6 @@ def test_single_sentence_bert():
|
||||
trace_model_and_compare_output(model, data_gen)
|
||||
|
||||
|
||||
@pytest.mark.skip("error with pytorch 1.10")
|
||||
def test_multi_sentence_bert():
|
||||
config = transformers.BertConfig(hidden_size=128, num_hidden_layers=2, num_attention_heads=4, intermediate_size=256)
|
||||
tokenizer = transformers.BertTokenizer.from_pretrained("bert-base-uncased")
|
||||
|
@@ -7,7 +7,6 @@ BATCH_SIZE = 1
|
||||
SEQ_LENGHT = 16
|
||||
|
||||
|
||||
@pytest.mark.skip("error with pytorch 1.10")
|
||||
def test_gpt():
|
||||
MODEL_LIST = [
|
||||
transformers.GPT2Model,
|
||||
|
@@ -7,7 +7,6 @@ BATCH_SIZE = 1
|
||||
SEQ_LENGHT = 16
|
||||
|
||||
|
||||
@pytest.mark.skip("error with pytorch 1.10")
|
||||
def test_opt():
|
||||
MODEL_LIST = [
|
||||
transformers.OPTModel,
|
||||
|
@@ -7,7 +7,6 @@ BATCH_SIZE = 1
|
||||
SEQ_LENGHT = 16
|
||||
|
||||
|
||||
@pytest.mark.skip("error with pytorch 1.10")
|
||||
def test_t5():
|
||||
MODEL_LIST = [
|
||||
transformers.T5Model,
|
||||
|
Reference in New Issue
Block a user