Skip to content

Commit

Permalink
[hotfix] skipped unsafe test cases (#1282)
Browse files Browse the repository at this point in the history
  • Loading branch information
FrankLeeeee authored Jul 12, 2022
1 parent 79fe7b0 commit 7e8114a
Show file tree
Hide file tree
Showing 14 changed files with 22 additions and 2 deletions.
2 changes: 2 additions & 0 deletions tests/test_fx/test_pipeline/test_hf_model/test_albert.py
Original file line number Diff line number Diff line change
@@ -1,11 +1,13 @@
import transformers
import torch
import pytest
from hf_utils import split_model_and_compare_output

BATCH_SIZE = 2
SEQ_LENGHT = 16


@pytest.mark.skip("error with pytorch 1.10")
def test_single_sentence_albert():
MODEL_LIST = [
transformers.AlbertModel,
Expand Down
2 changes: 2 additions & 0 deletions tests/test_fx/test_pipeline/test_hf_model/test_bert.py
Original file line number Diff line number Diff line change
@@ -1,11 +1,13 @@
import transformers
import torch
import pytest
from hf_utils import split_model_and_compare_output

BATCH_SIZE = 2
SEQ_LENGHT = 16


@pytest.mark.skip("error with pytorch 1.10")
def test_single_sentence_bert():
MODEL_LIST = [
transformers.BertModel,
Expand Down
2 changes: 2 additions & 0 deletions tests/test_fx/test_pipeline/test_hf_model/test_gpt.py
Original file line number Diff line number Diff line change
@@ -1,5 +1,6 @@
import transformers
import torch
import pytest
from hf_utils import split_model_and_compare_output

BATCH_SIZE = 64
Expand All @@ -8,6 +9,7 @@
NUM_CHUNKS = 1


@pytest.mark.skip("error with pytorch 1.10")
def test_gpt():
MODEL_LIST = [
transformers.GPT2Model,
Expand Down
1 change: 1 addition & 0 deletions tests/test_fx/test_pipeline/test_hf_model/test_opt.py
Original file line number Diff line number Diff line change
Expand Up @@ -7,6 +7,7 @@
SEQ_LENGHT = 16


@pytest.mark.skip("error with pytorch 1.10")
def test_opt():
MODEL_LIST = [
transformers.OPTModel,
Expand Down
1 change: 1 addition & 0 deletions tests/test_fx/test_pipeline/test_hf_model/test_t5.py
Original file line number Diff line number Diff line change
Expand Up @@ -16,6 +16,7 @@ def apex_fused_layernorm(self, input):
SEQ_LENGHT = 16


@pytest.mark.skip("error with pytorch 1.10")
def test_t5():
MODEL_LIST = [
transformers.T5Model,
Expand Down
2 changes: 2 additions & 0 deletions tests/test_fx/test_tracer/test_hf_model/test_hf_albert.py
Original file line number Diff line number Diff line change
@@ -1,5 +1,6 @@
import transformers
import torch
import pytest
from utils import trace_model_and_compare_output

BATCH_SIZE = 2
Expand Down Expand Up @@ -33,6 +34,7 @@ def data_gen():
trace_model_and_compare_output(model, data_gen)


@pytest.mark.skip("error with pytorch 1.10")
def test_multi_sentence_albert():
config = transformers.AlbertConfig(hidden_size=128,
num_hidden_layers=2,
Expand Down
2 changes: 2 additions & 0 deletions tests/test_fx/test_tracer/test_hf_model/test_hf_bert.py
Original file line number Diff line number Diff line change
@@ -1,5 +1,6 @@
import transformers
import torch
import pytest
from utils import trace_model_and_compare_output

BATCH_SIZE = 2
Expand Down Expand Up @@ -30,6 +31,7 @@ def data_gen():
trace_model_and_compare_output(model, data_gen)


@pytest.mark.skip("error with pytorch 1.10")
def test_multi_sentence_bert():
config = transformers.BertConfig(hidden_size=128, num_hidden_layers=2, num_attention_heads=4, intermediate_size=256)
tokenizer = transformers.BertTokenizer.from_pretrained("bert-base-uncased")
Expand Down
2 changes: 2 additions & 0 deletions tests/test_fx/test_tracer/test_hf_model/test_hf_gpt.py
Original file line number Diff line number Diff line change
@@ -1,11 +1,13 @@
import transformers
import torch
import pytest
from utils import trace_model_and_compare_output

BATCH_SIZE = 1
SEQ_LENGHT = 16


@pytest.mark.skip("error with pytorch 1.10")
def test_gpt():
MODEL_LIST = [
transformers.GPT2Model,
Expand Down
2 changes: 1 addition & 1 deletion tests/test_fx/test_tracer/test_hf_model/test_hf_opt.py
Original file line number Diff line number Diff line change
Expand Up @@ -7,7 +7,7 @@
SEQ_LENGHT = 16


@pytest.mark.skip('value is not aligned yet')
@pytest.mark.skip("error with pytorch 1.10")
def test_opt():
MODEL_LIST = [
transformers.OPTModel,
Expand Down
2 changes: 1 addition & 1 deletion tests/test_fx/test_tracer/test_hf_model/test_hf_t5.py
Original file line number Diff line number Diff line change
Expand Up @@ -17,7 +17,7 @@ def apex_fused_layernorm(self, input):
SEQ_LENGHT = 16


@pytest.mark.skip('value is not aligned yet')
@pytest.mark.skip("error with pytorch 1.10")
def test_t5():
MODEL_LIST = [
transformers.T5Model,
Expand Down
1 change: 1 addition & 0 deletions tests/test_utils/test_checkpoint/test_checkpoint_1d.py
Original file line number Diff line number Diff line change
Expand Up @@ -67,6 +67,7 @@ def check_checkpoint_1d(rank, world_size, port):


@pytest.mark.dist
@pytest.mark.skip("takes too long")
@skip_if_not_enough_gpus(min_gpus=8)
@rerun_on_exception(exception_type=mp.ProcessRaisedException, pattern=".*Address already in use.*")
def test_checkpoint_1d():
Expand Down
1 change: 1 addition & 0 deletions tests/test_utils/test_checkpoint/test_checkpoint_2d.py
Original file line number Diff line number Diff line change
Expand Up @@ -67,6 +67,7 @@ def check_checkpoint_2d(rank, world_size, port):


@pytest.mark.dist
@pytest.mark.skip("takes too long")
@skip_if_not_enough_gpus(min_gpus=8)
@rerun_on_exception(exception_type=mp.ProcessRaisedException, pattern=".*Address already in use.*")
def test_checkpoint_2d():
Expand Down
2 changes: 2 additions & 0 deletions tests/test_utils/test_checkpoint/test_checkpoint_2p5d.py
Original file line number Diff line number Diff line change
Expand Up @@ -37,6 +37,7 @@ def build_pipeline(model):
def check_equal(A, B):
assert torch.allclose(A, B, rtol=1e-3, atol=1e-2)


def check_checkpoint_2p5d(rank, world_size, port):
config = dict(parallel=dict(pipeline=dict(size=2), tensor=dict(size=4, depth=1, mode="2.5d")),)

Expand Down Expand Up @@ -66,6 +67,7 @@ def check_checkpoint_2p5d(rank, world_size, port):


@pytest.mark.dist
@pytest.mark.skip("takes too long")
@skip_if_not_enough_gpus(min_gpus=8)
@rerun_on_exception(exception_type=mp.ProcessRaisedException, pattern=".*Address already in use.*")
def test_checkpoint_2p5d():
Expand Down
2 changes: 2 additions & 0 deletions tests/test_utils/test_checkpoint/test_checkpoint_3d.py
Original file line number Diff line number Diff line change
Expand Up @@ -37,6 +37,7 @@ def build_pipeline(model):
def check_equal(A, B):
assert torch.allclose(A, B, rtol=1e-3, atol=1e-2)


def check_checkpoint_3d(rank, world_size, port):
config = dict(parallel=dict(pipeline=dict(size=1), tensor=dict(size=8, mode="3d")),)

Expand Down Expand Up @@ -66,6 +67,7 @@ def check_checkpoint_3d(rank, world_size, port):


@pytest.mark.dist
@pytest.mark.skip("takes too long")
@skip_if_not_enough_gpus(min_gpus=8)
@rerun_on_exception(exception_type=mp.ProcessRaisedException, pattern=".*Address already in use.*")
def test_checkpoint_3d():
Expand Down

0 comments on commit 7e8114a

Please sign in to comment.