[PT FE] Fix mask2former model marks in tests (#20717)
* [PT FE] Fix mask2former model marks in tests * Use better machine * Add more models * Update .github/workflows/linux.yml
This commit is contained in:
parent
cde757d66a
commit
f029ebb8e2
@ -10,6 +10,7 @@ albert-base-v2,albert
|
|||||||
AlekseyKorshuk/test_reward_model,reward_model,skip,Load problem
|
AlekseyKorshuk/test_reward_model,reward_model,skip,Load problem
|
||||||
alibaba-damo/mgp-str-base,mgp-str,xfail,Compile error: unsupported Einsum
|
alibaba-damo/mgp-str-base,mgp-str,xfail,Compile error: unsupported Einsum
|
||||||
allenai/hvila-block-layoutlm-finetuned-docbank,hierarchical_model,skip,Load problem
|
allenai/hvila-block-layoutlm-finetuned-docbank,hierarchical_model,skip,Load problem
|
||||||
|
allenai/longformer-base-4096,longformer
|
||||||
ameya772/sentence-t5-base-atis-fine-tuned,T5,skip,Load problem
|
ameya772/sentence-t5-base-atis-fine-tuned,T5,skip,Load problem
|
||||||
andreasmadsen/efficient_mlm_m0.40,roberta-prelayernorm
|
andreasmadsen/efficient_mlm_m0.40,roberta-prelayernorm
|
||||||
anton-l/emformer-base-librispeech,emformer,skip,Load problem
|
anton-l/emformer-base-librispeech,emformer,skip,Load problem
|
||||||
@ -71,7 +72,7 @@ facebook/esm2_t6_8M_UR50D,esm
|
|||||||
facebook/flava-full,flava,xfail,Tracing problem
|
facebook/flava-full,flava,xfail,Tracing problem
|
||||||
facebook/flava-image-codebook,flava_image_codebook,skip,Load problem
|
facebook/flava-image-codebook,flava_image_codebook,skip,Load problem
|
||||||
facebook/m2m100_418M,m2m_100
|
facebook/m2m100_418M,m2m_100
|
||||||
facebook/mask2former-swin-base-coco-panoptic,mask2former,xfail,Accuracy validation failed
|
facebook/mask2former-swin-base-coco-panoptic,mask2former
|
||||||
facebook/maskformer-swin-base-coco,maskformer
|
facebook/maskformer-swin-base-coco,maskformer
|
||||||
facebook/mms-tts-eng,vits,skip,Load problem
|
facebook/mms-tts-eng,vits,skip,Load problem
|
||||||
facebook/musicgen-small,musicgen,skip,Load problem
|
facebook/musicgen-small,musicgen,skip,Load problem
|
||||||
@ -92,6 +93,7 @@ Geor111y/flair-ner-addresses-extractor,flair,skip,Load problem
|
|||||||
gia-project/gia,gia,skip,Load problem
|
gia-project/gia,gia,skip,Load problem
|
||||||
gokuls/bert_12_layer_model_v1,hybridbert,skip,Load problem
|
gokuls/bert_12_layer_model_v1,hybridbert,skip,Load problem
|
||||||
google/bigbird-roberta-base,big_bird
|
google/bigbird-roberta-base,big_bird
|
||||||
|
google/bigbird-pegasus-large-arxiv,bigbird-pegasus
|
||||||
google/bit-50,bit
|
google/bit-50,bit
|
||||||
google/canine-s,canine,xfail,aten::slice: Parameter axis 3 out of the tensor rank range
|
google/canine-s,canine,xfail,aten::slice: Parameter axis 3 out of the tensor rank range
|
||||||
google/efficientnet-b2,efficientnet,xfail,Compile error: AvgPool: Kernel after dilation has size (dim: 1408) larger than the data shape after padding (dim: 9) at axis 0.
|
google/efficientnet-b2,efficientnet,xfail,Compile error: AvgPool: Kernel after dilation has size (dim: 1408) larger than the data shape after padding (dim: 9) at axis 0.
|
||||||
@ -105,7 +107,7 @@ google/owlvit-base-patch32,owlvit
|
|||||||
google/pix2struct-docvqa-base,pix2struct,skip,Load problem
|
google/pix2struct-docvqa-base,pix2struct,skip,Load problem
|
||||||
google/realm-orqa-nq-openqa,realm,skip,Load problem
|
google/realm-orqa-nq-openqa,realm,skip,Load problem
|
||||||
google/reformer-crime-and-punishment,reformer,xfail,Tracing problem
|
google/reformer-crime-and-punishment,reformer,xfail,Tracing problem
|
||||||
google/tapas-large-finetuned-wtq,tapas,skip,Load problem
|
google/tapas-large-finetuned-wtq,tapas
|
||||||
google/vit-hybrid-base-bit-384,vit-hybrid,skip,Load problem
|
google/vit-hybrid-base-bit-384,vit-hybrid,skip,Load problem
|
||||||
google/vivit-b-16x2-kinetics400,vivit
|
google/vivit-b-16x2-kinetics400,vivit
|
||||||
Goutham-Vignesh/ContributionSentClassification-scibert,scibert,skip,Load problem
|
Goutham-Vignesh/ContributionSentClassification-scibert,scibert,skip,Load problem
|
||||||
@ -300,6 +302,7 @@ pie/example-re-textclf-tacred,TransformerTextClassificationModel,skip,Load probl
|
|||||||
pleisto/yuren-baichuan-7b,multimodal_llama,skip,Load problem
|
pleisto/yuren-baichuan-7b,multimodal_llama,skip,Load problem
|
||||||
predictia/europe_reanalysis_downscaler_convbaseline,convbilinear,skip,Load problem
|
predictia/europe_reanalysis_downscaler_convbaseline,convbilinear,skip,Load problem
|
||||||
predictia/europe_reanalysis_downscaler_convswin2sr,conv_swin2sr,skip,Load problem
|
predictia/europe_reanalysis_downscaler_convswin2sr,conv_swin2sr,skip,Load problem
|
||||||
|
pszemraj/led-large-book-summary,led
|
||||||
qmeeus/whisper-small-ner-combined,whisper_for_slu,skip,Load problem
|
qmeeus/whisper-small-ner-combined,whisper_for_slu,skip,Load problem
|
||||||
raman-ai/pcqv2-tokengt-lap16,tokengt,skip,Load problem
|
raman-ai/pcqv2-tokengt-lap16,tokengt,skip,Load problem
|
||||||
range3/pegasus-gpt2-medium,pegasusgpt2,skip,Load problem
|
range3/pegasus-gpt2-medium,pegasusgpt2,skip,Load problem
|
||||||
|
@ -276,7 +276,9 @@ class TestTransformersModel(TestConvertModel):
|
|||||||
return [i.numpy() for i in self.example]
|
return [i.numpy() for i in self.example]
|
||||||
|
|
||||||
def convert_model(self, model_obj):
|
def convert_model(self, model_obj):
|
||||||
ov_model = convert_model(model_obj, example_input=self.example)
|
ov_model = convert_model(model_obj,
|
||||||
|
example_input=self.example,
|
||||||
|
verbose=True)
|
||||||
return ov_model
|
return ov_model
|
||||||
|
|
||||||
def infer_fw_model(self, model_obj, inputs):
|
def infer_fw_model(self, model_obj, inputs):
|
||||||
@ -297,8 +299,7 @@ class TestTransformersModel(TestConvertModel):
|
|||||||
("google/flan-t5-base", "t5"),
|
("google/flan-t5-base", "t5"),
|
||||||
("google/tapas-large-finetuned-wtq", "tapas"),
|
("google/tapas-large-finetuned-wtq", "tapas"),
|
||||||
("gpt2", "gpt2"),
|
("gpt2", "gpt2"),
|
||||||
("openai/clip-vit-large-patch14", "clip"),
|
("openai/clip-vit-large-patch14", "clip")
|
||||||
("facebook/xmod-base","xmod")
|
|
||||||
])
|
])
|
||||||
@pytest.mark.precommit
|
@pytest.mark.precommit
|
||||||
def test_convert_model_precommit(self, name, type, ie_device):
|
def test_convert_model_precommit(self, name, type, ie_device):
|
||||||
|
Loading…
Reference in New Issue
Block a user