From f029ebb8e240690209e4b97360463b30350216e6 Mon Sep 17 00:00:00 2001 From: Maxim Vafin Date: Fri, 27 Oct 2023 19:04:30 +0200 Subject: [PATCH] [PT FE] Fix mask2former model marks in tests (#20717) * [PT FE] Fix mask2former model marks in tests * Use better machine * Add more models * Update .github/workflows/linux.yml --- tests/model_hub_tests/torch_tests/hf_transformers_models | 7 +++++-- tests/model_hub_tests/torch_tests/test_hf_transformers.py | 7 ++++--- 2 files changed, 9 insertions(+), 5 deletions(-) diff --git a/tests/model_hub_tests/torch_tests/hf_transformers_models b/tests/model_hub_tests/torch_tests/hf_transformers_models index 31a24b681eb..dd41a18235b 100644 --- a/tests/model_hub_tests/torch_tests/hf_transformers_models +++ b/tests/model_hub_tests/torch_tests/hf_transformers_models @@ -10,6 +10,7 @@ albert-base-v2,albert AlekseyKorshuk/test_reward_model,reward_model,skip,Load problem alibaba-damo/mgp-str-base,mgp-str,xfail,Compile error: unsupported Einsum allenai/hvila-block-layoutlm-finetuned-docbank,hierarchical_model,skip,Load problem +allenai/longformer-base-4096,longformer ameya772/sentence-t5-base-atis-fine-tuned,T5,skip,Load problem andreasmadsen/efficient_mlm_m0.40,roberta-prelayernorm anton-l/emformer-base-librispeech,emformer,skip,Load problem @@ -71,7 +72,7 @@ facebook/esm2_t6_8M_UR50D,esm facebook/flava-full,flava,xfail,Tracing problem facebook/flava-image-codebook,flava_image_codebook,skip,Load problem facebook/m2m100_418M,m2m_100 -facebook/mask2former-swin-base-coco-panoptic,mask2former,xfail,Accuracy validation failed +facebook/mask2former-swin-base-coco-panoptic,mask2former facebook/maskformer-swin-base-coco,maskformer facebook/mms-tts-eng,vits,skip,Load problem facebook/musicgen-small,musicgen,skip,Load problem @@ -92,6 +93,7 @@ Geor111y/flair-ner-addresses-extractor,flair,skip,Load problem gia-project/gia,gia,skip,Load problem gokuls/bert_12_layer_model_v1,hybridbert,skip,Load problem google/bigbird-roberta-base,big_bird +google/bigbird-pegasus-large-arxiv,bigbird-pegasus google/bit-50,bit google/canine-s,canine,xfail,aten::slice: Parameter axis 3 out of the tensor rank range google/efficientnet-b2,efficientnet,xfail,Compile error: AvgPool: Kernel after dilation has size (dim: 1408) larger than the data shape after padding (dim: 9) at axis 0. @@ -105,7 +107,7 @@ google/owlvit-base-patch32,owlvit google/pix2struct-docvqa-base,pix2struct,skip,Load problem google/realm-orqa-nq-openqa,realm,skip,Load problem google/reformer-crime-and-punishment,reformer,xfail,Tracing problem -google/tapas-large-finetuned-wtq,tapas,skip,Load problem +google/tapas-large-finetuned-wtq,tapas google/vit-hybrid-base-bit-384,vit-hybrid,skip,Load problem google/vivit-b-16x2-kinetics400,vivit Goutham-Vignesh/ContributionSentClassification-scibert,scibert,skip,Load problem @@ -300,6 +302,7 @@ pie/example-re-textclf-tacred,TransformerTextClassificationModel,skip,Load probl pleisto/yuren-baichuan-7b,multimodal_llama,skip,Load problem predictia/europe_reanalysis_downscaler_convbaseline,convbilinear,skip,Load problem predictia/europe_reanalysis_downscaler_convswin2sr,conv_swin2sr,skip,Load problem +pszemraj/led-large-book-summary,led qmeeus/whisper-small-ner-combined,whisper_for_slu,skip,Load problem raman-ai/pcqv2-tokengt-lap16,tokengt,skip,Load problem range3/pegasus-gpt2-medium,pegasusgpt2,skip,Load problem diff --git a/tests/model_hub_tests/torch_tests/test_hf_transformers.py b/tests/model_hub_tests/torch_tests/test_hf_transformers.py index 8b595e54256..8e3ea5ecfdd 100644 --- a/tests/model_hub_tests/torch_tests/test_hf_transformers.py +++ b/tests/model_hub_tests/torch_tests/test_hf_transformers.py @@ -276,7 +276,9 @@ class TestTransformersModel(TestConvertModel): return [i.numpy() for i in self.example] def convert_model(self, model_obj): - ov_model = convert_model(model_obj, example_input=self.example) + ov_model = convert_model(model_obj, + example_input=self.example, + verbose=True) return ov_model def infer_fw_model(self, model_obj, inputs): @@ -297,8 +299,7 @@ class TestTransformersModel(TestConvertModel): ("google/flan-t5-base", "t5"), ("google/tapas-large-finetuned-wtq", "tapas"), ("gpt2", "gpt2"), - ("openai/clip-vit-large-patch14", "clip"), - ("facebook/xmod-base","xmod") + ("openai/clip-vit-large-patch14", "clip") ]) @pytest.mark.precommit def test_convert_model_precommit(self, name, type, ie_device):