[PT FE] Fix mask2former model marks in tests (#20717)
* [PT FE] Fix mask2former model marks in tests * Use better machine * Add more models * Update .github/workflows/linux.yml
This commit is contained in:
parent
cde757d66a
commit
f029ebb8e2
@ -10,6 +10,7 @@ albert-base-v2,albert
|
||||
AlekseyKorshuk/test_reward_model,reward_model,skip,Load problem
|
||||
alibaba-damo/mgp-str-base,mgp-str,xfail,Compile error: unsupported Einsum
|
||||
allenai/hvila-block-layoutlm-finetuned-docbank,hierarchical_model,skip,Load problem
|
||||
allenai/longformer-base-4096,longformer
|
||||
ameya772/sentence-t5-base-atis-fine-tuned,T5,skip,Load problem
|
||||
andreasmadsen/efficient_mlm_m0.40,roberta-prelayernorm
|
||||
anton-l/emformer-base-librispeech,emformer,skip,Load problem
|
||||
@ -71,7 +72,7 @@ facebook/esm2_t6_8M_UR50D,esm
|
||||
facebook/flava-full,flava,xfail,Tracing problem
|
||||
facebook/flava-image-codebook,flava_image_codebook,skip,Load problem
|
||||
facebook/m2m100_418M,m2m_100
|
||||
facebook/mask2former-swin-base-coco-panoptic,mask2former,xfail,Accuracy validation failed
|
||||
facebook/mask2former-swin-base-coco-panoptic,mask2former
|
||||
facebook/maskformer-swin-base-coco,maskformer
|
||||
facebook/mms-tts-eng,vits,skip,Load problem
|
||||
facebook/musicgen-small,musicgen,skip,Load problem
|
||||
@ -92,6 +93,7 @@ Geor111y/flair-ner-addresses-extractor,flair,skip,Load problem
|
||||
gia-project/gia,gia,skip,Load problem
|
||||
gokuls/bert_12_layer_model_v1,hybridbert,skip,Load problem
|
||||
google/bigbird-roberta-base,big_bird
|
||||
google/bigbird-pegasus-large-arxiv,bigbird-pegasus
|
||||
google/bit-50,bit
|
||||
google/canine-s,canine,xfail,aten::slice: Parameter axis 3 out of the tensor rank range
|
||||
google/efficientnet-b2,efficientnet,xfail,Compile error: AvgPool: Kernel after dilation has size (dim: 1408) larger than the data shape after padding (dim: 9) at axis 0.
|
||||
@ -105,7 +107,7 @@ google/owlvit-base-patch32,owlvit
|
||||
google/pix2struct-docvqa-base,pix2struct,skip,Load problem
|
||||
google/realm-orqa-nq-openqa,realm,skip,Load problem
|
||||
google/reformer-crime-and-punishment,reformer,xfail,Tracing problem
|
||||
google/tapas-large-finetuned-wtq,tapas,skip,Load problem
|
||||
google/tapas-large-finetuned-wtq,tapas
|
||||
google/vit-hybrid-base-bit-384,vit-hybrid,skip,Load problem
|
||||
google/vivit-b-16x2-kinetics400,vivit
|
||||
Goutham-Vignesh/ContributionSentClassification-scibert,scibert,skip,Load problem
|
||||
@ -300,6 +302,7 @@ pie/example-re-textclf-tacred,TransformerTextClassificationModel,skip,Load probl
|
||||
pleisto/yuren-baichuan-7b,multimodal_llama,skip,Load problem
|
||||
predictia/europe_reanalysis_downscaler_convbaseline,convbilinear,skip,Load problem
|
||||
predictia/europe_reanalysis_downscaler_convswin2sr,conv_swin2sr,skip,Load problem
|
||||
pszemraj/led-large-book-summary,led
|
||||
qmeeus/whisper-small-ner-combined,whisper_for_slu,skip,Load problem
|
||||
raman-ai/pcqv2-tokengt-lap16,tokengt,skip,Load problem
|
||||
range3/pegasus-gpt2-medium,pegasusgpt2,skip,Load problem
|
||||
|
@ -276,7 +276,9 @@ class TestTransformersModel(TestConvertModel):
|
||||
return [i.numpy() for i in self.example]
|
||||
|
||||
def convert_model(self, model_obj):
|
||||
ov_model = convert_model(model_obj, example_input=self.example)
|
||||
ov_model = convert_model(model_obj,
|
||||
example_input=self.example,
|
||||
verbose=True)
|
||||
return ov_model
|
||||
|
||||
def infer_fw_model(self, model_obj, inputs):
|
||||
@ -297,8 +299,7 @@ class TestTransformersModel(TestConvertModel):
|
||||
("google/flan-t5-base", "t5"),
|
||||
("google/tapas-large-finetuned-wtq", "tapas"),
|
||||
("gpt2", "gpt2"),
|
||||
("openai/clip-vit-large-patch14", "clip"),
|
||||
("facebook/xmod-base","xmod")
|
||||
("openai/clip-vit-large-patch14", "clip")
|
||||
])
|
||||
@pytest.mark.precommit
|
||||
def test_convert_model_precommit(self, name, type, ie_device):
|
||||
|
Loading…
Reference in New Issue
Block a user