Mark models according to last validation and fix some problems in models (#20324)
* Mark models according to last validation and fix some problems in models * Update tests/model_hub_tests/torch_tests/hf_transformers_models
This commit is contained in:
parent
b630bffa14
commit
0fec05ecf2
@ -1,4 +1,4 @@
|
|||||||
COCO-Detection/fast_rcnn_R_50_FPN_1x,none
|
COCO-Detection/fast_rcnn_R_50_FPN_1x,none,xfail,Tracing problem
|
||||||
COCO-Detection/faster_rcnn_R_101_C4_3x,none
|
COCO-Detection/faster_rcnn_R_101_C4_3x,none
|
||||||
COCO-Detection/faster_rcnn_R_101_DC5_3x,none
|
COCO-Detection/faster_rcnn_R_101_DC5_3x,none
|
||||||
COCO-Detection/faster_rcnn_R_101_FPN_3x,none
|
COCO-Detection/faster_rcnn_R_101_FPN_3x,none
|
||||||
|
@ -3,7 +3,7 @@ abcp4/mymodel-test,mymodel,skip,Load problem
|
|||||||
abeja/gpt-neox-japanese-2.7b,gpt_neox_japanese
|
abeja/gpt-neox-japanese-2.7b,gpt_neox_japanese
|
||||||
acl-submission-anonym/EAM-spectral,examuse,skip,Load problem
|
acl-submission-anonym/EAM-spectral,examuse,skip,Load problem
|
||||||
adalbertojunior/modular-test,modular,skip,Load problem
|
adalbertojunior/modular-test,modular,skip,Load problem
|
||||||
aerner/lm-v2,open-llama
|
aerner/lm-v2,open-llama,xfail,Example input problem
|
||||||
afonsosamarques/ardt-vanilla-combo_train_hopper_v2-2508_1336-33,decision_transformer,xfail,Tracing problem
|
afonsosamarques/ardt-vanilla-combo_train_hopper_v2-2508_1336-33,decision_transformer,xfail,Tracing problem
|
||||||
aihijo/gec-zh-gector-bert-large,gector,skip,Load problem
|
aihijo/gec-zh-gector-bert-large,gector,skip,Load problem
|
||||||
albert-base-v2,albert
|
albert-base-v2,albert
|
||||||
@ -21,7 +21,7 @@ ArthurZ/jukebox-vqvae,jukebox_vqvae,skip,Load problem
|
|||||||
ArthurZ/persimmon-8b-base,persimmon,skip,Load problem
|
ArthurZ/persimmon-8b-base,persimmon,skip,Load problem
|
||||||
ashishpatel26/span-marker-bert-base-fewnerd-coarse-super,span-marker,skip,Load problem
|
ashishpatel26/span-marker-bert-base-fewnerd-coarse-super,span-marker,skip,Load problem
|
||||||
asi/albert-act-tiny,albert_act,skip,Load problem
|
asi/albert-act-tiny,albert_act,skip,Load problem
|
||||||
BAAI/AltCLIP,altclip,xfail,Unsupported op aten::numpy_T
|
BAAI/AltCLIP,altclip
|
||||||
BAAI/AquilaCode-py,aquila,skip,Load problem
|
BAAI/AquilaCode-py,aquila,skip,Load problem
|
||||||
bana513/opennmt-translator-en-hu,opennmt-translator,skip,Load problem
|
bana513/opennmt-translator-en-hu,opennmt-translator,skip,Load problem
|
||||||
benjamin/wtp-bert-mini,bert-char,skip,Load problem
|
benjamin/wtp-bert-mini,bert-char,skip,Load problem
|
||||||
@ -79,7 +79,7 @@ facebook/musicgen-small,musicgen,skip,Load problem
|
|||||||
facebook/opt-125m,opt
|
facebook/opt-125m,opt
|
||||||
facebook/rag-token-nq,rag,skip,Load problem
|
facebook/rag-token-nq,rag,skip,Load problem
|
||||||
facebook/sam-vit-large,sam,xfail,No node with name original_sizes
|
facebook/sam-vit-large,sam,xfail,No node with name original_sizes
|
||||||
facebook/timesformer-base-finetuned-k400,timesformer,xfail,Shape inference of Add node failed: Eltwise shape infer input shapes dim index: 1 mismatch
|
facebook/timesformer-base-finetuned-k400,timesformer
|
||||||
facebook/vit-mae-base,vit_mae,xfail,Accuracy validation failed
|
facebook/vit-mae-base,vit_mae,xfail,Accuracy validation failed
|
||||||
facebook/wmt19-ru-en,fsmt,xfail,Tracing problem
|
facebook/wmt19-ru-en,fsmt,xfail,Tracing problem
|
||||||
facebook/xlm-roberta-xl,xlm-roberta-xl
|
facebook/xlm-roberta-xl,xlm-roberta-xl
|
||||||
@ -199,7 +199,7 @@ kiddothe2b/hierarchical-transformer-base-4096-v2,hat,skip,Load problem
|
|||||||
k-l-lambda/clip-text-generator,clip_text_generator,skip,Load problem
|
k-l-lambda/clip-text-generator,clip_text_generator,skip,Load problem
|
||||||
k-l-lambda/stable-diffusion-v1-4-inv-embed,inv_word_embed,skip,Load problem
|
k-l-lambda/stable-diffusion-v1-4-inv-embed,inv_word_embed,skip,Load problem
|
||||||
KoboldAI/fairseq-dense-13B-Janeway,xglm,skip,Large Model
|
KoboldAI/fairseq-dense-13B-Janeway,xglm,skip,Large Model
|
||||||
konverner/qdq-camembert-apolliner,qdqbert
|
konverner/qdq-camembert-apolliner,qdqbert,xfail,Repository not found
|
||||||
krasserm/perceiver-ar-clm-base,perceiver-ar-causal-language-model,skip,Load problem
|
krasserm/perceiver-ar-clm-base,perceiver-ar-causal-language-model,skip,Load problem
|
||||||
krasserm/perceiver-ar-sam-giant-midi,perceiver-ar-symbolic-audio-model,skip,Load problem
|
krasserm/perceiver-ar-sam-giant-midi,perceiver-ar-symbolic-audio-model,skip,Load problem
|
||||||
krasserm/perceiver-io-img-clf,perceiver-io-image-classifier,skip,Load problem
|
krasserm/perceiver-io-img-clf,perceiver-io-image-classifier,skip,Load problem
|
||||||
@ -309,7 +309,7 @@ regisss/bridgetower-newyorker-a100-8x,bridgetower
|
|||||||
rinna/japanese-cloob-vit-b-16,cloob,skip,Load problem
|
rinna/japanese-cloob-vit-b-16,cloob,skip,Load problem
|
||||||
Rocketknight1/tiny-random-falcon-7b,falcon
|
Rocketknight1/tiny-random-falcon-7b,falcon
|
||||||
RUCAIBox/mass-base-uncased,mass,skip,Load problem
|
RUCAIBox/mass-base-uncased,mass,skip,Load problem
|
||||||
RWKV/rwkv-4-169m-pile,rwkv,xfail,Unsupported op aten::maximum
|
RWKV/rwkv-4-169m-pile,rwkv
|
||||||
sahasrarjn/interbert,BERT,skip,Load problem
|
sahasrarjn/interbert,BERT,skip,Load problem
|
||||||
saibo/genkalm-medium-gpt2,genkalm,skip,Load problem
|
saibo/genkalm-medium-gpt2,genkalm,skip,Load problem
|
||||||
SajjadAyoubi/clip-fa-vision,clip_vision_model
|
SajjadAyoubi/clip-fa-vision,clip_vision_model
|
||||||
|
@ -20,7 +20,7 @@ class TestDetectron2ConvertModel(TestConvertModel):
|
|||||||
|
|
||||||
def load_model(self, model_name, model_link):
|
def load_model(self, model_name, model_link):
|
||||||
from detectron2 import model_zoo, export
|
from detectron2 import model_zoo, export
|
||||||
from detectron2.modeling import build_model
|
from detectron2.modeling import build_model, PanopticFPN
|
||||||
from detectron2.checkpoint import DetectionCheckpointer
|
from detectron2.checkpoint import DetectionCheckpointer
|
||||||
from detectron2.config import CfgNode
|
from detectron2.config import CfgNode
|
||||||
import torchvision.transforms as transforms
|
import torchvision.transforms as transforms
|
||||||
@ -32,13 +32,23 @@ class TestDetectron2ConvertModel(TestConvertModel):
|
|||||||
assert isinstance(cfg, CfgNode), "Unexpected config"
|
assert isinstance(cfg, CfgNode), "Unexpected config"
|
||||||
cfg.MODEL.DEVICE = "cpu"
|
cfg.MODEL.DEVICE = "cpu"
|
||||||
model = build_model(cfg)
|
model = build_model(cfg)
|
||||||
DetectionCheckpointer(model, save_to_disk=False).load(cfg.MODEL.WEIGHTS)
|
DetectionCheckpointer(
|
||||||
|
model, save_to_disk=False).load(cfg.MODEL.WEIGHTS)
|
||||||
|
|
||||||
model.eval()
|
model.eval()
|
||||||
inputs = [{"image": image,
|
inputs = [{"image": image,
|
||||||
"height": torch.tensor(image.shape[1]),
|
"height": torch.tensor(image.shape[1]),
|
||||||
"width": torch.tensor(image.shape[2])}]
|
"width": torch.tensor(image.shape[2])}]
|
||||||
adapter = export.TracingAdapter(model, inputs)
|
# https://github.com/facebookresearch/detectron2/blob/4e80df1e58901557e2824ce3b488d30209a9be33/tools/deploy/export_model.py#L123
|
||||||
|
# This is done only for Panoptic models, but it may be incorrect to do that, because one of outputs of panoptic model is getting lost
|
||||||
|
if isinstance(model, PanopticFPN):
|
||||||
|
def inference(model, inputs):
|
||||||
|
# use do_postprocess=False so it returns ROI mask
|
||||||
|
inst = model.inference(inputs, do_postprocess=False)[0]
|
||||||
|
return [{"instances": inst}]
|
||||||
|
else:
|
||||||
|
inference = None # assume that we just call the model directly
|
||||||
|
adapter = export.TracingAdapter(model, inputs, inference)
|
||||||
|
|
||||||
self.example = adapter.flattened_inputs
|
self.example = adapter.flattened_inputs
|
||||||
return adapter
|
return adapter
|
||||||
@ -75,7 +85,8 @@ class TestDetectron2ConvertModel(TestConvertModel):
|
|||||||
cur_fw_res = fw_outputs[i]
|
cur_fw_res = fw_outputs[i]
|
||||||
cur_ov_res = ov_outputs[i]
|
cur_ov_res = ov_outputs[i]
|
||||||
l = min(len(cur_fw_res), len(cur_ov_res))
|
l = min(len(cur_fw_res), len(cur_ov_res))
|
||||||
assert l > 0 or len(cur_fw_res) == len(cur_ov_res), "No boxes were selected."
|
assert l > 0 or len(cur_fw_res) == len(
|
||||||
|
cur_ov_res), "No boxes were selected."
|
||||||
print(f"fw_re: {cur_fw_res};\n ov_res: {cur_ov_res}")
|
print(f"fw_re: {cur_fw_res};\n ov_res: {cur_ov_res}")
|
||||||
is_ok = compare_two_tensors(cur_ov_res[:l], cur_fw_res[:l], fw_eps)
|
is_ok = compare_two_tensors(cur_ov_res[:l], cur_fw_res[:l], fw_eps)
|
||||||
assert is_ok, "Accuracy validation failed"
|
assert is_ok, "Accuracy validation failed"
|
||||||
@ -86,8 +97,8 @@ class TestDetectron2ConvertModel(TestConvertModel):
|
|||||||
def test_detectron2_precommit(self, name, type, mark, reason, ie_device):
|
def test_detectron2_precommit(self, name, type, mark, reason, ie_device):
|
||||||
self.run(name, None, ie_device)
|
self.run(name, None, ie_device)
|
||||||
|
|
||||||
@pytest.mark.parametrize("name,type,mark,reason",
|
@pytest.mark.parametrize("name",
|
||||||
get_models_list(os.path.join(os.path.dirname(__file__), "detectron2_models")))
|
[pytest.param(n, marks=pytest.mark.xfail(reason=r)) if m == "xfail" else n for n, _, m, r in get_models_list(os.path.join(os.path.dirname(__file__), "detectron2_models"))])
|
||||||
@pytest.mark.nightly
|
@pytest.mark.nightly
|
||||||
def test_detectron2_all_models(self, name, type, mark, reason, ie_device):
|
def test_detectron2_all_models(self, name, ie_device):
|
||||||
self.run(name, None, ie_device)
|
self.run(name, None, ie_device)
|
||||||
|
@ -114,7 +114,7 @@ class TestTorchHubConvertModel(TestConvertModel):
|
|||||||
self.run(model_name, None, ie_device)
|
self.run(model_name, None, ie_device)
|
||||||
|
|
||||||
@pytest.mark.parametrize("name",
|
@pytest.mark.parametrize("name",
|
||||||
[pytest.param(n, marks=pytest.mark.xfail) if m == "xfail" else n for n, _, m, r in get_models_list(os.path.join(os.path.dirname(__file__), "torchvision_models"))])
|
[pytest.param(n, marks=pytest.mark.xfail(reason=r)) if m == "xfail" else n for n, _, m, r in get_models_list(os.path.join(os.path.dirname(__file__), "torchvision_models"))])
|
||||||
@pytest.mark.nightly
|
@pytest.mark.nightly
|
||||||
def test_convert_model_all_models(self, name, ie_device):
|
def test_convert_model_all_models(self, name, ie_device):
|
||||||
self.run(name, None, ie_device)
|
self.run(name, None, ie_device)
|
||||||
|
Loading…
Reference in New Issue
Block a user