@@ -164,7 +164,6 @@ def init_model_configs():
164164 "transformers" ,
165165 "AutoModelForImageTextToText" ,
166166 )
167-
168167 TasksManager ._CUSTOM_CLASSES [("pt" , "llava_next_video" , "image-text-to-text" )] = (
169168 "transformers" ,
170169 "AutoModelForVision2Seq" ,
@@ -300,21 +299,46 @@ def patch_model_for_export(
300299 return Qwen2MoEPatcher (self , model , model_kwargs = model_kwargs )
301300
302301
303- @register_in_tasks_manager ("qwen3" , * ["text-generation" , "text-generation-with-past" ], library_name = "transformers" )
302+ @register_in_tasks_manager (
303+ "qwen3" ,
304+ * [
305+ "text-generation" ,
306+ "text-generation-with-past" ,
307+ "feature-extraction" ,
308+ "feature-extraction-with-past" ,
309+ "text-classification" ,
310+ ],
311+ library_name = "transformers" ,
312+ )
304313class Qwen3OpenVINOConfig (TextDecoderWithPositionIdsOnnxConfig ):
305314 MIN_TRANSFORMERS_VERSION = "4.51.0"
306315
307316 DUMMY_INPUT_GENERATOR_CLASSES = (DummyTextInputGenerator , GemmaDummyPastKeyValuesGenerator )
308317 DUMMY_PKV_GENERATOR_CLASS = GemmaDummyPastKeyValuesGenerator
309318 NORMALIZED_CONFIG_CLASS = NormalizedTextConfig
310319
320+ @property
321+ def inputs (self ) -> Dict [str , Dict [int , str ]]:
322+ if self .task in ["feature-extraction" ]:
323+ common_inputs = {
324+ "input_ids" : {0 : "batch_size" , 1 : "sequence_length" },
325+ "attention_mask" : {0 : "batch_size" , 1 : "sequence_length" },
326+ }
327+ else :
328+ common_inputs = super ().inputs
329+ return common_inputs
330+
311331 def patch_model_for_export (
312332 self , model : Union ["PreTrainedModel" , "TFPreTrainedModel" ], model_kwargs : Optional [Dict [str , Any ]] = None
313333 ) -> "ModelPatcher" :
314334 return OVDecoderModelPatcher (self , model , model_kwargs = model_kwargs )
315335
316336
317- @register_in_tasks_manager ("qwen3_moe" , * ["text-generation" , "text-generation-with-past" ], library_name = "transformers" )
337+ @register_in_tasks_manager (
338+ "qwen3_moe" ,
339+ * ["text-generation" , "text-generation-with-past" , "feature-extraction" , "feature-extraction-with-past" ],
340+ library_name = "transformers" ,
341+ )
318342class Qwen3MoEOpenVINOConfig (Qwen3OpenVINOConfig ):
319343 def patch_model_for_export (
320344 self , model : Union ["PreTrainedModel" , "TFPreTrainedModel" ], model_kwargs : Optional [Dict [str , Any ]] = None
@@ -3501,7 +3525,11 @@ class Qwen2VLConfigBehavior(str, enum.Enum):
35013525 TEXT_EMBEDDINGS = "text_embeddings"
35023526
35033527
3504- @register_in_tasks_manager ("qwen2_vl" , * ["image-text-to-text" , "video-text-to-text" ], library_name = "transformers" )
3528+ @register_in_tasks_manager (
3529+ "qwen2_vl" ,
3530+ * ["image-text-to-text" , "video-text-to-text" ],
3531+ library_name = "transformers" ,
3532+ )
35053533class Qwen2VLOpenVINOConfig (BaseVLMOpenVINOConfig ):
35063534 SUPPORTED_BEHAVIORS = [model_type .value for model_type in Qwen2VLConfigBehavior ]
35073535 NORMALIZED_CONFIG_CLASS = NormalizedVisionConfig
@@ -3634,7 +3662,11 @@ def outputs(self) -> Dict[str, Dict[int, str]]:
36343662 return {}
36353663
36363664
3637- @register_in_tasks_manager ("qwen2_5_vl" , * ["image-text-to-text" , "video-text-to-text" ], library_name = "transformers" )
3665+ @register_in_tasks_manager (
3666+ "qwen2_5_vl" ,
3667+ * ["image-text-to-text" , "video-text-to-text" ],
3668+ library_name = "transformers" ,
3669+ )
36383670class Qwen2_5_VLOpenVINOConfig (Qwen2VLOpenVINOConfig ):
36393671 MIN_TRANSFORMERS_VERSION = version .parse ("4.49.0" )
36403672
0 commit comments