@@ -121,10 +121,11 @@ def check_available_online(
121
121
"BaichuanForCausalLM" : _HfExamplesInfo ("baichuan-inc/Baichuan2-7B-chat" ,
122
122
trust_remote_code = True ),
123
123
"BambaForCausalLM" : _HfExamplesInfo ("ibm-ai-platform/Bamba-9B" ),
124
- "BloomForCausalLM" : _HfExamplesInfo ("bigscience/bloomz-1b1" ),
124
+ "BloomForCausalLM" : _HfExamplesInfo ("bigscience/bloom-560m" ,
125
+ {"1b" : "bigscience/bloomz-1b1" }),
125
126
"ChatGLMModel" : _HfExamplesInfo ("THUDM/chatglm3-6b" ,
126
127
trust_remote_code = True ,
127
- max_transformers_version = "4.51.1 " ),
128
+ max_transformers_version = "4.48 " ),
128
129
"ChatGLMForConditionalGeneration" : _HfExamplesInfo ("thu-coai/ShieldLM-6B-chatglm3" , # noqa: E501
129
130
trust_remote_code = True ),
130
131
"CohereForCausalLM" : _HfExamplesInfo ("CohereForAI/c4ai-command-r-v01" ,
@@ -142,24 +143,26 @@ def check_available_online(
142
143
"ExaoneForCausalLM" : _HfExamplesInfo ("LGAI-EXAONE/EXAONE-3.0-7.8B-Instruct" ), # noqa: E501
143
144
"Fairseq2LlamaForCausalLM" : _HfExamplesInfo ("mgleize/fairseq2-dummy-Llama-3.2-1B" ), # noqa: E501
144
145
"FalconForCausalLM" : _HfExamplesInfo ("tiiuae/falcon-7b" ),
145
- "GemmaForCausalLM" : _HfExamplesInfo ("google/gemma-2b " ),
146
+ "GemmaForCausalLM" : _HfExamplesInfo ("google/gemma-1.1-2b-it " ),
146
147
"Gemma2ForCausalLM" : _HfExamplesInfo ("google/gemma-2-9b" ),
147
- "Gemma3ForCausalLM" : _HfExamplesInfo ("google/gemma-3-1b-it" ,
148
- min_transformers_version = "4.50" ),
148
+ "Gemma3ForCausalLM" : _HfExamplesInfo ("google/gemma-3-1b-it" ),
149
149
"GlmForCausalLM" : _HfExamplesInfo ("THUDM/glm-4-9b-chat-hf" ),
150
150
"Glm4ForCausalLM" : _HfExamplesInfo (
151
151
"THUDM/GLM-4-32B-0414" ,
152
152
is_available_online = False ,
153
153
min_transformers_version = "4.52.dev0"
154
154
),
155
- "GPT2LMHeadModel" : _HfExamplesInfo ("gpt2" ),
156
- "GPTBigCodeForCausalLM" : _HfExamplesInfo ("bigcode/starcoder" ),
157
- "GPTJForCausalLM" : _HfExamplesInfo ("EleutherAI/gpt-j-6b" ),
158
- "GPTNeoXForCausalLM" : _HfExamplesInfo ("EleutherAI/pythia-160m" ),
155
+ "GPT2LMHeadModel" : _HfExamplesInfo ("openai-community/gpt2" ,
156
+ {"alias" : "gpt2" }),
157
+ "GPTBigCodeForCausalLM" : _HfExamplesInfo ("bigcode/starcoder" ,
158
+ {"tiny" : "bigcode/tiny_starcoder_py" }), # noqa: E501
159
+ "GPTJForCausalLM" : _HfExamplesInfo ("Milos/slovak-gpt-j-405M" ,
160
+ {"6b" : "EleutherAI/gpt-j-6b" }),
161
+ "GPTNeoXForCausalLM" : _HfExamplesInfo ("EleutherAI/pythia-70m" ,
162
+ {"1b" : "EleutherAI/pythia-1.4b" }),
159
163
"GraniteForCausalLM" : _HfExamplesInfo ("ibm/PowerLM-3b" ),
160
164
"GraniteMoeForCausalLM" : _HfExamplesInfo ("ibm/PowerMoE-3b" ),
161
- "GraniteMoeSharedForCausalLM" : _HfExamplesInfo ("ibm-research/moe-7b-1b-active-shared-experts" , # noqa: E501
162
- min_transformers_version = "4.49" ), # noqa: E501
165
+ "GraniteMoeSharedForCausalLM" : _HfExamplesInfo ("ibm-research/moe-7b-1b-active-shared-experts" ), # noqa: E501
163
166
"Grok1ModelForCausalLM" : _HfExamplesInfo ("hpcai-tech/grok-1" ,
164
167
trust_remote_code = True ),
165
168
"InternLMForCausalLM" : _HfExamplesInfo ("internlm/internlm-chat-7b" ,
@@ -187,15 +190,17 @@ def check_available_online(
187
190
"MiniMaxText01ForCausalLM" : _HfExamplesInfo ("MiniMaxAI/MiniMax-Text-01" ,
188
191
trust_remote_code = True ),
189
192
"MistralForCausalLM" : _HfExamplesInfo ("mistralai/Mistral-7B-Instruct-v0.1" ),
190
- "MixtralForCausalLM" : _HfExamplesInfo ("mistralai/Mixtral-8x7B-Instruct-v0.1" ), # noqa: E501
193
+ "MixtralForCausalLM" : _HfExamplesInfo ("mistralai/Mixtral-8x7B-Instruct-v0.1" , # noqa: E501
194
+ {"falcon3" : "ehristoforu/Falcon3-MoE-2x7B-Insruct" }), # noqa: E501
191
195
"QuantMixtralForCausalLM" : _HfExamplesInfo ("mistral-community/Mixtral-8x22B-v0.1-AWQ" ), # noqa: E501
192
196
"MptForCausalLM" : _HfExamplesInfo ("mpt" , is_available_online = False ),
193
197
"MPTForCausalLM" : _HfExamplesInfo ("mosaicml/mpt-7b" ),
194
198
"NemotronForCausalLM" : _HfExamplesInfo ("nvidia/Minitron-8B-Base" ),
195
199
"OlmoForCausalLM" : _HfExamplesInfo ("allenai/OLMo-1B-hf" ),
196
200
"Olmo2ForCausalLM" : _HfExamplesInfo ("shanearora/OLMo-7B-1124-hf" ),
197
201
"OlmoeForCausalLM" : _HfExamplesInfo ("allenai/OLMoE-1B-7B-0924-Instruct" ),
198
- "OPTForCausalLM" : _HfExamplesInfo ("facebook/opt-iml-max-1.3b" ),
202
+ "OPTForCausalLM" : _HfExamplesInfo ("facebook/opt-125m" ,
203
+ {"1b" : "facebook/opt-iml-max-1.3b" }),
199
204
"OrionForCausalLM" : _HfExamplesInfo ("OrionStarAI/Orion-14B-Chat" ,
200
205
trust_remote_code = True ),
201
206
"PersimmonForCausalLM" : _HfExamplesInfo ("adept/persimmon-8b-chat" ),
@@ -209,8 +214,8 @@ def check_available_online(
209
214
trust_remote_code = True ),
210
215
"QWenLMHeadModel" : _HfExamplesInfo ("Qwen/Qwen-7B-Chat" ,
211
216
trust_remote_code = True ),
212
- "Qwen2ForCausalLM" : _HfExamplesInfo ("Qwen/Qwen2-7B -Instruct" ,
213
- extras = {"2.5" : "Qwen/Qwen2.5-7B -Instruct" }), # noqa: E501
217
+ "Qwen2ForCausalLM" : _HfExamplesInfo ("Qwen/Qwen2-0.5B -Instruct" ,
218
+ extras = {"2.5" : "Qwen/Qwen2.5-0.5B -Instruct" }), # noqa: E501
214
219
"Qwen2MoeForCausalLM" : _HfExamplesInfo ("Qwen/Qwen1.5-MoE-A2.7B-Chat" ),
215
220
"Qwen3ForCausalLM" : _HfExamplesInfo (
216
221
"Qwen/Qwen3-8B" ,
@@ -236,8 +241,7 @@ def check_available_online(
236
241
"XverseForCausalLM" : _HfExamplesInfo ("xverse/XVERSE-7B-Chat" ,
237
242
is_available_online = False ,
238
243
trust_remote_code = True ),
239
- "Zamba2ForCausalLM" : _HfExamplesInfo ("Zyphra/Zamba2-7B-instruct" ,
240
- min_transformers_version = "4.49" ),
244
+ "Zamba2ForCausalLM" : _HfExamplesInfo ("Zyphra/Zamba2-7B-instruct" ),
241
245
# [Encoder-decoder]
242
246
"BartModel" : _HfExamplesInfo ("facebook/bart-base" ),
243
247
"BartForConditionalGeneration" : _HfExamplesInfo ("facebook/bart-large-cnn" ),
@@ -280,9 +284,7 @@ def check_available_online(
280
284
"BertForSequenceClassification" : _HfExamplesInfo ("cross-encoder/ms-marco-MiniLM-L-6-v2" ), # noqa: E501
281
285
"RobertaForSequenceClassification" : _HfExamplesInfo ("cross-encoder/quora-roberta-base" ), # noqa: E501
282
286
"XLMRobertaForSequenceClassification" : _HfExamplesInfo ("BAAI/bge-reranker-v2-m3" ), # noqa: E501
283
- "ModernBertForSequenceClassification" :
284
- _HfExamplesInfo ("Alibaba-NLP/gte-reranker-modernbert-base" ,
285
- min_transformers_version = "4.49" ),
287
+ "ModernBertForSequenceClassification" : _HfExamplesInfo ("Alibaba-NLP/gte-reranker-modernbert-base" ), # noqa: E501
286
288
}
287
289
288
290
_MULTIMODAL_EXAMPLE_MODELS = {
@@ -298,8 +300,7 @@ def check_available_online(
298
300
transformers_version_reason = "HF model is not compatible." , # noqa: E501
299
301
hf_overrides = {"architectures" : ["DeepseekVLV2ForCausalLM" ]}), # noqa: E501
300
302
"FuyuForCausalLM" : _HfExamplesInfo ("adept/fuyu-8b" ),
301
- "Gemma3ForConditionalGeneration" : _HfExamplesInfo ("google/gemma-3-4b-it" ,
302
- min_transformers_version = "4.50" ),
303
+ "Gemma3ForConditionalGeneration" : _HfExamplesInfo ("google/gemma-3-4b-it" ),
303
304
"GLM4VForCausalLM" : _HfExamplesInfo ("THUDM/glm-4v-9b" ,
304
305
trust_remote_code = True ,
305
306
hf_overrides = {"architectures" : ["GLM4VForCausalLM" ]}), # noqa: E501
@@ -335,7 +336,6 @@ def check_available_online(
335
336
extras = {"2.6" : "openbmb/MiniCPM-V-2_6" }, # noqa: E501
336
337
trust_remote_code = True ),
337
338
"Mistral3ForConditionalGeneration" : _HfExamplesInfo ("mistralai/Mistral-Small-3.1-24B-Instruct-2503" , # noqa: E501
338
- min_transformers_version = "4.50" , # noqa: E501
339
339
extras = {"fp8" : "nm-testing/Mistral-Small-3.1-24B-Instruct-2503-FP8-dynamic" }), # noqa: E501
340
340
"MolmoForCausalLM" : _HfExamplesInfo ("allenai/Molmo-7B-D-0924" ,
341
341
max_transformers_version = "4.48" ,
@@ -361,8 +361,7 @@ def check_available_online(
361
361
hf_overrides = {"architectures" : ["QwenVLForConditionalGeneration" ]}), # noqa: E501
362
362
"Qwen2AudioForConditionalGeneration" : _HfExamplesInfo ("Qwen/Qwen2-Audio-7B-Instruct" ), # noqa: E501
363
363
"Qwen2VLForConditionalGeneration" : _HfExamplesInfo ("Qwen/Qwen2-VL-2B-Instruct" ), # noqa: E501
364
- "Qwen2_5_VLForConditionalGeneration" : _HfExamplesInfo ("Qwen/Qwen2.5-VL-3B-Instruct" , # noqa: E501
365
- min_transformers_version = "4.49" ), # noqa: E501
364
+ "Qwen2_5_VLForConditionalGeneration" : _HfExamplesInfo ("Qwen/Qwen2.5-VL-3B-Instruct" ), # noqa: E501
366
365
"Qwen2_5OmniModel" : _HfExamplesInfo ("Qwen/Qwen2.5-Omni-7B" , # noqa: E501
367
366
min_transformers_version = "4.52" ), # noqa: E501
368
367
"SkyworkR1VChatModel" : _HfExamplesInfo ("Skywork/Skywork-R1V-38B" ),
0 commit comments