-
Notifications
You must be signed in to change notification settings - Fork 158
[New Model]: Add support for rhymes-ai/Aria #337
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Labels
Comments
Currently, only models that work with group topk (like DeepSeek v2 and v3) are supported. |
@SidaoY Could you help to fix it? I also want to use Aria in Ascend. Much thanks! |
Thanks for reply,could you help to fix it? @SidaoY |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Your current environment
昇腾910B3,依赖:torch 2.5.1
torch_npu-2.5.1.dev20250308
vllm 0.7.3+empty
vllm_ascend 0.1.dev1+g233246d
transformers 4.49.0
🐛 Describe the bug
加载LLM:
报错信息:
在NPU上加载Aria模型(MOE)时会报错shape is invalid for input of size,但是在GPU上加载时不会报错。 @Yikun
The text was updated successfully, but these errors were encountered: