Update config.json
#13
by AntonV HF Staff - opened
See https://github.com/huggingface/transformers/blob/0fe03afeb82e1a435a75704d1f434c47e49a0bbb/tests/models/ernie4_5_moe/test_modeling_ernie4_5_moe.py#L172 for the correct class, sorry about that
Didn't use the auto class in the test
Such as JetMoeModel, Qwen2MoeModel, we consider that 'Moe' is more suitable.
Fair, let me change this on the transformers side then. I'll close the PRs on the other repos :D
AntonV changed pull request status to closed