You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Each conversion causes many compatibility issues, significantly increasing Development costs and time.
Since HF models are updated daily, AI models' architectures undergo frequent changes, ensuring compatibility is quite challenging.
Give an example.
Mistral offers various models, such as NeMo-12B, Mistral-Small 22B, and the latest Mistral 24B. However, NeMo can only perfectly convert Mistral-7B and Mixtral-7*8B. With significant changes in Mistral 24B, converting HF checkpoints requires substantial effort and meticulous validation, creating a considerable challenge for developers.
Has NeMo considered being more compatible with HF? Directly load HF models without checkpoint conversion?
The text was updated successfully, but these errors were encountered:
Why isn’t NeMo compatible with HF’s checkpoints?
Each conversion causes many compatibility issues, significantly increasing Development costs and time.
Since HF models are updated daily, AI models' architectures undergo frequent changes, ensuring compatibility is quite challenging.
Give an example.
Mistral offers various models, such as NeMo-12B, Mistral-Small 22B, and the latest Mistral 24B. However, NeMo can only perfectly convert Mistral-7B and Mixtral-7*8B. With significant changes in Mistral 24B, converting HF checkpoints requires substantial effort and meticulous validation, creating a considerable challenge for developers.
Has NeMo considered being more compatible with HF? Directly load HF models without checkpoint conversion?
The text was updated successfully, but these errors were encountered: