-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Quantitative testing of models on HF #6
Comments
Hi @ochafik, some of the models you tagged as missing templates (like Jamba) are base models that aren't instruction tuned, and so they don't have chat templates! For the others, I'm not sure - they may be genuinely missing chat templates, and users are just doing formatting manually somehow. |
Hi @ochafik and thanks for pinging. Congrats for the release of minja! 🚀 Re. testing, do you interested in testing in a larger quantity of models? My proposal is:
For the models not having template, let's check with @Rocketknight1 to see if he has any solutions. |
@Rocketknight1 Ahhh, makes sense thanks! I’ll test Jamba 1.5 Large asap (edit -> #8), seems to have an interesting template
@ngxson That would be fabulous, thanks. Happy to let my box crunch through them all (once deduped; I’d be surprised if there were more than a few hundred actually different template contents tbh) |
@ngxson suggested to do a larger sweep of compatibility tests on this llama.cpp PR (cc/ @Vaibhavs10 & @Rocketknight1)
To start with, I've added the models listed in https://aiworld.eu/embed/model/model/treemap (top liked / downloaded) to the tested MODEL_IDS.
Known failures:
Some others are playing hard to get (where are the templates for
ai21labs/Jamba-v0.1, apple/OpenELM-1_1B-Instruct , dreamgen/WizardLM-2-7B, xai-org/grok-1?), but otherwise so far so good :-)The text was updated successfully, but these errors were encountered: