-
Notifications
You must be signed in to change notification settings - Fork 303
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Error Model Loading / extract-and-run #1756
Comments
I have a rough idea what's going on: the base branch in the PR is not main, and the actual one was not passed into extract_model_names.py. |
Feel free to skip it while I work on a fix. |
There's an error like this one as well:
|
What can cause this error? |
Problem seems to persist in #1805 |
Maybe we should check in tests if json file is correct without downloading models? |
The script will scan all model metas and check if each model is present in the JSON. If a model is not found, the test will fail. This way, the test won't download any models. |
I get the benefit of not downloading any model files, but which JSON file is being checked against? How does this test actually validate that the model loading? [edit] Oh, by json file, do you mean The core issue of |
The main issue is figuring out how to add new models to |
Good call, let's go with your suggestion. I feel @x-tabdeveloping 's case (adding many models) is rather rare. I'll work on that today and will update when ready 💪 |
Maybe this could be handled similarly to tasks' descriptive statistics. The wrapper could write all the information about model downloading to a file, and |
The suggestion above works well with adding models, but not updating models. The json check alone does not influence the test outcomes. What we missed was whether the current target key gives the value "None" (means the model loaded without issues, if the script was run). Here are the cases:
In the update models case, if a model passed before and passes again after file changes, git will not detect a change for the json file. NoteThat said, this version is not able to check whether model load has actually been run or not. This means that it does not stop someone from hardcoding |
Maybe the model could be run on some mock tasks (or all mock tasks) with the main score and MTEB version added to json. This could be used to check if anything has been updated, and the comparison could be done using a git diff by tag. But this not aligning with previous idea |
That could work, if it could stop someone from hardcoding in the json file. I don't think it does :( Another approach could be only run one model from a changed model file. Right now it runs all models from a changed model file. This might help side-step our issue with |
When doing commits, it appears
example run https://github.com/embeddings-benchmark/mteb/actions/runs/12712643686/job/35438686654?pr=1749
@isaac-chung
The text was updated successfully, but these errors were encountered: